var/home/core/zuul-output/0000755000175000017500000000000015147514250014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015147550036015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000362711515147547647020311 0ustar corecoreϞikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB?KHtufX]]6f;l6? ow K-b6"οƼ>UWm׫Y_?|uݗ[y[L-V_pY_P-bXwûxwAۋt[~ _P^~&RY,yDy~z]fs,l<L& " d :o5J=nJw1f /%\xiƙQʀClxv< |N ?%5$) y5o? fۮ?tT)x[@Y[`VQYY0gr.W9{r&r%LӶ`zV=Tooz2¨(PQ wFh k0&S V3M.*x6Ql"%qYHz^?cI۫d@FNsdxό?2$&tg*Y%\ߘfDP'F%Ab*d@e˛H,љ:72 2ƴ40tr>PYD'vt'oI¢w}o٬owko%gQ(%t#NL֜ eh&Ƨ,RH 4*,!SD 1Ed_wkxdL3F;/u7Taqu5Ոӄp\2dd$YLYG(#?%U?hB\;ErE& SOZXHBWy|iZ~hal\t2Hgb*t--ߖ|Hp(-J C?>:zR{܃ lM6_OފߍO1nԝG?ƥF%QV5pDVHwԡ/.2h{qۀK8yUOdssdMvw`21ɻ]/ƛ"@8(PN_,_0;o_x+Vy<h\dN9:bġ7 -Pwȹl;M@n̞Qj_P\ Q]GcPN;e7Vtś98m1<:|a+.:a4nՒ,]LF0);I$>ga5"f[B[fhT/ɾgm\Sj#3hEEH*Nf äE@O0~y[쾋t=iYhșC 5ܩa!ǛfGtzz*з 55E9Fa?Zk80ݞN|:AОNo;Ⱦzu\0Ac/T%;m ~S`#u.Џ1qNp&gK60nqtƅ": C@!P q]G0,d%1}Uhs;H?)M"뛲@.Cs*H _0:P.BvJ>mIyVVTF% tFL-*$tZm2AČAE9ϯ~ihFf&6,֗&̴+s~x?53!}~Z[F)RH?uvͪ _5l *7h?cF_]CNnW)F5d,0SSNK9ް4:ÒozsB<^+鄌4:B%cXhK I}!5 YM%o<>"ہ)Za@Ι}YJz{ɛr|hxY/O$Zøu32Eʉ?i¸ `eH&MJ!&ᙢ(<<-ja0Tazkm{ GYә7}U>>a~W;D=;y|AAY'"葋_d$ Bshe NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'W'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJB/_xY.# ſԸv}9U}'/o uSH<:˷tGLS0l/LKcQ.os2% t)Eh~2p cL1%'4-1a_`[Zz㧦|k˭c ĚOρ_} Ewt3th?tvͪ{~;J0= |JUԍ;Iw}/9nh7l%>'ct Հ}a>-:(QxPyA Z UcÖgڌ:8cΗ|U1,-N9 dI [@3YN%:ò6PT:”QVay 77ĐrX(K&Y5+$wL#ɽ 4d-bbdAJ?w:P>n^2] e}gjFX@&avF묇cTy^}m .Ŏ7Uֻ󂊹P-\!3^.Y9[XԦo Έ')Ji.VՕH4~)(kKC&;嶑, }t&&\5u17\I@ 5O? ʴ(aPqPϟ'$BrG XF{coT;ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~7Z~okJX ?@?gU3U;,ד1t7lJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c{xnln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?t/]ps/9܎ms4gZY-07`-Id,9õ԰t#-b[uemLi_󈛥^g+!SKq<78NBx;c4<ニ)H .Pd^cR^&|El~vm9ZNi$2?$>Q IJipqc2*3g 2wbc(K.$tx2lj0<j}fa%(%#g*{C.ު s^_0Mn}#~sz1}z}V? ֝8CWA˫kLznbWo 'hG iׄ?:JmE1y<6"` T7h|ْXe|if]voɼ;ʪ~5/nQ?s dpxu\uw>=QR-Mݚίd rtc\MU&IzhBYň(&&WV#@..$ 93XMBQϷmQm{mOߐ/ۼ[𹫘qݎt6Ym|M$ 6ڟ.x4 TMXbXj-P\jА޴y$j`RO6A"EkuS"WCW\tniV K`pC?fE?~fjBwUN!N5Cu!8WqRї`m]; lau]o@ Fi:K [Av*_9B0qV(pP4W =d#t/(n&P;)_]µ!doR0`pl`n~9Fk[ٺ+4Hhao-jϸ??Rl`#/ѡZ af#rjcl ^2BQG7'Dff^f!8:/paf LnU*P(8W[U6WXZoѶ.)r$KC%(;6.qp#YI$A@EEdT+w'+'A7㢢V"+aQ33^ќz9Ӂ;=^ۭ7h9 lr_qSq-XbsK، JBJbeOfOAsg31z=ʬ-'vlQ]mKͦAc?-֋6rR)? I?ytwpC'P/9} V1}g$>8ac#sĢB\PIPfwJQH;Qxm &GBf\ZA$Ba-zk$}  L&T+̔6vmEl 05 D"wO>"J~7+0_tK%XU͍ &dtO:odtRWon%[ZihRk71VDqiގ\<:Ѓ3"gJJčE&>&EI|I˿j2ǯɘCGOrbkI%8zPU:LNTPlI&N:o&2B~ LfZ eWzRSrH ֱw;SpM8hGG&ƫEޕQf7s#bd+SDL ,FZ<1Kx&C!{P|Ռr,* ] O;*X]Eg,5,ouZm8pn;$B/,X Q"SoUG6 !ȴ ,!NB ^WKy⩽]x99?u0軡*uG"  f.s{ukkN<ƽN!n<#)u62-anOIq;6z( rx߅ eu-\q!<'J[FJ,4N:=6. +;$v6"I7%#CLTLyi{dPt/ ̿f3Nd0 ~n~YͤBOK&9<{w7U6=S0+ň+KMiw(W ]6ȧyԋ43]v_ r{i=s BG.k7%,ڃpzU4(6{\٢K 5XGU/m >6JXa5FA@ q}4B//Re&#c5t'B6Ni/~?aX9QR5'%9hb,dsPn2Y??N M<0YaXJ)?ѧ| ;&kEYhjo?BOy)O˧?GϧmI C6HJ{jckkA ~u?u'<?gd֐eZo"h 1YB s|-ö]u-_. l?Dq#@ Ba RQXV-aH7C:?pM  Sy5crc-X/#=Im415ONNLu\lA czZ*8/wMmeüze1N싰s~5$lISG*"oI'3".zɪ)u #T)D:fci[*`cc]崑!"QH: y/hUkmA A{5tR<$}KrTW$md1"#mC_@:m  u:KY < ҷBq/Q|FZEP_O1NFPba_odh(+PKO;!\;3Iam:xg|Qo~ˊc,߹,=;2lVǻXEo3p8튠هynpEֆgȤ8qc3SR&&;h8_ߍu>MPsZ,K44|MZ'0-bB$ !)v+xFd q w帖+8VKĸ@8x?4=\HGbIqEǴfW I8!mR94G t̆xɯAߙRўń M*IHxʹǙMVA*J'7FYFJPvBg+.;q7Kǰfξ}?ke`uR*%iQIlC8P_funl8P銗KDi'Uvg5|@+ ^wm 4\xNt]#(MCu!n[mi3)WfsF:M"uo?ϹhoY5-- /IDdv0G{ ̏hi|9Y"mmasSb_b'u*;B5 @; KE.a<M1@gI͙A~IQ%iܚ RWgޟ{u<<}mݻf~xk@o5iw`*,ל{=whJ^p\R_gDJnxkG:8@ZO'rC6xӁvGT.+-+Ox2c6 .1zҪR "^QK TF )㢥 [ Ƣ x'Oz=CTU_L~Tb_,n֪ʽw c%lV1ù=LAnzpvǸ-[nf &E1K3W^5@ӐCʤYVX Fq4I1'$ge),~jO'E]'' ( o 8qܦznl5tu@{Lm`Mi C4tr۾֌ Z3q0)pQ}Tgĸvބ%=89 ㈤͉Ĵt2zG;ro&7?tNGO\AsfV('I rcb2H1G<5 Nj̒fGev.F꽐&h@U_a4/6X;[nqCRu1y=㊻Is5÷-CL(lT@}3,ʐoHKFϗԝ(3ogגm'^dI<|5* lH#=M7\`'$tYWm𓶝?Di륍sF,]V.SSJCҖԻq=ky^L6/R%eZ;i&8r [EEp<SHN:ӫq{L$ߋQ{ζ(F_E{j3mqfΤP-j)H˧]Tq~g d0/0q߉!_yB.hHޝ>+EV؎x0bv6 fd1^ 2ӎ H1H#Vu]dڙ@'h9D]:U:焓nf}c/⨼0Κ_pL/q7j!dT/E n/p=t[h+s?Ny=M|o'f|l.g_v\TBZy{WLE,^jS/halE 98 {(oѢ-Р΅ 7ցl6618ł_1+=du).s¯?.c[;''=m2i\y6Kt3X7QUV:;.1& ,5΀j:<< +OYO?78In'abXIǣO;&V\DŽ0,9f ħOO_#[l:h8wݣ19\:f6:+ &3}9uzCC ٹẏS<dz(̯'CqS<;1$E[nPr/b;E'x[M+E;,Qƙ1*dcsO bSŧULy8۬ E;-S*v|oãee@7.z!<}Kj#IkXW Q:U>fQ*y Isc`% s/A[R$Ocnи%G_"|ؔ^K_y׏<:n:!d#[7^.hd}ӾP'k2MؤYe/k!ca g/^wT j˚ب|MT_qkyub6()ęJآSz5;)Q_EAVح m3 Ō7$j1\7XD&?]\9Ȣg:$1`+vur?]8=%Ml%.İȖb?A,tpʼn)dk!SV nm=\ȁ-2=|5ʥ zi 8$ s8aK4%V\ t!Lku`+=% h&)RO*GUڇvI`b0ο0qoI`b#FOf_$q^!iA9);q`F:E Ec|֨r[RR1҃UE(Av1Xazn2a/"(9/L1X]~r9_7*rn |S.Z K9mUxBa"'4T[^2.9ȱ ] QAO=Ƅ`!1Z."ћV'rX-|_W8ʐ߯C{{K10>"=a'Kн|~+:)tpՉ8Y4cGT~$)*517l;V 6|~AVĴ{3Gl& Kq+%*?V6N{I_P:9Z\"t%>7ES5oà␻,?XAe0bX@ h0v[}Bf*Ih Km|6d61D -':l ܇Bz1U_#GXUE6u 4.ݻ^XڋX6|`zR$@VU^scG"i]qt)ǪCZA^jc5?7Ua,X nJV!; qoz[X=>NSAWE92g u`Y1%rXgs+"sc9| ]>TS"JNرWB-zҦՅu;3OgCX}+20G.@U#=7) ^EUBuYCrçң~0Ac`u0R=ljR!V*Ƅ\순*O]vÞr6 g _k@BS %fee}).~n~%r~MMp!~?~?lkdTc/wIA>px|ݸ燇*WuxxM?]g)EuXr|Z=T*Hmc 6~\i]u]=ݛoVb}y%wRwOתROmqtuO{ߟ+[{_uOq?u-|?WS_tOq?Eu-L_p_Cv .e ϿWѦUt׃wN`4ڄC~ uނ`b duhq[-Nk"-Kj'32Dz O\!f3K.qx):.qS qYқ>W Rl{y :gkE$"YDE֠Z4xK%k.%tLv7Ull- }c| ]| ęjnli˚| Id Z]0hdmD>hB֡#-&tWN ?YN: 3 xH "}C[ kӨAG4eրG&/EV$Ժ?wϰ:@VcyBFƈ?\H(m!?I#bX9nW ՈRcepO}[ s\Ve;]Oq%]X-RÈlб m5^ AjSؒd 3]%j|l#*˟ R ˨aRڛc1w|o*+ŘOi ? lT z+ZU;=eT|X-=҅CrFwT#\b~?/>, `ۢPltdr$i+tHk 3tl7h#3)vɱxMSLjnKȻ \ >ls&}+ uf^b$2[%ֶ/:掔i2lG~ V85FfwtRZ [wB X16`aSK_!;WK_3U8D'+hZ9| !8RO;"w O@C;'>|tL#LjnۖF+B9Vy"UP۾u2Ja>0ta.]1-{K1`HbKV$n}Z+&kv'ˀ*Ead<" ZW?V g>el\) Z.`oqD>tDN _7ct0D f"!!N\-8NJ|r^*A cist{=xJOd[s)t|2M++XGX߇ ѱ LKV:U}NU*7n-zߞ_EAV$4 {%V[niE|nF;9&-I,t*qȎlo㠙ܲ'w'Xq6\X)ى?Nwg>]dt.kam }Bޙ>ǖ_J ZJz܅E9t6FZXsreHhlw+ [Cr:I`+CLع )%ut\2+A!"Lhrٍ L.۪]ӵ sY4Ew`[x,!9V"R1I>aJ` UL'5m1Ԥ:t69| 3Q#tU)16^Yuatdq>*cct/G~- } :OVA#&<6JJ4E88AOec ܎y,()i7-Ո: :!8) B 8dJ`3ccK/\ct!&i㧃$>lĘ_E6=W|$/ -{$1h.$LG^FbKIjdHJ6S zp!m?e8 "(\N+:Y¦&"u8V䝏)@+|dvv?juj`6۾h#ZdXBF{Ɩ;!87Dw-~e;G\2Lة&*cUAN=Vޮ$D >iԸ>U{s]^l`+df^9c5}nZjA%sPX!8/G:GT0^1Yg~?ldTc/w2MjګU%7ߎY<eyw=xU)w7ߋߦ RۘV.nC_}S}/绻y0߬VݽǛEu]>\_W7uw|U!rx%Ϸ?sH΋P4w`[Qj0ւSzOAР]I,Ѥ1H|T=cK'X iצ) e5 gCj4[5Ui/N6H+ȶ!6g ҫ{.9Uʍ.6_~Oz_Z p!>;d9OUo_KR+|t><-BLqWՙsoY5mi^\b``Q3$[+G~HC+ipӘ͏_DAV$"{6} % ``'c8d )q[x[ERnY$Itbuuݺk*c'+bMttʴtJ,gl^/<*㠨-( ~"ZDQsPy|dۦkZY/!<"kҟ=֛7#+t-1+R#x/=R Y~V他oudrA["%9Onᅯ.,"G}xsYOac(O$2A}+Fu\EI 1@u784e4KחLa0 &m<ϲ($Bˊ2s<@E>o#i40SFyU&譬ޝ]?@=!)WQ:?ق4z|!Y-N )$x<w\5F,N g?8mÅe#mpWOi1$.yT́6P(FyJ0Whw?%|pa_X,M괽cRuU \1{?؀5ѧ?.}MV\,C}:3+a#-<0߇U ɓϤ\@ hw @WC,CiOL/"+t",ɠKT"SnL}j'g7_Ohʧ?csO041"g;Y3˶ȭ'1χh& @\6;&Uϳ4HVe9/B@{KI[C }:oI[o+%67>O[ǸsL*R/sH-iՕ< ABuJyOgoOi&-dYMH:k||7,MH^Di*E+֧nB>q5?ຟtMpIL0n>PtzH`J>@Aph휷~fҶp m?tyV.L%E@{MeHk:q!A!t k| $b!0" 2xnqQWL:8 ݟSDe0b+G]Buk.esxzAK+Gp|F5r";у*6iS,ߌ[7ߣ>=4.Vq#5TPF NKJl$ 7{ i..5< x|-<9ŵhqDd*)L;Uv#,>> y\E`К-|d&zICb$+2+||teͽw,FY?:Ǣl}<-M/$/`VarM!K&HРlֲRVk4m8@Ud>W64 &zc[UYy<6xJk, J/`l`;9☞v(r@ VUuCṆ%ƾʻJK)\ _jOmRͅVwi.嵔9zvЂ͕ZyTѮZA%i {*E;oѦQ*jx5UEb47hp:oLOJ 6ΖXu hDbaDRn) d<ɻ8Phi"ܴm_(6bX~U3H{K\b˳TV+m M$SWSmp0ݱA|.DK E;jeXǯZ±{|>dU$ydhS6י6&yPVW6ֹt--[meVXPO#VD6Sy/5_6R4; s1ɂ\[c]V,]ι鸯-0o4e(|{jXoӀrװ<Բn\&5쮽 zi5 VնqvتuU|0$j> GוtnD9b-~ʾMlҲ4^n]G*M첅IDj6ZԹ%L^SOIt<GbGO%xє DsT5Z(hUI{7?_3$M]h`iU-6Hxmj/Sܣuqc,_ gKL%m)K!39m/Ԅ!w&͞U)2 r$d^7ū:c3+e"*ѫMk^аyγSbWqٞ"G+hg=jn"Cjry{v-}Hb*˳8-{?Q1U+3ɗA-G$f~&e-KI3S>z\[6ndsoi|c'l.eb[bZGa>@wr2pk/Sau$3샺Ѧr5`p~IfA/ …=޸ 8? >%?|`7:`N;ֈ{POiYq-?ϥf: 9e&71*ں&lX'u%8vUsB%9IiڼfiW|ЗW2w_]81K*$whs[rbߧᆭ!'7N EGvk|S׵0ӟZ|S7TʉϓL6?xJT{pVE̳ ]?*oo(~(ވ4ӐM$ _[k]m:n V`w4yn%滕A}|k|?N˓ȿoa^;:~O=x1xwo W?M9x'N T@b` fXŵ@8 V+["ba}YcC]=?`"`vYƶQ%6 PJ`g^o~ #mI /)E\۽ D'baѠS $@i0ĺ" nOF|Wuuh {? * SCqP1óxnjSP ק"  & .̤ZzрtAS>Md`ӞKЇ(_ߏ`nKFB싇% ^ XߝS*3- C71CVbR/aCB.{}>3ubtQht|pbr~5~t;v_N_N٠{gXoF,8Jcf+\ Ё:eowmN:f&aHmc7K:ďNٜ L#шGbLNdnOmeO- ߫<|g}9r /ѲQ ^ Phx!+5{FD]u%Cox! uK`4l:"ΨC){rBNu^ҝ~Bn~Ne=1tf 5ڞm>zoӆ&t ]پ2 Ι AJXx0+*<pHlcJޤs@q (ۂ#J?NHtGj>7C #c;s<&{bg0+YF?G+YWm4emhxayg̜$ӄOieB\R L$(>Fu[,@P[uL -hɧO096}/Alyk0#o[KƱZjtDjnLo`x*M}Jcr,cspX>v26-uc. )g~./g g|6?24,Ӳ SfgIEưFFs9r-lv60YirY.yLx~Qeu/SQ vVp^`Dɠgy>&,,@, AI zyUG xưQVds蹮o{OyRx|Xq~|y[H1)_y狷|auh\&9JgEyN1);q{, V^\Ǹ~j}\?2d8^Al9 >sS'bb3^Z1eqb{AiR3X6".--mª(4㬊|d0FZl]}4ǚRs jCw^7tFZ-fuu6'OqUpc 2E<‡*"j]Hs- S&y\f 6UBSu N v-먩r|U膸HypkP!T́ \fP>q Ci\q l1"M.};7Bp;%E䭑V,NG1Tϰ1H!vI4C?mK0A;2xj"K^>% Rs,=} <ֲ~NX4dzxns]jymxO2N$C-3EŸ.tz8LPe3(/_A[kepo7bt{#e138s cTå@)p}ZUgQW .@jѡC +8Y;Q4f"ͷuߥ}[L{Jx6Th@ӂo- c@WZ<$$f#w>rQA9)3F1;N3Ѕo@ O=Hld:SRgI>e`3KߞkׄK390`>1Y7 )j3ukY8Q$̅(Z`#m*9wkiiI`/Nq\2uҨ@9(O*X*D6AB@7Zߢ?K7+rVJYC^Vr9o > F!xsD(ضxs:ZXOZ5tlye]XdBAthӡ;peTTYDd/7}1<-%ON 23B$W/QX~T Hz+fH& d +:H!yd+v1',˧i* J} gYoN>Үk/GVSTqЩĒo锺a~T5m]>FE2|۶bOl|wO38%Q qg t i{]%jˏjBjT}RI_0`Y5N52:Y ;  \_Ndd阞>+1Sq^➡!??pO #\@tBαoZIy6$ t^! ]U=3oHL M )Tz`kJRk-xx*2=8>NR\QS.M$â,Y4ׂ]Z4уt%NI f,N~$I2>YȿmS\e[!]<8'%EpZcWhQݏ!Au߶gx?BX.Щ~T:6/x6??Bi9e>S -.3 :e>[x֋ayg̗&is3##]=IGhk?%g)s-k& 8 Hєof?Q$_?oJ͞.w+?\t@PIQ2ZLR1 yc EZh ,>@̀ NB6R@ BpOIpoDL-,8.(Ig* kj$%:}ڍIKE ò_I͂dJtB+ 9e&Ҵ"D7[,83G[@Lt rnc=ު[uHqw}̓){Ox)@ j"$߃@|^Dc 8 T/]V!&A&A ^8m)L)be3[(3X7U+hQ &H: ΂̴姥5d]h*H6{h-**n1ß3]nbzcFV V& uM:[z)rQqE` ê>We=]DG JAlDtIjG"݃qrx;r`StiMs)=lVe#,^@ &bJ)fxFlbE@KVRxK%W) .*;ZJW^vcdkS^Fر+jɤt'YÏ: $}L۰(;Ty[& 26}qhCi8R$V͡^>stDP @R4tΎr6B, lsAT+ S$ޅ5 f.ZADL YIV4^N$Nᢒы:%^Kυũ~:t[4(h)-$({4)NptWn(&etT:Kכ3!G, R2/ݱv g ǚۻͦ`XxI =a 'jM/.8$AMX,ڹ)Zpc/ eUkJ5Uu5Hbܰ sOhI+1DעKt|YYR .*.wpdkrQ`ՓxOd96 <I' \=q~|No Y0sIHWVaopA߷ԥ/zItWݹ['7nn2C#90d8)h+M)^PxP 6P|[uͲ85<C`LO85/ VlG|mv1}Ձh&p(ˈbn5}TEN3hr("+M91MaYwa(bjbhO[XKIԑXCH:*nXpz25Ԝe"j#|)ԪA )bӭNlR$d'/A;jBl vWRh RZ&ϴ7$>5&]rA 5E{?b k IpU ᲚvJ{!QL,fb tr$1 E]ڌY򢆗$/<&~jnZݚs zHHH 2VeLI'' ,8jQ.HGqZ9]Fo.W;H%ln h8OYr4]J|]I'`5-G쫅0G)1aMثƤZL綟)A<^q޲/k${шN~3,q]ƭ{$̂JtJ.~$n&7J\2vIq@-VB[t~uT:$>E ?2q>r-i c4e <,)'# Fj<iPQ潲 "R$Oi9y{raD`9rX(- bZ ;ȜmjFbI4A w0)eYłdWܖc1үϮ^M.H.:"xVcN6b Bd6kth)/7C!)p ^p)F/7;Y&&wx9JƋrl >հg EmZ`bR%wI'1Pwx Gc FY],v(!h!74^m5Q$n^혒k C`yoɽ0R ÛWzϵ-7'cZ83\e<}ǻJB8WB^Y59?O/i'q{BM1'~[8Pjim.aјIi-%c{%#\]f1z¡3qήn3 tYݨ> hG~%9,k#rQ$r0o7 z6;-o8ZOJ>3 +- umW:t'vu`ȹ.~yml8A]SfӦ9>vBa!f!9'uŠ0JxBw!6r=mׂ⋍义?k[6W,Nd{B2VM_RxO+eQe#%ߝh'ϞT&K/ d 뢑YE8Hz͒_8 7M ܌cm'?=Tߙ;ur3Mg lTmȝpâsPECD $Z An2MkBHbmBrNUc~&76E`|ksŨ[UW =onahyVtJ~]#<{%+t,ZxWBs > SZGzKMb"s^<ȋ} wblĩ] jOƿTg;(AbC}jKX1pEg;j߈3E݄PXptެ5V"I''g-.{e|z4lؑq [v2̂T/R\Luruu*kqNwӉu, J]T~1B)سyh 4M5I: It N<\Yp<75Y8aɭ""[or0XtC< R*$-eJ؆"y= PS)#Y^Mn->wbe׌=^M^dh|ytwAux̺69U="?즰8ya乩7]$en O1+Z\bo3:WJ fsasrr6҅ۯ%b<n^`fHP$ NmܗɭtkUP`ToEbb2+"Nݘ]NWo.C/,6{۶BڞÀQInopAl?l c8%E{dG)ɤmpl9͙<׼~?v0[=_2v:&؞sT K"Aŏ .ȁANKL04{vHzlga.=~ vd{R90,.HB{Gq¤uJ[ʋpj'9%$ՆmcYp3FN1sMhRׯ-#OvM"] Ǒ)7x#$w+*^Fy"ƝyG_T7/o~m{MrZg|*FVh=o' _Xڴ]Xs}:jv=弻N`ld Ũ(ϼkWb2C?55rq4)FWM +C?=mN<F8N/㣣ű+c4>(Mݗr|ð8/'o&&'hp9'^dTbD=LTأȠY@^9S" Tљeol`Lஅ]hNnz8P?oZwZ0C8j?$>7F_5#~q"flafScώ!lGLOe'. /Aks` }P<ظ>zd>S~fMAZ ޕ+ߍ_Aķ=6dl9.Jt:#߮~-?AOcT% *ܟWXF3 >FʲH r0rDAO s`&6bc|T> i @rZNcWwG8Mu$%հځ%J~aI?ߵO"9d^Ax o_C,s>7FJ.I^{c {_.-՘л8[f˜zTP\㐄? 7ݛ @m4R7)rF !kNҫ݈HXSQ7͓k=x.GP޻SޖřZfq5c p1{ds7} f?nx20ӳѥ QV‰FDM8_NkSkb.fø%Qi9,7K8j6x6&Ee)U'ߩgMeCoTv9F_Yk .I6zd|4t7]b82V˾I$gǕ;[Pэ 8ꃪڣ'|qbi<lUT}i'y #݂ՈS{с<98]DtQZZ?NFjm`f#Nx(ZW9j,Qgx|ڥ9\*ZrCXA$YUByׯYQ 뙨䟃TsVfziM\e9r|9<.<@>o Ⱦ{>%8nmg9!!=(0( ZAiu@MVA>AYiF.Y֌lRf|֓JlM=ނ4' :ƢWT@= % UPW'_i)vLM+Y{0'(f@`F .'yϴ'}@jC )vlVG0"}mF_ڠeA+mX*0#m[C}m`ZQ> mZ$IbN޷(sz\:Yҿa4Q>_O{4`(Wh {Aaz h"K{V(O|<E`P֛?CI^Տ]yl(:6ᶏ 9q_A_j1VZGS&V3X3R:s r P&97tX4ȪR=.UT#eF[b-3Z+sF-c|[d1j_//YfD6lyxvMJI'$H XZMf%K{f*J+T'v͍7<])9Xt՗Dh۰I*QG8Q?N2m|@n'^&i$Iӟ㒜-e~PWl'{B ɎGg)BN|d$=@osyY?:^Q..-kMIh5Bּ䏇˻Xߓ1esMz?o(.^tX o>U~rC$[0o|ȥ#$N qӞIɱy;\I!ї,#7fۊہ HY30. _ӪJm k]5:Lt0 w8RmWgdKC`w(eDH+pwO@)ELtFW;υr7,9 ro !Z6$Wk-kxۊZxA+ROs9=p %%.ZA, <a C[E(-}ۊiF!*5jM@9ԚxA2P1H7Ak ;"PBokkFۊZiFl1eď]}1)` 9]}3qk-j4W%Z}/U1jm[٧ZK-MEv<|G`y:6jNv1ˇ,fL43DRUzz7?)CF)W-5oQD*bftgxAWC\6SOڱ@ |mK 7L\^ж.c+@̯8[Y_X|b{mGsU.j4foڠ j|ќ똇jUz&u7j!cRi`N.o̡FGL`ͫ*2҄s wM~=)`ȶ7Ů*aX7c,H$O\UܙLM6.޺Auj?vp L-N5LViT :Y*餼g/~7o:)k ֫ctD^`1ڙjCĽNqq` 횮Gk}Bڽ)<&YBT!*~b_ MYnEt'u96;fuf&+Oq? ?bhFDA}|+F,$ה|m0@u@SJ>o/a7P82ՆVJF#Tp]R B_|߾>Y=p:1εW,jC $ ^qm`5 XHQzÓHSM0H߭Q;N#9%,n$Be3q},n" nVk"|a0'2N#NqH"+ _Hp'Әɜ|K|'(Q},֠?CKQ19#bi fG,=wHYh+m=%56hf\rm`5l ZS6 BV8S%:^Jmfkeg,#m kqqyynqiN8Neyn3BJrOq`\@)R@%Jqc`/]D1m8tU[eKOMxd CY0)uWV'Fb>5@Cp !#`r[8Z}n `7*F9i xrDX* sjmbz=VRRFbaْb a%˄ {@Z{=zHm`͵OjD6jbȵyP.HWƻ: `,$)  WDž:7Sq>׹:7w4NMT$KZ.H%̳灹j :930OTJY%"%DK6ɱINX۔FJ)O$= 9J`[eռr$ 18Gxdp%[J= mܥm)?Pmx̭H;(Ƌ0gkk QFG9j OHˆ 1Q)T3L~"m,)rĩ\ysE ?=R!{{(\iC Z@9*|]L!޵q$2 p+Q~=d/6$luD)9Q RP/v8kEꪮJp[V\QغRk-0:s!/\x' ݦ7 +$: ,/ƇqftUY+d;o,x==yYc(>->[at(2ʸR48L34kt|)Q啩kUxv20;f !7"79+Mq^s]j5J !cuŷ-ƎDŽҢp`̽5  Jipa+-4ZUT^HrkO:"G.ub.,Sn?jt1ixKkYεu=;II8G|#@F"L@= DMUBS0ׅ)+DYsU9RwHy4Kິ`$ƀ=q !dIBwbk.Á;E 5Nu7mupih2aSJ'OqXJ$CCK k&g瞣ӡ[ܜA&WX& k: [lExlAf5ݙ nyfv6Hb(f+4bi/Td[ea=l+a;]M@=f9ѢвֵJ`εLxii ;;J6 mQ6dV˼:/-S3~( Y}[4@. &0AtL*s4Z5uf W/^4G7 7=Y~xVYp&WWB܍938Mo?ɘ;p|m&C~C䍖CU|Hs4W ޞ([+YέX$0괊S9ؤ.լ; =:& pu\c_H03Ӫ}!lrlx 4\58S-,l40a!=~u ̖.$J:!?%˴q9(zk/^C*v1Lq3qdu_}6o/ȩݖb Is&S e(QP\7(/ϪV|GB;1 &M=,I3Q (~><<_=mt>PB{ף̾hZC!>?w_?IΙ8[F$nYoW5_`ju+-/C:ܚuy8ɧS4Z̞ST>eny,7ʙ|=>_3 S#)G\gq3-{1[gBS hͫ,,e6?Q'+Ȳ޼o?x?:ͧbs;vvxaqzyʨ_Oghh|¯L/"?=(b>A](ڵ$tl-FZh{McD279Oq'k~:B|D^q#s؅0 fYǠ5f׺ZJ9ħTp9G8=~;>9ܜ)a^9n"V2v-ks/}6&|0]CSV*-^)sϲ&.;y6= Ӗb=fWI~.kj"<$~:k׏t e0±|ZKI{"inUIs[)9 %m+@Ml?HᎫNQ_e?ͧ#~8)r_^Ohnݨ,~8I; #>VwK1ٶ|_Yu>LyQuNn)e?64oQCjN[=VsAAR)L7okxB]hj`]䲃#|`/zwpޭ kk @ˍ4o$X~xho7 $:.7(t[_HQSز DX  ,VPG~_@f5 !Q-)"$rVȕa`%/p%t =l]6INGHk EDe2{L)<6ՈpJQNN<_AmZ*6qS]66ܨdSt3ӃlyVSvvA+mAI>}?Y?uq6 _s/]Hڰܳ?2|?{;I-Fr  Xumd3eH.pDk'a$TI%A^#P+Wi<$$6"q"\HRКv! ԮEԇR$` `LCfTgSI$дCG:FKB_͕$b I*(SqΫhBzco{Vfpi:zꓗ@n>Ԏ{!uu'۟mn9Gɹ QC'֒??-:l a/}$T*}r`-ycg\HVw|IWD ?v±sTtRqiV x$IL̓Wrv=0hq1Ublh{Ȇ ieC¬cb\!X4tq̔44&8 ,}h܁FNE-v~U[-mUFfJKѪ9?76秙>n5"'NL⸃7q@"Z2}hŅ?=om G9 ~X΄j΄ـN! 6VSHB{Tܩة:֩j-f^ч2/>֧:iGVi݁ae4Adqh% NCtFt"A vU 0'xxTRqC<ʗP*&[}.4X9i#Sdd9_ޅ(N.}Xg=1];eܱ܃V$܈sDVhp-K"AB+ڳ!彠ؠKd #237I7IQ3II+EC{M[-}ȸb1)n]Oюܘv^юqK;Dfu>7yIU(9df=u$U1}x#Aa/(1޷܏ 6EɆMU;8EhOe$!Ci[}ZhZ/nN6 k`dk\!^-Rlz#Ճ kq&b=$鈳VHzG45r)7A?h- ڈ$KT6M),$O=vps(h󝵓9~ۂ$[[u#t`ޟr}Rh*ŎӸ?vIp-cYQpJj護Q w*z=A+.xbDEQkiU](fsP0PԢ @BT{gUYn3'2 4[&3{FVt=6md,$:-gZ^i. c%`q\Ul+OGt|9\7' uׁ~~p{ݠe2ƭ`{h9~g~=Ybҡzut5.rD|hr~4kXo/۫iw'b<>5iӋ ?89^VlBdIJ%&M;6giF$8Xy[ 8ȋ_z,k9ֻ>A{)"9)s({)33&`EG s6j2=n(> exZ>f㓃;;s?Fhרy~o0aac|,~WmpŽr>Qr/}7!Ga}R#I7BXIwtm)!S!uf9莁;wgxpDO|[~.$x{3LT{ Ӳpu;9ڌi$8'b5식fH kZ wgo g)՞u]S"18wl;fMc3?_m+> P9137`q\p5Z^M$xnԿl׽$aҫw;rpXs5! C_r4sxqE}n;V.Kd2C g8!YґG VAz}R?޾y} Hɱ(m5it8CF5=Im:InzBr$g^w))(dqH1r31eΐ1@zLoZHLwȼ㌑rz?bo>\w]Jkǿ~W+]US7Mws}ʁ.FTK[B2MU #:8lZj/˯6נadJ'M&Π&N8}z ni4Z 6L̹S,%4ZE"2]v'!NDʫHӪ $T (wo"<}":y[M2)S$Tl[Ss{H7h;E|"-[mXc'(@Tb0[W{ j~ '(tnJI25P !("E%TMVD^`\QPHռnou{O2%3KR !U'Fl;EQ d/nɺ.EnSZT0WK[ DmΧ_},mgF܉96I!oIx [ 5" PO6ɏ}p=CQh0La )C)8xr^!>g?H7m A5v_=SRd]t8]Xa`aOhuz+¦8 }BQ97Ra\Ok]1COBos"]r64cOc}U_tQ遭2&_ID9ך7R LSU[zvi:u0}B@sz|Y#BS߹DqY6XD  )%nc5m7ʒl\txY/Zu?[[|/\<|ctkuTg!Hd/v蝁4cobi}160Q:2v'uBl8Wg5ʸc1[_.҇ElR*4<׽f)/j _Z 5s4viL{fB^mBvyn ]~tY.7Cwﭏ=H-q m *qfnH9SqP֎v.`>Ip$ΓLN˂K5!鴬ye)F_|nnnIG;=8u)P\3A.=qv 9 @' 5,#z7mx7q&!ۨa6o2RNL{AB@ 6ONge` Z?K)Y0RsuE|S] uE+ |T,E@0kWjӫPWZΤƖeء䕶&ً(YFa5k .`&W <.]`p3R dU[\ SڪpLA. )65piOe8蛚~xi!˦,TR7g`H9 0z&4RT{<~0]Pe;RNxh i_[9) ꋃw CR.SFb@u]M f^|/+QwA&\'-vMj~BDtX[oEZh 񚦫112kar`LuI%8#i]YfI'3;’KhȞ 8J?+H10r_4d) __2?VWtҼyّ1a#49i #gr|nrC]9Hv&ߖ6–Ň9Ar)v*T>žK#ÜF:j~wRtRy32 VOqXh1x)4esHu|Q5khZѺvA}+jIf4If)J08V@7߆i7wdMt>A=;FƇpvGc4>miC ej'El w)<#@$8&:.-l C01w'3R B#/ ]CC^R:`RLB)" CC[nCCC^=TN$@Bf3m 6iznx;t ~{MR&8tD4Iaf*az:,vX) U{'ZL- P:("@ #ŀBUPlW8( #Cirżr@gCh_X͟>WYuS0stj% UFA10mׂwQ] <|/\ /9IFaLH' Vwh;FѦڿe`)1œ #"x PP(r~$X2 ub|![Z2̓67߿YgI+VҶLT@HٰH[%lP `d#g H;k(9M@2ksRk0ʕ']ɿLt8>K.pp,0R5t]Wtvb&G1Ѣ)p獔"hs~nx.ĜʬrbNBeHmFPlRp2#Đ<2#eO<( Uʻ/""x0Q!: )g)FʀJ:|ߚ\kᯧL?fM pkt`k~v_}jq s0r<.t{(_oᗛMSQz0o޾O>ݾgҋK]vϻ7} &>ġ송?UGw?ՇMGs&[{yOxg m.YL!/f 57|EqEgBݟ^PR,C}_@D }wBpKWh6w?nH_am0wdv9,p (ۈdWd[~%?vG3@ےbW,X ۶qfжK c+6hJIyxUqo]Ko|/'&ؑEZ@n^݂O?c!1c~㪕?LF.x-@ n{ -5 Or|[];}lT H`k`_24m*.5UȻ;ZWH"ňB@ek+χ r,{3A͐1H401n?F,~ƶd|7NnRY7>$uo?oS*}n7F r{'[.+VZzxknCp/ u؃? 7w_+5rc@W /\I,BS~l4}Կz_뾝)w.8Txf5zpnt9ot!p-j \nodj@QQs<LRxUJղe|<~iZaJmY\糠)5 @6g7g_Yi* 5c81%V~se{Ę5Lf?qIٿ&RvN7"b˼oiL"j|/Pg!=06QK%=O?e`9iEʳ 5cةc膗6ᥗKo#_j@+sb? /H(O37<-R3,:K>-?gSXr4v\D;jh㷧 2]EB`UEICM zbhq<;P-b@Yt:r |:1 CO$$VD1&sJ!-HōK~ڜ4RmT(Xm@*D1JO>L-Y,[>T|8>Urڑ8$n<@'ɒ6Y&KdI:KB+>gKUihEkL]ʒ4iMɒ6YXj\Kg\seK7#26T zàRCSĎ}sx7EJgY;G`ӒB-r-B /a_k9ȌNeIchoi41!YKÑWC?c\;`W+;MnAY??Pk%tkvZ}jE! \l74b0W *OK>WZRN?f;ۜ\Iz4԰~i#G6;Љ&n]Xx )a+:+oSGsVPQt_W4[=^ `)ʮW|I{:yQZIQ 0<äHt9Ot6l7Z "u7i^4xayF| 0 s77棃gy[[Uݴ ݵl_Ayl`CMkAkn~-⦵ Y߹l҇A,uZ}_-7JѤ^˹^nqbAb+,"?wXMdgz5R/'clB(6A 3⧾!S bp}֔&zR33Y-ڣ+܁^s|g[4>C/FaB5zo n`b &.*b3+Jn1j9;%^МgeE/ϛ7ӟM$M{=\{ Kq~3|>/n?m+ gӵQqmi~A嚗ZΑ/)09Z}xKxZ/gc!*8\΀/[:cDDxjcjס)?6a!ߴdc Ag? $44 أ/f6^>K[cЇ``#w tYΝ 4I |dY|j^JcY1hYww^MuQ}oY^lvY}쫻W+Sˋ:=d[^eGb]3  RQc$`&gƪkKz5D"s(ݰ~)%w.@taen?5c5d 8-0 k+O RЊ6#e*S+jCd>)*NXGL)9XP}b''ㄌG  D+>-bJ5iiZ'\$ ]$+$.fu)K|F ~Q]т^ rP+ߗ_5kQWԘi{j " Gd0n4&+BfA:Uqܕw䟻HWԘQqmDRNDS#*NDS'"jHK גZQN ~7x\z0k.F) U˿yw|6|y&YgӖkvoy|ztZq>W8]X~W[7&.{s n~/3)•fl 3`XHT;M1q#YBHhQ[hH,Rpc,cG}ӳgAƌ_ <3 ];@c&H6 BZbR%O&fMqr|:hD./J88NpZZCRͳB" {G•YiTzv12;d`;|=:b{;fVSw\D=DR[ k/&i5C;K)yߒ0Xy8[gzG)-:=C\PKT/9iFi ~xz~PK`ЉA=q:-?# IԢt\hΗT|G1mC+l4V]c ˻*Mnu~5zŭ0?T8;M\hiXe A+`4x49DEô|YNZhXPq5fm+axdo'є>7 &9J_Q^âT8c 9:p_:g !Y\JŇh":ˠ(C7*=A'PEȟ ӥ!1If;k,"L[ Yih2? ;N}a,MLԗ;q^v5AvZ5;G L :y`RC`x`RO\v 嚝G`"*&AjU> L60IPUJWk1 A^H@ 1)BD|'<>izp,PKZEV` D_ `bיD~!a1JkaIA=TjUu!xa  a{ x`[ Зl;N3kc~JŭEXp\3\ Mµ?SyћKw+oӁ^S@ 5F/޼J{bg,eǍmgѱ^~k'v}NF3Qa/3 s4?. u6>JSog[Vqo (?V*Φ{÷:\p1КN؇ +o~pu.=Q!0iCfP 2_7{l4ݬnωU.v0*k(%\ EqBb,b ԍVT(ņ2Ozv?~"{޼%ipo ߝ{ߊ0Puoߧ2#HjEHT$9fpJ`96S&k) NWI,5M̰r%:TF8qBlBIJTD0vXmR6B'9F땽i`R*l0RA cN4։`,7Ĥ(2ƜB 20"CJ+3΁Tv.Qqba04faHL1Mlhj4(-V%Vא~2JɖҔsĄ r1򉀔;sN2f!j > EM*6aS%X3p̔[A74$M2ҩAB&RaEc0f< ]Ɖ:&8.6x*)͆#k" X"8Naax~.oA b`85I*ZJä> n ?Aʥ^T*oMI9$Fj‘i@hFB˥u60n>񍫰 nX|ѤbB'ąaP$RFVWe2NOjEJ%s4mh{* ×Te2jnun›aqT o*~tXtunR' ;UQϳW5 V1¶,9 A";_ r{tJwu2W)K?&XR ]b=^EDB.`b>MzC*rK }@9B If@_K)۽G,C)9i0 6#:ZN:G!w, >B5+ɮ$}b6̤Dwjz@J7^tz-2%Yhص/F!t&M03%Y aA?A.0jأwG{c5E{ĢP~R}cB8*g PBЌDES6pk. %=beB:b8k4Q8tDj@gc3r0hmJiފ{)pz+2ߤn%,B Hjt_CieT LCbZihm/H54`eSFuL!Ͷg3rƋБ$.Xzꢀ+9b@4rhMm]~DZjwW4"o2B [!= V G^m!ai-nZC_ i-0ƥA[POsaXvϽ% e%¡"[J|qکV70Nw[\QW;[ǸXg|45y[8kM7WV>Z{8퇓P~mo=^.lNۋS܉?279m wشw*gۯ}/o[q+ފs.C9[qVlŭp`cV0pzۊk o[qVܼ7o[qVܼ7o[qVܼ7o[qVܼ7o[qVܼ7o[qVܼ7o[qVܼ7o[qVܼ7o[qVܼ7o[qVh+gG1C[zut@U:P0>;72u8uaPAuaPAuaPAuaPAuaPAuaPAuaPAuaPAuaPAuaPAuaPAuyz5'zu& n@5OGvG3A63à: 0à: 0à: 0à: 0à: 0à: 0à: 0à: 0à: 0à: 0à: 0àΤ?z'B u}fP(huSKnP8=-Pxu>ݩW?aPg*P!TǨOPNwWuaPAuaPAuaPAuaPAuaPAuaPAuaPAuaPAuaPAuaPAA;+߃P1NxӪm8~7׾_wE-Z6en}\ܑ-U:Oj7\.5+7e\9@#*ЉX̂DR׉X4BNĎ#m'baZN:MD'b&z٠0A8Pl8-v"VjkDl[m-~=۾^|7ce^/\aAuJ#Մ|lM?~wD~%OW nJy飩.dLk]P㛾w?nIqR ZuRHU❈5:30hKfI::qSA*?ekޮ/DmϋjKM2&~/O|-ߌke[\Z(Cloa,V|J"{2SLRtc\z1SMvsʧ{u88skqةD4X̝ͼ( Ν5]V,H7,SHWZvt%^)Tu?tJvZzj*Mt ӕ2E(]5IWK2ZzsLWNYiy"`[_-(8sz弿jśWWt74[uA;+ڛ*R,^,68%ol7S>.^}.L$! W}x6,Z7۳9~N4\^^G U.+AUO_dO.Ƨn^6JW,O#2vxcι2߻bzM\m CE+U}q."篋޹'ӝ ?njs&N}P9堪Zep{M c [1FP>zQy 9|R{ܲJEX&j{#H%WcW{ܺʩT߆.evSir*`nE1}mS*@h~k79~\R1s;aھF=9~ ݲȥTlБ):Ynں^_boAeuv jWal=w{iҊDzϥM94V d5Q=_;w֖ʃ&7SST&n9mmD?&xsI1FCz;u8a?ӏ3`觓/)^$qʎӭ ۈE=o BKm]nI\?:=kOU[~*jmlӚAKu78(_t`v2˛w\fl9Mvl|jV1Ƶy\iζ#81_>$0M?>}l̫S_f7>V_.W/'.5YnR\I=!^OVweQnɒs[_v\ ̼4 cC f,y|h/F1N/b9yHy7,cn`ղ6QԤX1 ooPGi4`=Thx? z;;sst!UO_s z. &*~N  jB=i9ӯ2_JM[GƃTlv9&~eŧÚKAZQ- p<\aKmcͣ~o~+յs\ww641C24/wocރy铖F9 raޱ$2&\)roʴY#x1/c^>_q_e[g3 N>v58/q$cA_7q_7z-;FyY @@4'.ht_s7y(G\I(_SRmRӂ%hCE/(#y9i|_zzlB/d g?QۢRW_cGa sHtGdTRyK,Ֆ {/9/@d/"'_Dv7yY ](M*id1厹huAfS2BK "v cCbipuqGFUK~E U}wO{{s+ߖod}9GQ'yU e 1Ŏbs=KXg0E,l1I zTҪd[حM.1#rC@n:?{ rbchjx[U(pjǡ^S]ۥ>ե>]S]S]S}݈GMuyog: @K3K42[/¿.y=s5 d $'~sчe6>,>l~nܯ:I)%-DB(]0yn h~ho?tR7U2/raYY(M 'V$mw)#ϸԌyfyW5ך[OH/)n1%vS'|Y(#H q[ cr q]Un\o*R O77+:ߜKkt FdJQB%JǓ#Mv0iGQvE^m_䩌Jg D"-8wY 2v*GUlK-XxX^&GqU0$|!㎓ӝR1Ք#TӟR 9H EtIvBģEMrpxs{/^}-N{0H($hzM2D cr q]UR-'D\2ڋ@w)QP$:znCaN͘kjƑwUM,y&vF/Z&mj%-K|NT)4KWăurotkER$H4qIF.N|[TS-湚jq]U;o匿%>/$C[SC g!UҖbQu3EMᎪéWi9A~9珳{Ra~'CO5o'#Dq$6c4Z 4DH)mГ 2M9SC5()~m@o0֌i7ޅ5;wJ$~F bU3"Ƣ0**{ +f,VStl`hK dF,qp"gQq;Aw2XU>$u>`iI5N2+3E"m=\z\Yx"=cV9 "=uDpLK”.,y`VF xyWF GUNzx׌"E/ ť.zR"h˜ 3Ql*M9ʆ 1XlA.n1}*ia6מbfz3ϖ`nlӶ!⍓UNyN6qtՅ)rj*7ý/(s7.2 ,cjtXN ~,e1p=vEͷ%4=߰- *yz0o射 , *dl=dEQŎX8D -Fyn"ڲa-+s-cĻ!m`cx -R]/7-0#qv|!ƐzIynt;)Bu!nڞ-v&ӽ.9~;¯~s*U)šu#W2{jm9^(qD){<.8YַWKha9` 5oMxP06zlakG5%7k CK%lOϻ]neTM*4T_CըK"FE,s9>bByXLMݨ;8;3-SUAAƈK)1`0 t_Z/dR*o-6<=AF*9apAFEdQUa3jQn(ՍE 7&իAƈ;3͆ۿ/e٢g4Z.IbDFғY.?eS9GM8*1m&U. פ-Sd8hAfdIljT%'G\6skE%B"IlM5S KD-Ӑ,d>Ғ4^ CU=?1t 2];Hm5W;f'Ut4Q^9ctV[ӟ18mjHo{PآA1p)#A ÒR$ژkkcVd Wv ENAwԶ1!ZnKp襜m>W%˛|%X=ފ*Mz2*?G"r1\R.zڐhĕ9 ?pG| (`fiQ!:@wUf蚩 Qh1X`qǬCc^ܨA%l4r$3zo_M^}m i1b3Ȫq1 Fi$ЌhwK)D`c1Oo{Qլi,*S3c۵6}xo؎:=gjw^047/dΌ?ϪO'Ť"JOYRw)VNRB̼(fę3N+GNgZamA_8䀵dm-dRrS,>߭vϙdz8o75U4q#Wt@S`p!sfw%R=v^[=AQ-ɶd%$G,vY,#JSGdbү(1e48`ErU>-Z-vUwD!$;WKRGQk-[ ruИO zE+4l 3cW8g+/1gO_|rPge0Fe ^֦l-Oq*+P6ܠzJw>k|W%ƔA#66NEFx\"CZE9-|(5 iaM>=Px.5/tdrsRRVmpL#mu#!#4x φv: ~s3Miww:}%&=G>  Ub ;l P]m1 ;.zUE=bXzy[_\)Hr!q A^e9W]n4|3Tģx=cRN BKb$2q"\ҖU~Q1'%>:`қ^(E"g$FBDő|?ШNJ͓@&H&3f \КGq7Ӝ!8OPpf#W6ϵ 0=(>o714RSXHPSaV<7{,ęH[JuH 8\c4Yd9!F> mWf}Wuk-J}>`$lA~ts*36Ik# @2>Zr,.h,= !,TaeݻTg"tĖ'wxAwwpf7 Xd_v 1gӐةge# G.eqe"G1J#LȕgpI| w$R{Dro_p&g%ؤF4iWҨXr|(xPcjg'JP3ΑR{x_T7ll֝[jU[x-|,u/񛷇o4pI='o+{dJZ pJSt#\ht ڊL)AƬwIHx 9IJFIRk1ב`MODn5xkjGdWa4Io-QFFPnAFdw@wm8 dEeەI\](@Տ҄P wIl[(R4F׫ӱhonؙg!*9]r vJtTז*%mI-(#Eδ#P)esgu8/-XRЫ`1IC>uV%K4 Av+ف]kОG`IAi,#V]dC7_%:Zt0pMT$܆y9k8)~ d-2k% )5 հtgP}/|*A^ɶ^S Kd-̬V,eBWdfS@ HAO3ɯo[R>IjfG9sv'DKt}USKtiLqRLqI3Aw *h&EP%|-(x/$!*IнXNcHh΄MEάm~;6.c;0}fXp;PNjJ{_JAzA`.1ɶVL)Kx J:@$Bo}7;1Jցef)N/K2NdyBsä m**`":B40T<~x6ɗ:gm[|kjNvnEx۬< { 5>@1Oڦ;=>dv%=.])}t ^r S2~l &D7&&(uqiu^+OVq̩C[Zz C%ZNc*:^Wk?41,%4҆rlbpLtSI.w_6lkJ2ov'6|\40 a7qxM(zS ;$W,qd9da,]o\(].Bq^$]fxñ W 0WWLcc7jtk&vI|\J&F^y2eՉIRʄԪ*յ0FD#^z+*bbf܇i121D(u41X^j&GMdxNv t_Z5 IۻBH tÁCY)1a4ܯ/&%N2fB[{n_^|l cx;=YۼVPbSkQEfR_mu C>l6\2]}O^~So?o?exٹ > q\Ґ1K }.ʶOMէϸk J~m Q뽎|R2޵ZUqv1΢ϓ+yhy\o.g+j}Fko| Ft ֮=FV+/릗fMfFz=S#s-#"-{ۣe98ASk((&F"# 9A!5 7c 7Zƚfđ?3(r&F#e: LzFZ4ߪf1/9@NkA9e"xD,Ο,#s]{ߤ|. >#ސ/:Y/pO&(+ƌ3rt%\fWtsRRi‘ A&~ml#!6 Ƴ@B%Z <mSCԬA[Q65.>R%_: 96D"E$4yYsT|(S|ALab0`q(lkF Ѯ<3N" Wsg}86w[7w"kS4/ 5)ܷlgSX T׻Wp't6@=tCanPibH&KAPD#(. ̱!`캴!YGA{C^~¯ˮjop NZX\E΅=I.>ejs׻LVǐgc! w3aɞiqP6xϕQ!$ RL_~mYBD2`Ȼd-XOc z[R(Y"X ~Y%ZЫ$ C%ZNcŃ!݆L8Mkx@=P(-rdJX pjECBI)o2MNļ<@8Tk5aMziq 肌^!]"dyT><3Ġvcxx>IlEp 2b]f5F_緝^Do>I;$:FJk)"%U*zkڄШZm;krI"oB PΘeQyaVB%j}^i4 ݑ4ژLhzfXPyڟWSIdLJvzVAj0ANKm-R( xXB^Wf9xko{Әdk20$Kax?^_g+]{u<^x:K\%B.Ġ,Bj ?Yayn2ˣ,բ=W֘/~ˏOH_!ߝ}!0~{9WXN8gQc6_|\\LUȫlxtǿ(^v|u:("4)[BX-b.ni䃗(@-T_n=BvШc4Aa@" u_wI 9Zk2iO#/7?̕3l.% =v|܇@BW~Urwu?ko (#c6;ޣ$9|f|(R$E@3գ6띣4eE"- ArHX4\Ғ}| SL'vy5hłq&6 rWr Z#h=R1L+[ TƌXjoY$je2HvD;vM)-SoDQ^O< y+AH8ȥ=8] '*U:\|̚rл*#7X0D r.TgAWGOcNa2S=DK1@ν-1Id ATc:xlv4&h$BFA]o:x)iL#QM甜i]j^VYw]dSmmx#uWQmXXb7i W_!`-1žHHΙ=EܑS 0T4&%'3` S{+$R1צ%yYDKZ ^]-J$4TI :,;k*5p{.|UQȼ)x||}݂x.iL cyߗyBfMvg7/ቖF{,2ޓ8#+B<^x~ҔJU#L]YRL׃`;`F;"VLJubl.[AC4t.fM&qbט:p% AhY8V j2/(t Ӂ'Ehv`gA&r>9#ȥKD11[zQTUjR-;NC;J%U﵅'U? u?#56 X_@F_ڻ$ėN36ƇNw4Ӑv+2oz ‚EJyzgKTġ|px>[CE Fhɑ%fXYkB58lxÉz:pd&8[•*J*YEԁ! RRQUB,d: UwGQ8ZwseGYtىn'ƉtŹ6&;U4Y >ifW£Chyo&J3~>(5Dif&8N.8׷NN'NӔN뻵f'7lTqic?Ww NJ$H=suR{vӔ*0tN55{fv:ۭ+zgYB! V(u,Htc﬜Dv"<WNA- 'أꁓىA]kVۯվ)_~ssbD==}6M__\[`DB(2NQ"H4:Z=3A*1R+꡾h:!ƓByB^h==IJő8rӲ]4vͼ,Rԍ)Hd*Hp"Rk46zxSuէkO5) 7Y͕qqGV<{K Qݯ ;+fB{jGSP%xͶgte 1{ejCTQ QB(\8w:)GU=vj"2z=/n.V_ț; 4̤ l:_?t$WqdPR-q,*jETThnkdmJUQ5}133"JG$ţAJ8B>'#Sfp(J)xUֹk°&G!682N7工[͐^XC!Y4dɩQaΐVIIp5I{鯋ag?Cklޣ(<ȷYgldwޣ4VGf۫l^|3C1%LぞO3[{j^p<Ts 2_ B|)X^J\TT9'3LgyY}Yx ixЅ$3y _}ʀPQI"!R,$">{-MƆEZYjCYv^53 2~f0=q>> ,;pd&=z㤺z6_kK A2SӃ3Y,QG;YՊSWH[đ8~M+GiOXԡ^*3c_u!F2{  bxiʣ;pd&^.5{DgМ8jxup%+ 6j7(P€jmq Q2S]3r@axWv\Kv\g Rz ;pd&RtOTTIY:۠2רGf05ЏuH=Zq Af KcCNQZ{JwQU':E]*H3%[9 -;pd&:">NGzu4$ZҔ˼ig yg JuF>F"+[ z{qV AT٣σ.2{z^iNtb;P~< ١C3Ҳ#R# vj 󇉟.&Ij2'͠xMI'isAzؗ_^x& 'fO&om'lj?m 5?ǥQyiͼo7R?ZQ|g3:,m(PNwTp^Bpx 60ʫŵ'j^1Wɸj^d,0ЈY}02L0Ra*W3 +h|=!اIOcY(3i޼iϦT[o 02vbBT=}MLqq3M?iTPB:p}5)tu95mRjCH3Nd~UTySs.s,NqцAQɋ/hI   ,6Yl|-3 s_x&c!B5 "Y)LQΉh|rv贜Ԏv]mJW!ZA>r~Ak)1FpYRӔ:1UU[Gvxw kmo1Czc\ӊR9Pswgkrtޣ|yOoc+( &ցFzTB8 rt8,ixr)*#*?=С +/$)WW_ѧԋ,ֳ޼ȱysKGhQwv SR%ehfս>tN8|Q!_\K+4psC;c/}>} |&rE g'0 #+YpuG2g~1$gRC%[6)S$UCE{be@ >Q|rGqAw(١N)γ5|;΂γ6p(dKAl ̦n<+~;"VdwƼ^~uii O!Q 7";En9+8(c͖{j$jMF\<Icg̻^%<㦁 N3nm 8ָB唝(I ';w_QU(8rEy(~rXÅOzƳ6Q5F( d2VQ"HF(!+v!@,w!!Vu!,Fސ4; 8'Pa\E5 c%ɝoݧϟy}^m}oJ{ZaJN4e]R_2Q ]N}kǸGZTߵ#bTBߍʷXwA{G4fWcR;PPw.緸z{.fR=;׫n(Ճw7wP\Hk܍BP'R N}EVQ3)׆7w_k/x tt7o1/X ֔8B*jKk Ss%ўcqa5|;]oc R @wPC{\7[6[+)$ M5d,\Ԃ,~0 ׆ ɐkdl-m4(喙[wYT~ EEK[9 "ߎ ..(W==' HUcN%#;[xG")"^Drs&Cսfge..v4S/ǿ_*iCͳԿϮ -_, Q*cOQrFu44Pģ&Sb)ٷB$ iq1&4,_dhzv9|e.-,"vGA6ԗ'ђ Lչ=v``}sCpG}Ni}j onU|r]\hQ &ٺ}(:Ϩw0 %-hÇh,|Awq%HnF-wgEֺ*>̈WV8QI"#D!K^)Bݗ,CNSQ-6IpN~?FE|Wm^.dǐo?BaZoGkk ]1`\i%(%#F*\U9ٵQm;ց0z܆[-16~_k%`\А^rcGs>F&@dJ,G'G ׄP]W:W] %r޲(-pm߾=Y}|R PзgomGF?F3#i&4>>.E*m24)x!$܊=%G|g[jmnCm wv}bU]x%pc{ h7n|2G\(r0(Yzg`u6 i?rk57Eu @ x~IW (%SHޓiMx 94+t8OC4RnM0[ݽ]ƑgsJ{Z-.7ȄKdb*{L/\,7L%ڸLe ?n-xk"4<"8?lĺWm2 hnǯfi[w0~ juW*p͗l"r!Wt̾As $ }+HW;V M6Gi&`si<8R@ddc۩ Ũs͙3Rħ ,) Svp^r,rv| / œAʝǡieOVTNGBLš)^!D+JklCTxz(6 O$*LɘZGfx6S ): !}ITum8bQHZy2`I2k$ZF.UuF]«(T5SU@Y^Ɨg4 'lBai2IRjyga>0N J5RW__^x*R%oZ 3fߨePyŲU0ѓQn_Tvx~2:r_ mԶ[gsy6i[=N`RpA.eƺqa'Q⍠g[Ɇ+󘫷eB G0no 2ZPU9C&u# 2JkϔLhVK\0Y"JY뀐qrvpfm ϡ0+G!(d "kW;kj0[~aA[$haA|Pm mx9qe"_΀k Jqm AiH ? Bp7 ,e)}iʻ!YKe⸭r&պխo^՗(`99G@-V&$N3k{Q  2;.8)R05 FXtz#>R2f(۵qrD:hbci[鹙75m[>I)Oj$g NI4pf!]X&_4pL:'؋*XW!i83Zh.8>hOĚw+WSVsw&&L2R8jexHδ 'p$!v ړ,BhҀ^ :B`~dXrp)b-ʨ^/%2V!jn+yt*͒ C|<BAPIHKȑ60s`,)r 8@y0k0rO݋?dUg]Ax.:T;jrwGН6Ӥ*31'&&y7:Vjkӡ٬([).?>^sIi}b\\ډz Hf` gЈi`"PR8A@o|;TsϝY\Ui.-9g,vO3 4rv@_Z%w*y_r6I.fqnu\Oh0Z0e`F)q@LG/e퀭<=[:m gV,nLg!_2ymPdb't ͪb3\ӹչ]Ncs?8sA՚xzC(]Vʮgq~I/_&(J4(+ߨewnyְڏWW%IE0wIg}xW<ߋ\ixſ/'Yg7o y?w`mw{?* uD{zGw}L, fN:6j.-]o 6KN>\o=;¹igo8A瀡O8m?N-hӅj|q0r\r\뵃؏*>ã<'-{dt[GocE~^̎۬t׫x.mjhx(rls4-5I$_ L__a}a.M4A@1cԌ.Zi>{'dB+޻}ҘMοIbH7"B z(ץɟ"Gqdԥߎ4EV΂xkeY1X=ً&5*+`jGWNYuWɨ$Y|Lf8G/2jضMSTa嵽$8ʿ}G: d|XN-' ]]e0.S?FL} $s.~ o[1(/AeB}:N7ʽS^.fb~e#|<7iuV4<0-ljPD" Pȭk,]l6#+'\'Oggb!XkE&\KSqQ$3Z33:"s#su>eGY&sZal7~S #\yfGe8yZ|:F/Ǧ\&HaWWp^`Y5y6`>Դ56"KI_8$bF)ɽ@3z !B3Pj$`{S]w ޭnC(ϿPD<%e6sgmԉuƗw-Raݸ]|0 c(4 i'%P35ъ*@[06\>c?nVz}r1= 7@s^'Y 2ުd~qd)g^Ch`Z ,KNh  1횕wUI'9w @- ;;yNo>h};ڜt :8ۯfrCSr/I @ >gTdB(ԡ!ď,zS%\Qo&"KmDDt{!qB\ yqqhEH!`/BZAI2Tn;3{~~@$%җ(Ğ]V Vcj@$Dse՜D2d!Dp(;{w@ :]}8S*n3hETF3̈ط<%E`wpKorQxCh)=.(FA+ԁ V@y444d2R޻㣜 FP8#`aԋ4RRqja8 +{4 "TyRGAD_mg%ǔISEL`S3W #sQft?o{{;4qL0ϸFH0 Q$X$" )`qf֏{޿~ă,;V4cGϞn?~k(X7[S5̋m ߋ gRT|=:sД2!7`(a*Q O @q;K`wKI(8@"j^0q  J`5dbZJp]ݝ7޻N Zpi>wo;;w{/^5q= SHE\%Ѽ]E}łB2d$(>l/~=[ <ӯ2AfmVPJA5H8>"A2|D(_ HXJlvJ@*$4ᔲ8@12N(/\ ?+^t:gd#DNU`rXcD?%6vU0@?trtЮO ) Z j\"Ji!`Na$U>0`ihßf&Y=iJlUcHpsV"{^by@|N#<g؏2vcOo8›-C/Cj'II7I$e/I e,,jȝI؏Y.owj;o (4M&nfI*WYdY7ڏm6IIᶽ@ٻiK}xU-5\#nT!DV {1mqؓ*N5#,JllO__B_;piOuAs&|v۝ΛC>kfZE 6٬_9ğw}̒l.1ae۱iJ]iMaH¸?l=wLMw<,R#?WmLUL>(kJeb2ʼx,~ %i"? G x">iotp erNNAV!5 *Və3ZX`.WSs_h۶VU#}eqO^㓋$^ `;dbdT{.pڣhMbf2Gx5=ıMW[lkLMHazH ^65yi@D 0Np~Ns;{tnP6jJ\y#$|vief]QإXW;LooCKoe.>tFzCqvo^kᾗz/FmKLqkjz@W8 PcMv 4BpB$`J؃W\{/JF\=Bq%* VT<qe⡈+ iHm#WJ/h܁"HW`6ϭW(+3աL Eus`FUP8DFHraqסh7Gs+Z?Oɞ9̻L8Qh|WT؋<$mUg^crv^ltJtbo0M|#G,R~0{k4rdIL?OCi@v&?L6VWz %7K@d|N z?r;]^%rsT!p/MЖʝf vbJ؞w)s`wh9ηI39LHaclI^<~2<eo`K_Y Mqmʠ*ky/eFv&,fԳ:OӣiÈޢE!G-d~xh⃭195#AߖtR<& K?&GppYS(lEORS<$uw[7~9dnA֯Jo\z:I3:%0X8yb&oD"Uq %x%UfrbސmPj'yMbUx)~k8]߲)sk}z`_Xk@, b{cr6҆Us a%}mX)V 5^!t4L8l@g2T ִO/|.Fsԗ±=M0^H[ 8SW<-m]v_Hqee+ qŕMx+ úĕ",kW%d#"X_Sfk?> Zqfի=ժVy4I~d/$!( .#KrrCcHMvF& "tDYV*Vk6]k8jojojojojojojojowYm6wT{joh_+_XR J9 5 e,V Ejb< $N9 d\aiglo!PbEՠp'B,ư"Z8}s8Ի/ۻNtvu`z] :̧;?ۧvgs=8zwV`fQy6+QHbĠi $ 1hD"d'+PP.yv8q/:WS"A *@WI q1kL'>G 4BnO)T0%AaV-KāZ) 6$R5t)  F,QؒE$vWGn]uJ ى7Д7:-cF;;ҟϓM@sv@Ӂ(ޘ|4/W%5]*pqo'Yoڵ%k۴%.Ŭh[WoX;RFB*0*ц_<0Wn;{|k4jv_]J/\뢟g@;vwXmX7Wd2>t'/f<8)Vs }PZJqj"Utn+9`_Z!qp.vo~? TF} j!OΟh֓׭'''g?o8qcml66Rr[o+iqǶshwǰnn7VdBTPj8 dMF)PʋCt/.| i05{[EW?J}_tˑ~#sH֥2^z޲uŊ :\%q:/aQ/guln3tY1̬VZ4WWpi,eˊqʿ}W-#6%Vi+rG,V̢6[0)H_WS|`wo.Y:WLS.H һ vLnw0;f+cŚugO,b!Iی{4.mTʭഷ$~1@>)WYݯ YК3f֎}yFQ!Z1 n"Mel`Xbyx]}e90/J$ex+DjVciַm,`4~P.4aњ+SAiB| kcC@G׍-J[UF^gLZUhHCH#|GF0 ,u,^"Aj_Ib&-ʏ4SPjE_BؘV˄?FPEd_aO1}$$g$i( Z$}ʌV%x:O5/C ExFGB)Ϣ_|X5Z)lV " [\׸b+ZDl%xd(01i)79#CcOVDck\<¼31 7L853S/[1o8pPvi#rI$Q nf>gKLg9 elH.ɍ"-X>R&J0Zz'II.P,N/So6Py,μg3am`In`"K7%|Yĭ #ckY\_-`b 9Np:ArJ0o0.niub^|i#51xRbd8d}~uA2 Ajze;Е~}ݓ7OSf֙jv* sZy[撵9J0G?0ѥcBY =G!2JTxڪ1!gPyJeq;A04d-,5C;|ӜJ0(.1eF[ !@VJ m°9#`eɣl-GmYApbOJvi)Anp- Y↘R^Xq*2|An1"Rcl3F( C'htqMBRiL#NV 7^,1|МdY vyVW]5[uY* V$`2"X)Z5xlT83B cq5[" 8 #-ar0o*lU^242FQĬE0sF(` 0#NKtX450,X~ W3B ѥu^ Hji=DD`S#`^3^0TDno 6via-,Z+00fsF( ipV Yʅmt>g~7wv!ܡ_[C}<\'K|k`{mDPy]y igmUs85YjGiȪ7ۚ Yc[i+[%m5)mRJXńÂ";./I}m$ʓ?g3U<fQ\D P`e׳|J0/gB!"CƧ%y5:Ǚ&@'6bjy κߞf={~)K }ӕ#ߏ'K~ϩi؏\nq~Vڅ>ZݼpIQ[r`thg= g G]\/|y A2s`%{}&vf1|0{xޭe tE~Q뷝O./Xjms} 4Mmpg'ͻjf}?MM*9p?:%__y漁64]/4m6s){'εH(FRp(HVX@J`R9[!8Eq *r!q?3B P ̌(XQ)Z9e"A%7FԕH[+< iK20  sF؟y u>VX"NzMEml^:q9#`rS:21 AJ 4݁â58b'Җ9xZlIVrZ2ْ|%ؔ^AIq.O%% VT ¨nPyQ$DmшSJ-bp)5i 01B %gQkMHD 9E $q9#`^)QzKUqi$rJDcqP=J06vysB^0ɧ_] [%,09Tw[쐳"_01XD5Oӌx%8C* A(czflhq¥; &r"#s)VSؿ1$ݾ 'TuYNYʵqvѭgyj.Šoj ]5[m>Ԭjrw,*,uW#gY`ixЅAڟfka vsT{UG͞.џ2+B$WJJeZέ!96NH`"3F+j9`tŽeoߓ0&WJQ^û~{2kLߣ˙ϳe[m1l!VfTmf[fZ:GY}jjyqgk0b 79|3)RIƠ:@N5^t*X["Ԗc)W+"֊VkEZ !UT]-zǮ}zǮhij̈*Jk[PVj ^TXՊ'XV`N%#WZG|D-QG|;M8JLdqes4RZ;@%_8 T#D6D \=\%*pup6lLp6|4p1 'jyWJYWPCՈ*h  \%j9>vJTJ\0k95x큨~$ , _t7p݅G.,Лb=Ba[$qg52rZoq)ח"hfhnz>_NS<J]P7@U5e_:J Hi%&u4A1e#O_]d?Es8`RfHaNDIѢ7-R YI$=:J !P磽ɔ_oW,ȊS-@ỔCiv YٛS~6߮>\N}XE Y.`.>?f+[r6wogtOlP:oOn< M%UJUsoM,fu ArIe{ 5?|Zs8%ߵ=E_7_`|5zq&HEε>k[h{_4—˙]/w`=]%iJfNW"&K+%.aSn٠(ʴ+x/'w*jn~Y2<(ŮSPm M:3~;e=9(63_B{yV4(VJ}dIMxΊ=DI򉔭V@Vj‘Ue"Q˥Edn52.1@/(]о鮄ּQ{yu%NUzIA}j8b#x8"+/(1"+R#9/"zjQr65@ݽ%2 Q%XȤ!So.2>im_9H#PBzqT 4X@-GMTRZ' h*+X*QA0pup4O`wM~]:r^b),0WGGc_u8%{&?::i3Pcz|!j'lglnEqyWW+<s!kf+ͅdH(R%[BTp\H9|;2jLF5z4M"WƣI*s PI0 z4و*v*Ndž 'U"%PK>vJTV:ҜQU"X#+hD-;JQ/}"\Q#tBǺJJ5JԚwJuWW,S ܿXRp2W[66Ԃ/B$z-UUD0ՈW3JU"WPK:vJTrZ E%\ϑ,W\W@-*QIe+&c:"*X*QR箶R W'W#+ Xa1JR9J vpTz*QIX+$U&\JFJ?w+-ziW^W̢L]R)U|W;w &N'm}?RzmR;SkQ̌\`Mh5;Јo9|۲UqPǕ!\==j{*dUVh-3Xԩ铮vnړnlhkIj9R! fXjJd T|5zNÀLRNlrpIvW\43x+vޗrӝt1[tSh@kIO7l}t]^ʙ/аU$ rꔛ/="bY&>ۑ+^82nXlE%Gv!o52 *rOPD/'zm1l) 3Iٜ/BҚ]=sGe7Avs 'xWǽ/R]bw~Zm_oֽw>kv3 7vYţ m\f˶Ϩzo7ェZ YhVYkv׫dWۦX Q"v&0!؍@ nV9dSbFkVvٜglkU AKZh**q#K+Dc`Zv݇`v&E&A!LP*Ȓ#t=0-K6)b86Ez֡>Uݭk tU<`xi ~=`DPKɒ"\SZ|^E:A!2&DW&{fh5:]!JK{:AL67)=q'ލA3> $%dv:;zuz5 ?csKPࢂ58H!plz:*ȆR8P}}Z:lXu`Ϣ;++U:8du%uA8ף\kze)e4؈d2kTfhzF(+{Pd4B&+lL՝#J+{:AE+ly:Lt%-}tut 6UIzD.T JuBtuteJ&DWXI ]!\ ]ZM;6Q> ͸4ѥ:\d?.t)d8~Z70-[YvW,Hئ?kX|mXUUGݸgOu,B=8#i_KǤ4wC'" uyԂo>6(b^¨`eNu&_58b>ޙ48.k!`XE.[ hu4uY~r3 c6xl(pÕβA%w?9~xa\~t.n*ZgնpQ:RUjC7YeqaWg>Vw.NF7WKlWxk7\?MϢo;xPTY}tޯzP.{RjYLgMc!l)Ѐw_WҬ2P߬-܍B߿Jٻ1(؍AH6|EtލFq~dkVB;[=D)?alw>zE5]׳sՂf)%3|nuvr8@9 UVN,j;'W?~[tb!J@D9gnS=5 }sHCx`feiL=I!Z\ 4rahΤ3%s(&R9+γ>90Y)X8Fl od̲Kw|(8. gGl2-nJv'<[bf;QA^=()8{$;܎ixUV ~Cڽ9pY\*8RdWqK9t<^r/ܒQ Vy_vWn>o^H m=(sdXg Dh WB # d2 ʤ@ѓO, 'ޅ?ۃKpq7Jiq!:ӡ KG}xoՀogXч}>#L(qGC[a!UJ?:l_}NJcjO؈ro4UgB!I  A 1"f l@ +Ora(צִѓ.%eKu1BP[ Q(!gӜ_< xl*Z^SnJw0Sx keTfї/\Bk(#չ}L5KҢEU:k,Z/ZDOpѢ KA9଴E}V+|;]J!uQ -KJB8K#m!MJrkEv5wZ}y2Dch3 v> تOFb(iuB#VG?e\SI&#dOO0qTľlP6k7ot΅ ҕLvF%:BKPVcN U]!`c+h2o<]!Jz:A Mt++H$]+DEOWL͡7g iX#Jfp-9.]5BpmwL2=]=]MA@MJ J =dz|;R](!(^ЁȝX H4$μOPl1InZ K* g[_堔 P7)2EYwk9’6jrlOy7&<Ҋ̒̚YSZHi-=d@5wsrF r5ǯW G2Xk\ފPvm|Y%c[67Btr'?KfJ !W}%tnF|2n +EeHbL!^]; oohpf1f %Fû gҏli~c44Z]ĝ-z[PHm9qI[gT(]`sq{evSqӚ[6w+Y4S>1l!h-Annc;Ι4E$ ,*BqGsE2*O1! |-8\jE!glå"ߖ;t8\+XZ`E4E$NrJ* -f%ګ}rmYrAM$\S!usgI<08+8(4 }ۿfdS]բʇ t,H ƜYj֠,Zt$:RN@NJM:F#h(3σTHt#X]P^Y̕ΙډDk\4Uds3j.>+]HB)"BFL%!Q +#.RX 6 A]5t)AI0fƂ1EA!,Wdsϊ" *$[9t  \q:vt my6d7q&@;syC Ct Z V% Eatm '0XzCU-"@xBhDZbŞ}妿xhb%HHkFhcl&_Cb1w)Qi0#5+sC|m3 †b a0(` UdѦ J,х:bUdPCx*!M)s1Ŏh'p Z+Vنp5TM VnNedRLh|A.*Aᒎ)Xm#hۖ*2!]:|fjPwV`ܐLAAXη.( 1?1 04Ǝ!Gr 0 PPJ ٙHՇ* :, ;lGDx.VZEBP{Se*spJ!\`cRwP,B,x'D)lHGL_(46 qJ `UW~TET㹏u!))"z_R}Bb -Io RH/ }S.SDA e1fe 10 Iu{PM29TUck!z? hDufA˝jo[q@\09Y|\1f PjAN0F!&$2{0g:.xx$BhdQgk#[[1#fch]bQՅfH~F5%wN52@6TB] !Hb"5xgc_q1O q(:̀Kk3`ˈ٥yH3h^A %y]%8(fF5H)"Tev0FupےBΨh `KVsNJ<QyHHzb"ŕPd'm1.cpal#Vq ߈Q0DxQydx݆b1۰q@2S3iU/#=(Z#jw7yg*jVC,*8Qeq@8gQa A1+'JC?tCң,qNU@HhYɜxѡJVWoU5poH"千nCJXM,/#_i,V!VMD9 ,(-\ /ƌB̆ 9%pU2'֞f+JO(TH,yP'M[Ealy(F-1˱ 0kMZ?dCl>.XؓI-˷73B]wT |J;R5?&|q9n>6{Ks\礉 kyPV̧%,(Nǽ5&<:Sk"Ge 9cD7by7'ޛ;?%OqMVTuX|5ڧsFLZW;Iqk5Ũ#F1QG:buĨ#F1QG:buĨ#F1QG:buĨ#F1QG:buĨ#F1QG:buĨ#F1QG:bԙͨc۝ޞ_S;j/y{gɗ.vȻ}l^!Q;x3o7MhOZ\G7YoGl@| ¿C[oo?\x_V7WWO{(Ϸ;>Q? ՙ+]7M㚊Jk(G\C _g gϕ#}xN吏 {cQ+Gaj'/Un-b7^^j^^sN{'^^W/xy+^^W/xy+^^W/xy+^^W/xy+^^W/xy+^^W/rdz28ZݍQj)-ި^Q:IuĨ#F1QG:buĨ#F1QG:buĨ#F1QG:buĨ#F1QG:buĨ#F1QG:buĨ#F1QgFb=ur[ۏQr]ƨi7)ub`Ĩ#F1QG:buĨ#F1QG:buĨ#F1QG:buĨ#F1QG:buĨ#F1QG:buĨ#F1QG:bYQ˝|ُq}˿X.>£Ѫ'_ԏ(il7"n"tQ|E+G2lOA0npr-+V_ S ;#\`ˍ\A׋wmJW'Ş`g)Yn?Lڸx\$)**Kv+vb Xm KUdsp\ vłc?Jdv$շtϮO/Jq5I:ɵj^\MSԹ!\MSrxJpWza \zjx\ʽjG2+c%WO~1u+>VSrM7[Y-٥U:)\_דzR0K0~rӼŦM3morakW\=5ѓ$wT\dzzt\JgW'+mh;\~m_w?? zMn~yy~ QryW |*M^tK*$'\=AZ<{{~Cuj/y ~@. 4; ^m.ndzo7ǯꟾiutX_7cLVyپ}_[yٸv|e#_IW㣞+x;s^52a$9I_w|Tߝ]pYoί#? 0Fԏՙi\SQi=!G79Bb^>F/O'm'.À^mfMVu %F]K{T''ĝ,_ҨaM9uu\Rk1D3ߦ:hXp觢aRXZ|E*xt*[TGb^u Xm\ǴAE\yy R7'?b6NW_˥ XmXxVq>!G lfP^:X% NW#\App Xn^pj$P? zǕg! WO$7q5Mi40\ vvNYWW,w=fbW'+w#\ApP\\Mvj).W NW<}i{7dx.?7߾luY:mdݾަXkն]ޮ_5Ϗ+;m~o}ݳېy-e~L- 9f25Uc1Ǡ';*31ku.\{ϭ֬:-j=fOَ?gc z9|ŏW_/q-vT,/mWw/C_# k\Eo] Og_lzn?Es{'>DB EBW&ٷNku/&u6Yrdm\G">_ɜέ hv8>^"*bV&R[i?9: Ta 9`}orZ|.Nhun=ț/-(Tٌ37ϵHxQbQ~\Z#b5AJalbCFGKWX 2(ka;b"U* SE+Vf|L9VG͙+-lgˣ C 4B9?w}1 v=|{߮UCS:V / m=YU&dgWO徘t%z)F^YWLkDd$/T$ÐCݭܴBUZ7or=äVn\Z CV6(r o-~j?d=]y_6 1Ц]l nWʾ;ݩtk/Yw*l6A3/,M02JvIU}; ]sq<;)ɱ#/ 6O^j\m?W|C# :C&,YF;h l|Un|2ZqT"J8%DTH Sd8Ef8HDro9L)8j@"J񸖘('F1k6&Nͣ*wT$} Ϸ UP0oX[lSʊ ?W kLE|NE[R:qH)9лk)A %i+u%K_v0c\c3\N-L&I5c 3ZI%x3{d!ܶG|_{A9D=:k$6L9,14('aF# pgX`c#u?CtS]MMui;Gj֑v9&<(Vg K]ZqAd?8fA/IC4WG7 _.qcH̗=羐wXįLwK.' , O&ދu4ʏG{ߓuOBO (#?G)(dΣeo2<﹁U9#f#l,B.2iqiTByLҷ4ꂱx_hס]<:.ɀd%ſ''RlR`ҳJAt RqYtm>`8)č?,TFV4/Ń 3fuO~L EOmk:quE'ZbyՉ_w  s0,l~p0Wq,#yMlm^Y??:sͯ?Q?Wӱt9\½i^ ;]-t]5֯q5ӥ6B\+g)RDLcƌƎF:fNYBc4‚SIܱfZwUzNZP.&=BL<\y~(+jsWj8BtOgHSٚ ^\֚ ^ZA=CR.C̐.`l\yU{ʋE[ U:\y)IW$'=[W 0!5pť-pK)tWR n \yqYkKءHI!•bq쒁a"BgR7/ F2TτW/,s"dXYY %E] +90ٶX&ۉ+v{7u+)5FW| \:+Q ch \yqW^Z-@JL:zpEAJ`u{.%%6'ҲgW^C{՝#6 0S+.е!҃_Rr+NVEp)CXDL:\y)K:z8p%UEp΀ʋZ C+/%S\=@ V3+k \yqYkK+>Rn!•bTa"]R4vppJ3h\f#VC+/+QsEq fp+`p+q5H}W[Iypop%:y7|W.L"`y"ZWgg4?YҺ0y|j  .fq EjCXbkS(F; ~|?ܿN<2Q3\>ZY_[Go*BEƅJKjQTZahI⏂}#zp6˛n&ugFznFuΆ&i^ZmL[_+S7D‹P2_erw>,oֽ6nYoG\_Cd?N]yK$δ"O-K }s?Ƽ \a)9@w2b&O, 0n2\q8Xtx9ek^Ĥq4ZkRj=ȒCҞWMJ6'%i H,Rꮮ_wD2tHRY/$cBM,dQzsنhFk on{eY \dԍ` z oX35xȐZH"$fRIVO-Ay[X4 ֽL;_b#Zs" l{m(ZYA\v?ȹ~fB= /'{~+H3d+- f+EGR8b=J ,Aj]/ 1l )ކbkp+[ОqI,@4“ e6)~;ʿ.Ԏfwiq :R۩_K΀+]ervMhhA@ڎnyL5Ldz ,wW5J_PjW-~xP A\.>Oڍ0Zp)WJ~9G'ZCw %L!W$j:`#_`0QXɇEˤd&ϡQM oE:m|_ <1f~9?h\R]]C޼^վM>-m]3Z;30x_dK|o_!tV[?cf:VnX:Ěqʩy"s&\\i`OZ΀ޑO'1q!g4lZ#yӎݢg(m5>SɁ3s|au$)94?狴_o#4!n)͊9 -ڝ=q7{[a[L^x{YQ&7cӴщç`kl%Ζ!ۋZvl-* )`Syӛ=Oι(t \ܝ}f_SQrF ~z@ %ǀjZ9ߏ7"7Bӈ+ MP#& C0SaPDkur-h-ڱ`X*@L4c 4-5F(L eZabI#MQD $FsLh 6,Uos6TD6si&s(%A#u0`t4 ^1T0bCR?U8XrPbIwG4hjy f:2VZ}{~ Y1zFc)5DDt$jYi}(ZEe<`cU7=UdVqa(S+>L!{M~VW#Y0dFug;V "":,ƇXbk{H2 %rTmş ⸷'ڸ73aFzdac!>weQcH&=32V,$*ժM%IK狅bt_/s- } E,h([hX{,wLEޥӹa2])6H!!po`^RۃƄ'Ld팠E/C17ZA!u yH+G } #E/aa[k} E);8qt8P7(8Oq7EnpStLST~%U?A:\YJAG,Jqp}*ZMzXVQS S1F~GlXu83ϣF̳,,oy^Rs۩QӁ (|U2h2 TZHjzDVY&\͒ia'$ A,8r]'ա`0t`}"CZ9 bnp$b$xToz_y5&#`NIw~F/yU(^fR~kA#TL;l/!̿/ȦS;:-荍EϤ׭(gBr~+nL+s82#9 ? d$q:ռ &{6+:`oC-\r P+>[ Ȩ hiiϕnlcA($~G{{57io--j*he*@ jiyWt~N֓\4!e-Cֻ--]؊!Ҽ漰"n\Da6 Jڌ9v|k"ǥ'l8 a{xfs"lWjF%h[s[S ,t qte9餹iS0:60%Gv'-mT9x[X1ʚb=$r5 ;]")N|ZqW'y۩̈́YWy u}mx{ #9mnj; R1^Z? &B/-2uW]5Prlm`n#%܅E/3]f%LF6azepұzCnCEHl7<ջڦ׋F}i!oRX59 B]F [5:|'5e=SV\Bc6+(u3s.Iv ˩5 ]a@Rp$mdRdo`mJD;[U܋cw;*-TfcF-SXls=RWv/*ٰ.`#Saa[k8mu[%Ky-fҠuV?Jيu.X7ޣ+KI.) ّ2 ,רWwO8!t+{c-.n?he`C; b[t 7xbL0˃iξ<5&fxo^1&4Y8>$UBv.=9ň8QbK+XhJTҒZqyiƙD,Kni].rp#{T}x'7i:_0᳻VEq&Kڑ+q}Ֆ};+~IP+(L Z8U /+ҷdt6hnT LL{ ,| "|D|?,W?d4/] *M?+&߳ќ,PT۹:)R>*}w[tj֫$!ɦd,;WoUz2K.":&sa":&sQQj E)R"fN%ԊQFHQBW3 #_g@OKvt/U@WJZjcGQZGob_AtҮ*[f3ܸhj~)I@0t0!چZ!6Trq% ilv}K|bwG0,= /E~u~;.5,cP-ęN'C+MLa ?o1(pʛF1:-Sj`i,E8n1&-^|!3|?$.,&QllBYA-2 ]_^Sd^9Jw^i|} FF؞r3d~HJsbZ%W*H"J9Bh#9O3!R1x c63ސKPxC-Cn*kNx-Վy<&_j>EbAMؖzԑU.?e75iC/Ek'pu@0'>uTc?)Hrv'٠}2yZH{MZ"YGzށw`>K:%Dl&ncRv&rK PG8Bra\#$EI=#I$`uH1JF,DӄqBT4N'I8dԏJhsZ ŀ>@X LmǐK"mi# *ki.r>J?rN qO.?8Tٻmz8G-ߨ4]A%m-9(:e/-{ 5(2Lzf/2)ҏly֕Go{dIfHn1M0n+=`?/ng5IՑqj .kң*K0҉vfޱӧQI CK'Cc1#\Oբsa \SREif,2D $! (%R É0ҍSv*um !)3l.Qf^VPˎ^*9w.,26qe=~qete+Ć I |7?~{6x~*,Ftz@}5hx:(xD`lPFiS8)a$@(%LGT)H*()j==L$tOQ*AARXEGxc}0MR% t q4&.֞ldd9W{ ֲ%^\Ε*909WD3̹Q> è~ Zc\PC-'*ѿD?ce8*'5[ZcR*_F 2Q(r 0GRodƁ)9˸xe!jL:4߈qoVs ?#W|0SCr iXPDH ,?{DZ oU]m@哗K`}SZY^Fܥ5KcH[WW 陇WVE اRpߗz(H7&k+ #@[ k#B(;g-P?Mȹg1A T$}E1Z>x$%78^"I(@ Ɓ*FP}RTѦZD1^:ogeڒ$d(lU+1`FJi#[gC>)au6Ut铻4p`In?~{;Z@puwvX/ЍIxQJ =$_,W+ߝiM=5Lm8{h'~A۝(Sʺ#N,Nם ? x-ՂF#ߊܭKzޫ(L[rv3~Rye06HL2;r.Y }+{@4(Qф,yl 2١ Wo-spi>)A<z 9t{xKEW18q;aGD)bYMYMwf/ &,#2Y $mhmI4ײA1gy;vs p={ʧ,UO['^yLnn LH:3RJpf ,OL >3O!͛WVl((hB-`&RP54J-}q.?+)u@qND}Tlޠ"%7bPJPJqn*z;Վ=;!s?G/sU}_QU+)z:7W BF|jjW4\lv9\"'c*/] Vc7WJfs͕FltU%WM&:vsUD;7hO3nDe|fَlР1&fus}U|[ZHfp=sǧXW?o'_/*ww0N.]yZ?7pɂ!4Ee-R,&))tZMϤ rq`ۃ A0DH(%ig݇ { M4& TFiRK^~E^5'(u׬1_^Mu@c]:4;lOQ9_RTQj@XMw6D7I9}tfFR] vI縥iO)vVcO+F)L FRO\1b2E5sZn*n6Wo\H0>Ԩi KM`T0:ӯڨh->>=B퓢<)zN%Y2j2ka2&L-ѷqn޼}ZXQ!wV~ǨR«5x0*EPARLRdg\O[BIQXZP:g2B* 2VF)E҅:H(t,A#DE @P@r.1r?P/{ZQ /eRE@BewVR HJR!$ +#!5}  #G2Nk 3~M北*ǬZ(&pqzVo"R t;W0Qx 5J²q6R]Dʝ2-)EK ȢH bby] qBfS8ĦS3h}zf:#TeHhС΄0Gf:Q-D@A wO.1 F&Wr5M%o+W*QL#:0` 'c\3cZyM*j>&UM=߾z>Y__wx40w現/`6r\"6?a>pK[/W9֍f5Zo2osY|ݲkfIfj(F< #)jդE1ch.~_E3t#N׮?|%SgqFv6Y,V&i3b{.o.߭-߮Vg/-l#[򯵿htǢlJ1dLfvw zl 4[b^ɪ#{&h>_'k§Jr\/yCBuOk\Uu?Av_vR]UboH3pymس/[==3eݞ%kmlpȕ5=jidF>$n"oB-Km3Õr"g͡~%;{([[s_9)b_"R¾R )+w&Syϗ==_u^LJXjOTɏk溜;#rNy?ˋcf_JkOU]jZM痟-Mm` 6f-,ռX<$( J!,K,.\&nl PWC2)q+dRBp@X62XO9`ЉsBҵpCPmnQlg9vS`'wJtֽ<@TvmwDe( 38+^fEb0njV$B `] -W -Et=X !eTV +e)" R@#E-s<3mu@4?`NWr;I9N[vI:I4 ;:8n{fhO@ \ҁ _3< SScTƬd3p[tn`7 2͖u rm.~EG>)X\;)d$BrFQdeDP Ym0Sj wոJ$wd;w7B1(%τ']wjIal)"ל2DT9"m:8)GP2%3 i źU%OͲFwa'1z~˻9y̻XmE×_6LهbB=LsP6:A!ZOް^\~Z-ݣ4:iϗԞo2Hi~,TVF(e(ї@[S8'LL"6tY=|*@U0OTO/}U~Vg'r2fiZq{yH( Bx[QwovY1vM%zF-N]S]鯐l}үU*zU,G@1k)HrD yA7Mg$ҙ Aw޵m->wy]`:^sPΤ$pN^FQ {iq*r{Tct6o(Uwa*7o7Nf#yjP3vV ٞkףRISO+N98T @멝&81ch*HqROc2lL&Y 28f5nTR'*٭6vjTQK.;$D}/?*V-.njmV𭏬pQˮ4VMNǐ~_rutXɬܞb9a-\g ,OyV,^AUgn_vAſ%( aJ݆3axJzvri[3 iT|Uru҅}<9ZyR%0H;5Qيkz BImMXc\W.%Kktb*f$TPd+2n1:uͩL?5ۧkSèy:Z۸N{Q9pwӵsuk]O(݅)[>2!Ro_LNH?|T0)15 G,Pn쮧yF*˄&[%;RN 档 PS8ֽؑX՛UHPtm BKs-,^Z^U=$ULNzo+k.e>s8(Jڍ v#G~ꊌr^xQ $`keoժgƩ,9LD*h~ IDnZ%C\ISPQO!APM,_: H\s]k?l4CxF/UAW]~P5[ē5K@ٮdZD ]¹9fJD"}`_H *ͫ X~1Y9Yv^M/UIԧp=+ȡ zmON*"_3NuUR80KB/7QHT7*w^DIiv3cx7/JEbNfKi|};W+8`ZK3-@rMv>;VN%|X(LDBxzZ/]D=z@V($̣4Xi-8Cz3Wuݱй5H zGQgqg`z؈J Aa!4eƯ]ƀY>,}-8Ƿ&ޟ y )^G_s/‹I vT }$vwcACf}KQqOZ7H7gyWLy{s|^׀C+{ {{ғ/r$I܀KX,tTfPɇfeU&ou5 EwY$KkWd5`$;siaC20;+|~fUY&E.tO xAhū#'Ӈ*JBJxIU_fZ6j/jcr| 2w?P'>ϧSNdpp*{"GK0^"wJRI?*p$h .T\aPwH3mPqBa"ʀV1-{IZ4AÔ7* AeAw*-#:@BXS4.ѽn4?ۍp.s9Xknr,1E8gRT TIA56Ogv.izO Z~=,}74rkfn/DxaEUc+ǥـe VBWo~tɨѢ'ޔXVYMu >02 s++z9}rFg" )1,3HS+0Iġ-Dez*4}Rw |\8~pr X=ڬ֝C:')iNmgd&bNe0fL!}$3 .' x\Vi(X'ߝPhh[ՎK@9ҖIAyv#h*d Ɨ=VX5[+8B:w *M693Z`p(0}jRR,܏}8-`F/J@yfo,`#daMf4êx]x\Z//FiXuz1gVJe%AXrQT-; ^>Q̆O>(:n<.{&,}e,qlU"V&r*aȴe\Z `,w-r8Uрe.!^ 6k LkS R `|%eqb˨" HXJZOGi < 'y܄TJ#sWOq74h+TK@;,L2 f߂VR+D pY]y_^0r#!\mG'wq 2 ;*R#m 7,r=B: , HF/X@]r9_ƲZOνeocLN[t>+e,η;10(O*@&~GKJ9e?sZ]$ D -LsJ?}|w)"`9՜01h r'TH$V2lw.0gRs~>r'ש} ܽ)=VEtN.w*7|GM3uzdua|^0Oُ*qa \G\V9R8"l dciCl%㰪ٻ9R_Bjt~=OH`#.N `wscjekj9[~T%¹x>Q}*ctu[C BGW9o#";"?V'E{։!/3zI1 O8 er ;R(M,b޻kNV:JVkt;.cI]pbyryr9*Y^hr}:mҌз;l46T4KGC0şpF1 w`l¡k}Gӊ ؾUi"رxsҽS*C]< "Z|JOm(Z4ݣ6Gz}RxX>ѣ#/r$0g$^~qɱX%Zle8ԛ'Փ'-ə"ˮt|r)ҾGŪj>x̦rL 1{_/ /,kʊ6S珿ϷKHIUNexz1% V`OSgj8zuӽ(#'4is_ )i`D c4(9e]h -鮤òF/1bGyo[5x-R i~Hc>__?jsiʿִxkR]vƵO-z4ra-*qfA6>V0ʔڎ-$d_E*ɷK2JBg~>}`Jn9aIT~>ZTx֨"*ej^Yqgvu'o4\_hŹ`$;s}$lCpUW&, Jnuz;0H+gQ-yH]Nz$I &FrY+K,@wrͽx@R~N?WWoM^7q˿b"dŠvt7P=NG|DR%Kv2Eٱ5@cq.CP(w7ž[c6Lşˈ Z&$,8-y݆aJP9o$E6e" @!e v-$,^lRsBH%>G"+H&fC.2e2@_&O1" ?~xRsWBF$3Z>wrp50(7] f>SzB^jZP?Fˀx|*%!#{{e6+[/elYxa2,'9 \u. 5x8MuOXfϲcPz1VU7eUQ`<Sn8{Cײ)4e6*i̞]ПX}FG(GhTđ?Ə# _ A*huf|gtr<ڑ`CC+afn^o|a-GW/us#}0ꤔ#f cQ$"FbO0$k\s?yi<P=iA!y^%m|&jv0@Nr /@sBPַ/d&-YJ|O|9$j~ S.+F\iq.!̝om*S"R49}Hb C < eD!$XAlL „(22 @}`5AOvݽ]gWzrيmPgXy+;Ѵ_ܰCJT0Ũلgvة؆Fvׅ7 DhTZzGll~ <(C}5ߖ1~xH>8WѥvjEcĂH AcŪ4LI#݊3ę'zZ 4raiRr"5dWc^sWIZ-VC:ɫ$ Ͻo+޽!G06 Es26Ж i1jkqvFiv>0mS" Y<5.jrB58$qbJ> ͽ73~DGwP++kn߄NE>] t:#nc)ݤ{uq}#'5ôfhak0ݱHȭw2/-D`\`؛x\u=_Cg㟇q#ž$01h9:C;5g!eU?C2Iʆ;& ;<;ìޛ0DcCtϊhpBGCӽC=p!wKj|9J-B;g?ϙt yйa54yrT"0s_CQ Sx"HRohOQ &qaтeFֲS58c!M;; $c:$!hm2MȎ[zV1!Ey('Pr"ԭ:^<"X4k%̦mֽH)6yo\ z{V3k؃Ά{0$2c$1(CshMk':y?"$%ڲ C՛<9mkg$g-$JɷrpSuޏT|g:by< ɼ|fm \RC8O9r鋰踅H#qPf{fy1g=_3ݨN$P%1'3üxw ƹiuv:u6DctXaUY;\RLjlj%g.4^/8r5QPUeDQבa\{.vK'PM< ۰:\^^+^uH;+Óy?Orkؼz]ߗA@G @Ȩg e.v3'eDžBj~㵋]v_GGI>fN Mρv'Ӻu}Շ7{;6Dhetm𐜢6/KTrWCC@gZx?vH#Ä ?DyOla±:yłq638ܥ[c\$IY4ATtma -?a:w#x0corQàE$:C9H1wm(8"ufra/Eu'8xQA~&s:-W"?{Py7䨤p] DإlFi)͙痐҂bϷޭP=xlݴ=J,M2kٰ8a'lO(  ˙gvh$мA]_`vtq]aGH*!JkR Uma4Ϛ|,0܂8GpĆ>[JH<'Ϛ(yԜz06uwFxkiGH'1y.S{\Rc.s{_DЂB!cG!$.1Ԥps􇟮:KTX] 4RYYfJ4u4y|SӸ(e_~r]˕3TKZ #/ԟ\YJŕfm^bH8 .}U桯{|֋ y%K'iz_aߋſzo3zG93݉'-l1,cwE\S2H9VEV&P'5YsūȊLVSXZ%D󝱣ĄX..=!3Ctobi^[߽̂Y6y c? uTzwe<issmW? ^т!0SFHS6 f1EoղDqg6;͑'8ԍq;v$6h)[ !$p:T@o۹U[խZZNC ?23aU_ EJC%A{aGƐ9}n6"ÙB)!v$pD둄=?8#B]k,z$;M l~͠˛̂U&pTx:+JeZ F@-0KIZCj/ "f2!!e>;0ꉮUs%p~N!IHYd{H^u,s&|b$`Qvs]ĎAIZ~g--sa= ,n6BڙOt_\ٯC'+>P9Ml~.} Dyru*k2m 5׈>ۖ`4qtsH2`D4m?!LXVwajE& [a\؂YR3+AQ̙%`=, !ts8ş(1| cO e!X>=&2iWhohtwJs@&8!!T@( 0^wMk$`L$`|85&ˤ+#4WH8Mu뾰̀eLKvg[<)v]ĉbS #JSeƂ'G)Q%N +_q1,AN*@9$HD0w~gv{M b_X]BR] %N]Kdrmxaqx툸!gK.pl,-Q~ 0>FV,C G-32: hrYYXCS=D8ʉ*٤+υHElջP-XSup, n^&ŜCF`v $RpbVN (tRNۯ- VUn.;Csr iLz58Nu-ÙuQ/&Gc~[FyH>8Ԋ\ϯD,bQ i6pIIBN9\?Ξ 7[7Ck]DoP kh[=$Bǜck9 ΁j'AO"ap 6l>mg*F@CϞ'SL\޵`!HX 01PH2k#y?jYW;(fN: ɸ]`WX0PWԵiA0$o,b < >GhU'U_Iqե8[.m̹bylA ޥnIǏR 0>T FQ(tΨ3h|%6O ⁥|?G⮫=ď2#7?/ިJiA0^}Eō_qNC?t:z:r0Og_yr }JHc1"B  R.0Q}ߗ7=*v呖ľ:З/({uQ5V-+#Հ>(~cNc]c_,價đ^@r?beʲ(k9`rZlEjL8! U+LoǤ&L$l)BB]nG:m2 |KxC )A@Y EmZSj!!/փ,źĵZ9kHL([nqbcbӦpt{VCD-p'5T[r?4GXeayD_ʨ*-{/^jH)o?_Wk4RзWBQ-SW_NbdiOj8>IF_+mES mKyםLYmAYbIvKA\jsi)Z4/2/宇[fz+N'7s> [}[<%Lb׽?/2\a\wBinSc%ZK:Y]C G  e&@ͥT%n&K:3 J ~XT^m;F馧 le2E&a?w|ԸK|rtRy0h Nw]7^~6Njݡg_[Q٭qzWnڭ!p~{ |S Q>KWFAӢLϳhrC .Z k{O9w'6P>|(q;+ቶt0iO+P:1}HKCIoC|SG.1"r$F渚EpОHť1Q"4ugl~Ƥf^1mrR2-K-[Hؽ?dQ~B`3ب@P,;e; /NaG ᅀ( z7+&37"0]hsfəg_sW!z.;h܏10q9z,%GLN"*}Wߠ=pWLa}!_7syRNZVJHgB8?p1wȓ7=̛f~Hۻ{х1{/F'}cҠ{K]H(_{ߥe?``:W7,MzA)j_$2~4h.Pc=RlN7^嶮Ppdu d1,`Ԙ~#aҰAU"w77iT%)R^xH(Msͩ*$CF YèNb|W6%ey*&}9L!Jú!Zs^[iHA1&z*ax?|vpG `4让RC]ln(UjIbVWy^LUox%q5oN(XN,x ?ƂǏcౘNyt jΉfpGr,yc,V4x/3- NO'+V:Q]&.mzxOy9~~Wl@PVn08Iгs~N-8dW=M&<f%2)mQ,tum^hߘ5e֟)..f .L4-xYt̺wV9x)k>AJ1K(q̕Hb\ vuBƓ~`Gaƚi1e3sN~CTdOޗE-u)-zF[P Qq0՜ltq$v)rJh4xZ0NXN-zojI0uhQG{4C8R,V$96ӖnF[v鎗 (Û,SޏCwmylE)ub8͙%+ ;4',AP9zY3O)LTGⓩ)Kԕzf. }| 60`ⵄ)׌O`вMҭ4g:֏k|< r^6SP:M;܊%FL=H(XϔkGs+ 0,6s;dL WEx[^T st0QL nr .hW4aF`2V[NpÍ.K'E$SUtC{mt&vv7(9"PBBTBฃK^#'b^I ЎޥT '1\):cѧ %=M~-4>$}roLRу</he;tEKSAQڇzist rVesL,p8VQ63B;JWD_WRDtҼݫ]bj$TgfDk,(܁^hFkߣvbUL9]*'ow^OI:f8YPTNqE/rֿG R]% P[E xnpJۊB 妀yfDFfM#/n*]@EpȘTKrڬ| aNm 7} (ҡ-02T#fdp cds7 2-%2J/dͽR)őx>"z b^E=p+bmL?;97U .%2\2]q2URi4{ 459L, 3AD\( Zow\LCONMen7S*P)uU߳ZORRv6b{t+XݣE>M'I8vm'‘.td/1woT߆*UYbGqD>ݪ*rQC)԰ٝ@o((R8D̅[nτq<Rŋ|3)S*@ i I kTzIE_z']1;%%BN"hLMKJ rV]||NyZy>/#}^h(%^"-5!v4j$ۡDE^ ^,Jze<3B_2D~\3>hBf'ìAO<"Jtf+ST.k7+yɰ3kƎXN|p3R2'1S2)DijyU\h1ԃ&7AͶx!cl? ,\t sG2" _UÓ''琧KPqRrQMe飒=O I%2N:V ܖ8|f 6+6| >1!}f@HWOۜT4PB|EpN vV L*-4}cR.L? tT$͕k):g 8gT-~ނ,Q腇Cd$8Ͽ.49*9o\'N,.z_kP$¤(a#0bB0?".44zq+@oמx8x|7~+B<*."Uf b1[˥w>85M;XqKhgF0}C1Y<7 |ta ˢIcMBc\2| Ba3`t3Bu@s&i" @t&ߩ K!9~:R;$}@ hʤeܙ0[1`R #G¬4r%,^:Y[;p0lY6tR>\b,f ,L4-:a:3@l#r3gDۜMZ;D~CrT %:"M<<:b#$*#uG)7emxg~fj/LnNO҆}wg*JvtMK / mpfHX~i˃phӂp[5g& ,"T-GЙl/pH@=8;QAyZX-9 ʩzFņA[wM`/4-@O3ﵿŧSk!m)*]z=L[13gȱU5L-AH@T9th; cyd,m)Q,M1#M)#g~="rtt<HWv<3w$%p[pi:3'GmKJ0N9Y ^?4à It59a"u}%Ӥ:M[.avU0AC1Q f8h=Xe&Bp3-z; +w 6ҙmL'Q6/бP֜ϣ$v&&r[[{lhp҂uJr`G%<92Dsr%<;>3F çk3%lz+Pw%5-}s9Mq#LdZâ0+qmM&Zkv MVf4.j)jd+Ց823RfDƷi+0b&?yЮ6GBfJrO^n/sJMB(]3<1`Hv6ZptG~P#-e qL*IJ o cfrM7QD);8,IOB%2N *Jxi.%W"-8!mt[i Rߧ?,y:RHvE՞~/hK $J|%N4<4jBI0:6NhCB9SZf,*<mOe9Ɵz=UZPjt3k*-C 9̝f\0%@gq6=/he;][S#ɱ+ SCݲ.D9G8|s+0I wg[#P-u F݇ʬ/Ąo^B9~;'')יڢ6V=GZh;]/{xls}Jqo;e$k`Ȭ''oy~PDr֑e5[ r6[&'8C5'A@Ӽ%ߦ7^OZ]iwiϞ [C\RxCo,F@8.*!ܥm@j3KV4=>KRum#E.쭦UbrJ)0ު{%h\7)bhD ^hZ7ЇP9yܭn *JW4~T',Ȯǣ}ː}9jl߰[}CEo:W>-I&Sw!P|k@v:UD̲f@ Th+2 v(Bju!0(ztbiٞci'{iw2얜rSEaNtxf]UbR&9Ei 4f[xkN/tls.1E^A\8 -.sgyZ>v(}mb|SI>-KQh`>*(}T*{:y>m=NQjoMB^nMpCdPV鮣MC :Jv쌕D];r:I];yQqhn% boX b6GGKoZ%.O>:c$ח__@PnHp3 Y7nd v#dI`v~ @0NFaBR|LX&ėπUT(SP¾a۾%(E%NCЏ#T1ꝍad[iӵ}:se!g[Um!MhC IF/ׅCj³Jt.L2QLU&:Լ}MI+CSդX@uk7i!)oPڡ*JwC4t\X-~llqWסVEj[u0}[,J=t몓l{:+-%g(/޹6L 誽G,jiVm678=81[4(QDn&hԾoNhTW}یfqWqź#FSTPfnpW|Pz6|r0NUk7q*Z}*zEJy0z#7!IelB&RqL#rzƳQ6#;Fo-?ǣQUValC5}'JM mlĉRh+WDrCJ 8cKjNh{J,pF%iԷerfx$M`Y ~G:%48ZZGV\PVS&ÓC%c!] ŗWdgA+*K.E Q 6Yy'Th*PEckPǐg|ފ bhAb80dDl4ܔp؂D j`FdRĄG~ BhxB8,;jsO.rT7xqɡ 4,N-\yo\L{]M?uWKW+7/p?0ʎ޸̤ XyHd+Dv.h|C[AU &"ajr[pIuGSzppϹW-7?:i:;Vz[ߝGy~}fwaUqI_.ffxEfӃ}FqcOc ~Y(,\sP/IWS=D35T͹^vpmƻ I_N[Z F2$>5MNQۘh֐'D_Yc3 ~GkfֈD=HrH! N"=✧x 1 x]gDX)yfh*xooM{]p#>+e&AUmN\R|ukŅ/7Ņ;Jtvp&@:VXt+jǮij<\OCy菓VO5 !=%E&EuLV4:+2&!*C eyaHPRF%OaBw뜺*mo(^X8-*xGNP!:\VPdVV)7 џd9BdR5.{yBcSI_U#_-D)9U$t "*q< )hTBfJ(+8mu­{ x@*D"5z*n Y\R^+\/ʙ@Mr;̓ǗٜMTEC#&dj͒`:C+ɛD2s/CUUFC'@Xq9 C ,W{ֻC0T;+*jK`a[a#6}y]h $RDޛN]h]XtlaccW[CDPߙ/>Usz.3d |=;]P&&`'#z19YPF@;orhP d!s.3T1`5qݚ 1Ef^o8[v*g)tk7Jer. wj,gz ː"{WDZc}*>*R7ɾ,]a)+ɶ$Ǿ~Y#4g#ǖbbUv.UZEcZ/946ѵgֹ ]}\]j`uNg/yŠT%gO I6KW'a^*䴿d̽j)Qbew{!C4g[ Ҕ. ҵ6+e!)ιPFNi"6G!Csղ΃@hٴPLh34A7cVSRqΉY>̋7%0y.Ș.V!Hƹ4;R N\VepmeuV[hE&NY4+,ʤ5k}I# 4-4V !`iMKUʜPإn8S5oL#?j:ֻdS5gX z{'$,)%צdM~VT[7sy)%3dF?ʹީ:ݎ8?G3qgG1]T@^4\ ^>XX}m_}_{?wd^ b s4*?NNŴ_<|R;~y1|Tfu%kl#\|1R\ѲA>?yB DjdYCP %Mg,2ks6*g~o< pj=gc>ЗGQrPߖ\οs>ӷ_IvA "*4Lx7;]:͝rDXǰ7]zm+~f *E.]&2h߼14q4B]loy Ne.MT[v2=W6A;fΉTA8Xݰ}Oh'ܳv0lG&h7A&c98\?.ΏUdvS1ƇX{{Z ~VN o N>Mv>fKE[/>4~X.;4YY4@Mi&D(C>[k2PA쳩@K6 C0z%s=V( )!?gkH@( :Pr a"29[[Z6'v7ClwFjZ tk#n}ȍ`'E'WK^Jx57uqX)%x*yYRI6kyxsO{f,QL)ت2b&VA,klD<=r GvY'rTe@(?(Dv"˲fU7tlыiOW=GХВ"!ʔ,ZpUcAL 1'qv>U U{B;Z,մ)QbI,f˴wa1?P젿ys!OL@2 n;Aܓ^U_ɣ#OǫLCU[9NwָnΌ+oq|uxZ󩛷<7hǤ Co)9٥WXķ[%RH:Rj3E -g h6[ׂ갡0|>ۻR_" M?zՀeAG/@TOmen%uɲ-RK))<2<]kJ"^c=mLOMy)3diK/ 9?rѐ!4ÔUȻ4AgLl|κF. ljH16P=D+#VCmC|4O]0Zch{gnM\?skꊎ^FBHqiu!0Ez05EjENH[.l7P l M(Lfz&$ͺ6eqi֏vGx 9Ԏ ϲԓue`Oֻ/:շiѺ lѺ[kjͭk]yoZ>jۚŚIΠDO vTs hK ;B`9E!<6hӬ  t`SؘZ zS!\Q2r~qq @(絭-h"6QPUԘVAFׂX\KE ӽCEd*'\ DE\qm;|>/4ں:~.+FRɂ`;i5QGl׭6b]I˟*KTEY2QBMZle$; pvp|]Al¯LQQE9Vb#Y_I0Kp:O(T]4Z2&_b !6L\̋ՋE}YM'g0,nD~u?hT5}X70jmBCBj G!+P4a$&Q<8PfYx_ٍPb?(:d93{3ɤfفz-55{+ȁ{ZÑ!x7Ӿ|Z/{qʧ],R(P^BcbXOiC t'.YR%q!OYccjŝ )Hq#%s{>?8+sXؼꭓTdFž| Ѝ٠7%iZo>"Y D^wBiLZz365mlX=ͭ▆<?Veǻ[9(\&>V1{aҒ١ˎsy +D6Q c_,K"VPob/Iٖެ&)gh&FtmbAv#GqBM=+~Y4,$DeQe Ez2Qu˕6bBlȬ޼= -Z^L>|^>H5Vr"aRO-Rr 505Sm6WrHcx$bgui˦ J8"-|ެjq. D@%*3 {jyYV;]&K@Xccm qjefpLc7v󁐧cl+\nv{$)6݆;G[ Q}{2m y2Q|4  B:hb\N1,Pϲq hqЄTaJk(ڲ 訅Zy` %",}^,DmoVsLcRjtaFn>7!K'zX6?/nu=ya]f[L%;*04JFqO?@kzFMٻX.T`.)b6M޳PepHa]_Z-,Y2dRISP:(YTUTj5M^xS0FDaW=peDX}(L<;cSAO+sө(UFҤ*:@>D* )'a,ud߹COkjF+w8Rőݝʦߏ?(Zޫ;=g_.rzY<7W.r}O^'xy\6>޺Gyň'CFvʐwޓu#ā[zК.u~]Q.Y KޮA\~qݯW7`xyy1;_Gfu%F>y#y27BoOީo9\91D.|p8X44o7NH]MUw5i-$XUH$2ȥ8!Qh"T܊pq*N Q!IUI%^|/5Ff忱؈ی[=K6n郊mn6l|{w7]]1_v&E/l'ּSkAA d߽.,εI],"]0[;>hv,})xb"x!!/FtwqJE3Gzҽ'/otI-M9R3̍ QW_* Dv [>O*[?]e nGFT:{|)L@k] T*QMs\()r )LJ cEg|!,l=B/!$/{= :8"Tw}r(=R}Ħ*S -^EZJ=d`ݙN4:s &/9.XDb8xUUCSҷaLlSG_ʒ!?y#榤tVYd_o х3گy w_MMM!_y>8pe݈!`4>~\2xFG_p+Z71y@ @:UD@1*䀘F uhz%z [D*%b7ӟ,1>K-St+T8Eǥ$Q!IԫP颒sJH,xA=jx-9F5W.Ceš$4]dqЁE4JVZRy6IZE$i(:2Ȳso!ݗG'Fz]i'VO;I}(d/r )m'Ê[)AQ@c1E_muo;*ᄽ/(fɥZd'>|vvDԎd_6yfSy>GEx)+{zv=aWI17[gaF*P:G9oSmT7_ur`v$|]Rt^9s;ބ8/1o ߼qhI.qD>>;}X-5fNw?*3pGvN jI {C/:$O=yv`—¯t:]Վ#$lU ͜_jھ ~ph[ ڑ[ qqv+;l{B@=4sfn"VhB ѹN1nMO#}jbOWd"Ö/xhEvφ"춭jDþ^nEvmGv:p/4llqv+;lFtY]_ CE0B+{hJWHl$XnʾŰUi|qSyxׄ{[NUFbjXKU1/Jzgb%ebkR# .sdk4XBqK$. F<.d6/<9ȴNm]0We7Qv)?dM9=Z8qO 86hޝO6gF}]씙)(NIAϵeP]ij̣@zp3()o[QPM> zO|qO@bFb*1_hxPF/M]5lg#.VLN+@qf7Z%^Zϋ gm }_UM2&['[S\rr%@.P+6ʨ(I2zlRUFLvNV UiN6RfW!!˂&jBejuA%HQSY[xY_BjK'ǭe&|4-%xY_'s} >q]gOG9ï{xeX(^098:+^\ȽBW3b3 ۶a?YyRfeЂan`+!Dk-xVwqAa"WӈqXٍ-J̭I.U,N6C]pQ{~Bcv&[h~9;pWTyz"&}18qsGU(gJ{:&(09ϸ=iqNxovCb~ƹpPHDbh&GE&܈0ɳ2gJ؛ET(4bZjw{C׮\R&\еk5ɩyUQIUj" GQQsbH)**nPY7ad_VI'盞c4~ yns:eS {(WŢܫ6TFl1(NTD$Eo>ꗯl^F-Zp*TN VLZfھ ]@~2b-m%G#vBLިno=_[>qѢ}&7SZs> j>& Xþ.dvŔYiXmVfZҜ}ԡ +p#)t]kKFNVPO+_juDN4o'h,H|jqa؉(f~̗%$L }Tة6.r!Eɒ&lG-%aZc׃BPI '67]Il\ZWͫ ЙWCmYR1ts;toyoR5hHYm K)Pm/Bl.R8/C2~C ٕPHc t UGKḘ׭31eKjRW B82}"3.޿Ko?(yW()FT|w֕LS\I,#(.s`WJ!tl-XZujT ŌQ MFk^r,KΝLw-mWEȬTq#fvlaYؾ{ Hh)R* #\8ZQ\'H@t=*㶯 lȃʹڠ(qeVgR0jX'sӶ,a~w'c&Ogs\֍k|ڪg2~$N sy7tߑ6rUk>]!Fi.B !ኇT!v.Hm #ﶸ53TЅZTJ^ sxDpY(f!O2yL$M5P^Շ\?61#VQQD!Bԏ7u4MbqS\aIbS2g˄$&_SҺMI @*hд>]Ӣeh;R]@ )h@ {%ĴS>譫i|@6 jD2byS-C齄dcna8`*. E ';%!.CfJ=}u,5[^%D"k6JM;fjh,'4JK/aA Xf{}K̑eMulhR`ka?% 4$mʈ˜ 0bfds0&`D]nU'"*~NKl=н޵=!G3?7y:I><f83/ 9ӧaff($Ȼr"f9LS~c/eO Oe3K7y:!xƣ@x4~8h{Ԩt[=PvF@Cק@4QEUgh'^Egm]5k/}Y^ غ}sGq|.}^V{۸8A2b|OFB5G3mF)VO'lBWֻ UpŧC~C#ҩHo9Q;r!f0/kvϏ#/~mF>u~]<pwcXňs"/+{F| |([~L%^"糩1[͡ %S'iBoWfJybE9>x:AJ^^:lj~jQL"Gl\md=2\] `6l.Vm7]!;2 s8PN f2կ`9KЧVEM':e"F~Zs>K^B5ac_;kH$#H'-x x{B-jT^dF߇t)ӏܭ睳i75i`-8 9 cMbG1z)کĢRYjƞRFc+a_|gJO7{KpTCuןGkFmXI\CŖ 4FbJ`G9k#4Ljs7/̎ \v)YsWY}sy|"@].t$1M Oi;F:F]-C*S[0.[ ГĭFlIịeeؐ+]H.S0'g*ެZ2Isea-y$p(¢2t6LhEG' mXMK\Z)EQK%)M5IֵQіH6J;+R|LEEʻcv\ߤ@kdb0hZ2G\aE za!BԔ4Ɛ)5[J}+\cKw mx}Jr]ɞ[,䛱g=0ך3ބ}O{dlm+-lA2U|Pr֩Yr> >~^֙4tS$(*܎1[.fDCc5TMK&`A5M<+: 'Alj}My#.xX>qy#旭 eA Pq (YPr qMneG=dߕU/ dS :VF 3-=ŜBo\ؿoﴹVsEߎ=N-x.F `rR@+5?4;V0nSr5z޼6XhldKJ.L YFacHX o\.[9,ؚATAҙ9YbRN0񘉭EFT19H˭bLnق-ګ>[|^4ZR4{o3\->w9*PwX7JwlCR &\gzWLJDi-Y*N1c-׹Y`3LFT fߣx'gxH{34jxdS>C=yN6~TY]wJ{y4츌bqpg˦ydNjrByĥSƘGKDlv9=rJܲL38mi^{LNFZ.,$CDjp,l,>jw7yp|fel~#ϸThĈA#,xV?]f\q?bF.&o夽;!//x2w;8,'ѥ!dyF΃q5#ɠUCkS`SWI, JAa͠0obobbǝ㱝(O&v:9";5v^q?wձuAG[W:Mq٨o9.̸%^DZ` ^~_N{ݳJ'Ɩ}Fuw]gl=^ǁtco_6y21ugB]$D`O^!~ '9cUNt5D3&O%0υG'ԕgT,)/ͺ\H0ׂ;S@a@}$\rahdb-OsPNUYRxOtx'ml,4]riV[l8`nel>`G^!T€ {"d; ?aJJ}Nob'gEhv\,d BY[Tf槯.{tj_жzܩ; tEDYdJfߩ+BVigdb{3٪CF9uꝗq^օlo}ATd˲ a YZ!2j}̱g`$06pRzg[TrDkmAa4*k0z 0^ ʹ<ΕRnMbF{]\%GH()RסhUޓ%ZI0t#<9saiT9;PHו&0Gn&DGO6Qo)xtxZ)QĪ\vYmFrN߶,[ExM ?򉥍_c28KYGGo~ohh^qN^3B7[?=Dn>~S3"38\Ms`뜮Uʇeq-Q̺\#sie|^3B22>.S(9 J(sJqpmR:63>.2UR n(ЎQ.P%5Q `m9T0rWAePQj[5yRduGvz.]L9}!ltB*r].]vxO 0bVsߓ\"h$:32`cQތI.ؓQ܈H^mfHd4vARIxW~1%@:_g3U{ggzY$SQwqv|<7-u=n}<Ɨ6Jty|7o|%W~PDAj4|HF+ӻq"Z?6>.\B*q3M.0MPd@*"Nva,Hx"(T<꼽r()׸ wFlD&+md|yYV}I[hu c&TD ŵEKMpJS.Mc bUjjN1 Y'Yc^ o,du2Y̪q\;R﷣=#M\obu|ށvRc] \]9io~nQ}ZC\xW@\Wl윁"\ͤҚ]PPY_G(%Vǫr~+.(3-f;"nM&ߞVY",""EwTSȔT`PeA4aIFLHZF~,*JHяYAqZ nH.Jw n 4BT$D$[g2%-դeno1b-rxs8ib#k7 pTW|3V }ɇ7>CYw O¯3~ 5A|e[`L- F`+,S<9T̝m2ViU9aAHM)+Cd$ pMDG YbZ1F:u(*dl0S?p(b (˱?M(`~siuވ8n#hA: GW< [Tz#bLJJ՜$eUVskvU74)A;E;RT!a}zȎz%OoQ+Ǝف.A#5-6SJAs ٝ k80Fe84Y7n|ZۣSԶwwjkx%ͷUSv,;%7V5ެfF'*‡vMܮû/v²s ћȢ/mt?Y6b}vqV~mn/[lAKV𞵆ʽ* \evp<`',&sRw1 wiM% XACMe$0(Oou] VLV .N-kh7^ $]q8'ܥApcY*(Y")|3ə  "zmC]<\7_~x3Jn Wm(xr|gxg}wLr)Md 0RaL T,PTpr\kϏή JyG~@FJDNQ, tyX $u]CHys)S\kCDk bjHG9|m٘%2 D]A'D62>EBD&6-(wǹ*h+>Ud*]qZQ' @ӇgoYM럣%k'-'xXnce7ZдYjuZԑRÁ;d Ck4L Od*IGڒ4hE37X@[!n6D^0iHi-':Rm".mMgL QhL>}sae Q)q4s>^Ie]e"]׷%l oؙfg=]LŒ>+seozJ#Laģ0g1&;rvQdgV1_\ϵo/hA὏vs+osaMd!˯29/Qo93p]Dz^]^ˮC&Q{vY\1d)/ǖRf#H}6v] ]&Q߇V{^ |L)VɋX7?ަ2 2|Vy z_onчex/䞞l-[3 n=772yWZX9 ќVGBk%UKaQ &-j,4%?Z/p7 y˳fX҅'lT?}p^6iZQSiRB`O/kyk_Ƴže@]u N<\4Mot< =&)K5p]|{ic6-~M?É>א7;uYr8DmSek@㪁+f.vӍ?|]Ϊ%P% opm!}sB=_v ZCjxk֪pzxpjkNaI>P:?{_e'T9d+[Q`}R0Xn {n)_[ (pjhG)a/YCK }\2b2_Rc?i?W&>-LAnfQAncg/(DwոٖbRpBZAk#{oÚMx)Ft6k/2n }$[?[9qHef(wlccLa!Foȴ}4%*vZ:QkxL߂!`tr|L2JKh^գF7JCGZS[ "fXO (Fnb{nUF ơ{ѻM 0=G|+vY=YHu> RroS{{DixO*焖W`( IK|p/߃˴z, 7 d0mLaEf1 6|c[BwN/7&_/T˺8$}ZD5 3nZ"ղi:a|O9=< j\RK`~!;rV<M)HƮa=~Ïq#`룍u}m`{WpzQ*zD!u~h*Jǻu"mTJ '*I 㔴'YʛcRO0C1*x@WHLʼnod\25>tiAqi"vAcW٧NcΕRZçթS' сū N|=vCif7'6 ,cJ0cA(랯Va%5o_b]uՇ +vŒ݌S} 岪@D]ˁ'HIεm+I))ZxS YOXtTgZx.iyQC4IcZ3EvןKŕSfpYMі|)?&ӫ$ЫY@d)ut6êO }_vpK3_.-ڷw0<^5 }Wտ~Lr讦#S̺w "Yal8BҌ2+ɘBxab |wf7Wim[hu TƔ0WFtb\SjmEoFSkD`爻ĝFOnh{æJVe %@6E[٢bhT2c¡:[;|Zm\sxrl8#>*d*jqL&(skxfzZ!lqRj]Paa@1B 4a> ZoʗH_SSWY4_?f$50uO?igO9-._̃Sw]8=Oo{E'\o!J% |=Oi׺O.Xu@]J6bnOÃc wJ8wn@B^WD67NbpohYyNϕR<7} gU,"MAM<`> gQ ˌtEPª-kۑ"Rpk.mRA#FLK|04Z4k7s2cn4ҍڽί2p;eL.$=iW-!"U:6ȮH;m4SrF40S b iox})Ueb)bT<xtQ8c3Kq ƽx2m}I" 3RHY)0ȣFwFK {)Ȳ f9Iu08P4FL7Rao4k3{5&J)vFE K:Lsyn }/~^s]7Ʀj/y(s jU0eoԙw\6!$LGmU)G&0Q)Ey`abŖ9g&cV2*MZ5||ߺќQRՋv `23 Oy OX2̰E2T'TR3,2 +=b{}n&X*#H@BB$ єfR{pI@RI빫)~EWnNeH5'ym Bz`ݦCp˼#1ǎbXLG o9r J#&/"[}p`/C ̵P ~Vھ jJp=/axc,bʂx'"f6m#[,#q XaaE-rH@'Bl-$c"Q]_Mxiye,/*'ݩ yA׎P U`\Ԋ&(CoGc\kS+SCX9xSۢ%t %0 Lsr|vYw~/8>K ~3[f ")E̓ &Hi>:5&ppXq0T+tֻ버BQk=I[ƿt_^9g ҿGɀfE9fd$7&ԙ }_xV/._!a/W)!bSΨyeƴ2PS_P'?9J8{0ނ3~Yϩs"H=.=7pCc;xyDȔB09`\smālAi6orΊ0yq}P:wbLi)$f^xh"4$9iѻ{W}/X"R)"u"R)"uQHhg8kV2GHLrPT iDCt,uTK>|olp[ۧƛԫ&jwNh[dc{1/t t׋{!^'5&& ~z}t/j'%w/[GWCGlu5RgCڸM38_Xؓx@G>Aj_ qHU>tv~\ Rև/28+{<*>7'wޤxbW4)eX 4%#jEJvmtT! U25in:,SZ3>f~L-t9}JH6pdt ke2h˄g8AeNFJ _ڼA<3?Fxח/V:h2s:VI*2#v`ph Ls2}/~9ͥa"5vJe)}õcC70'8@Ķ'PmVcSBWԚ{gA15֨9GFn5pIɝ/ʜ5pUjsw4zeu/.oN Yd.ޘ'-onj5*'j"J,bpFcdRqĭ(5V(xJz !g`㟢$H7^mSk9YMͿo|4\˟U^23| G(v99WuȧLiyn jYajɱKW؏E,X3ά%(Ajil 7VJߊ1 ǃ;!5rùi4IJAJWb&FQ( L2RR 5%d 6Тr֡+܇5fY{uxNAMޓXe2 %"Y5VE$ʃÞj \e9NrlEPi00?ZR4Zvw 9.i(#c`W2큠8nD^YA e*m"(HV*'8"#ǒ5e:R/3/1u63J)[YJ GFS\)E VL8\@1*,l@1AGJlh5rBuI&a~ ]n>_W%޵Dkl7iV*q_Ȗ}mVĩ;p)'>N`Eu#+k`S^{`*NVR<H 3,N0QzwӚfV@"!zyeTݝR4\BFG"f&: q1K(C8a.tJP1m^e-#`!Ec2ˍ̘ԥ+>.XqRғi@<|jvq3QB:xThóR]vwԬ%FJk1 H\PpY p=" ^D`JY0ȨP67Xd[` +ђaMj8Ue}SeeĘrJyN9d(#ˎ.O[[dݒlk1;X& HVUuMy4i*MJ1+~3WTZWL)N;-08-ܛL^{ݢAk+B7(x"] ]f*37?rh%Q2FGe' yI"Xk'\z ==GDկO5x.VV%QصiUBZФV%@Z4@[qX$[OZZ f pe؛} zz9Ńw$.rIe32V0)]LQ&58 Ofz1/Bh)Sxu}7tA<Њ2x/0!2CUsjYA;yCćes{zdHY$LEd<0~<~dbkB]Hҧ+u*HP~e܀qAb5!B;)a v&ج(6 -jtIbhs#TCVOJŦ-c< ft Q X+嶚K] wՖWRt&PJ6 .v7M19̣—7t9\RP`|3-)]/;O븷'@Ş`}g[ǵ; o/vdHk0F? ]Gz^;mQfs3fL!IVVnÊJe~5kXkߙ*C(ǓV[CVb (zN]sn=8qM9Xοq2bKshS1212E2WO~o?]NΎ>]m(+c {: a3Ntka uưXAo?O6otftOVۊAu@7 [l_諃ތ?M>@w?ܿ#<3I1#}ʲէLyiQoczcz{x=neS{gmI}Q7dX1{c\\`T|uRLåT]6 u>Wb3M—6SY<'Sמ8u{KU>b*l)ĝ!0+.*11.)Ręè;(Akb/,"n@@@ YmwskM3"QoKqmbW1i1h8. d8am\dRT?("˃_h2YUí-RsRKkh2Wg#ۯ$xZ J@]5|tDV\h'ln/P{8soAjeLP #<1J o"xpa!˅2;lPpI(2Vۥv~VkNu8Lō?_[<=p^|0xe E5wpۏ>/x f=|än^DD [@<k3zq<M޽^!bŸBwf6J~zU&Dহ"1 o^u5IΪן߹q+K>Lw^ D6P1A * a RW@J8ca. z:C!TxMSLp;U) ^Shpe*]N +`u y)B NWe(u WbkL$f@Qt`G@ ofn,y,2RC{(b/<\1Ivp,p?#rUԶ8G,L3 yfIؠwY43xq&iuf@ 0;0'xO)\T#fxmSWsd8Sj Ds * DŽ΅T!S,eFH _E>\f컳,TݜIdv0AhQYtbA U<1@LXm4liK؎ ڲ]P6NF|&AH|~ݪ"m24.H&U}*`5NfOˌ0MycVu>q؁TF4͵IUӡ v+#5*׭4DuΔ@zKZ6u^.!m5$suen<-kՉ5Cz1(3x7:aFn$j d.3 1$ueK-M]aѦ ;eחbm6BDs'UCky ya$6{"MI Z;Y꠆)ڳ\uT!ׂ,ZҒ d?,iDk,TBh岛p*!4@lk \[Z!]w=qMXS'WrEU_KpSxӤ]zC.ՉW;cP‘S39Q#G :Ck,<p)5u00Pp/wA:;gz[/΂+ͬ;~Xg&tZga?uR"A)9gHyڌa38U<10L"'& ex,C#?sH LhH{é7tY+;1)2sD)2zjM"6M^%Nfx[XdRΞ.LڛZ#CB;7܇CY^G5g[*w6 V(4?WwvU H751O}ZU ZDݎbcj90&}VVe4l`J) ;Wtf(OešfkuiaӯuP6%ex,"C-uHD8 sy)\}:DDwPw,VqQ vտy; cWkWWF,#PP]ݕhJqo@\ ck=jhjfr;`21CvLRHKT%w1a|#Qr%yUlRs1,X$ !=AL=`#1C[7B(&ns#%w6OcsEűaeYC"N܊+&X[t[ />^۸\zx[}Ÿ_,x#& B1iʮ{qfpRBMn|bڶHגz~3!sq4:Ј^'IŕμYEw BcSވE cJ* ޕ,"ad`J/xYfp1 fs"K %N[Mj6IYvXEvWKWj\uAQ*OӵHń*IlzXzjBjuFm?j# ?>E:@lb-X {{_G&Ur7?Β|$6ǹMqM1}G+ZN|.7e3$ < 3Ϥè|36qzD cO+Z#b#f[ 㑅lكi'6ӓε;pcEYMr~gJ̩ɰ5vQ1r;W2:wL"3C#mF5ת4]Oʸ3GeA٤Q5XVL6Tkʎ YEY}E[5U)W$g$0a (ę6{=$~,g9$B[h4A#sƑ7̮%0;?^c[JΚA b٠:3Ӌ6hlul)EWNx}^M#ciL2百$.a2nLtj~>OL[AkZ_U~ܝ^&Aq hs\Ù1]4`F')hZ}h+G*^>( mԊLyϭ Vp3X߮q]bOI`gkO:'M2G c/|j s MqV R:Q\ . (A¾f ]02: RIg.bNgYyt`݌w!s lԱ6̟QR*Nkl f Bm֨P(IV׺D/态f FXfRfӍT8t}t5xZC zʅjii%'X\mv8]VS ,1Kq%r1aZHǥ4p\2" $1ŬAq U\y`0_ ;mǔOv}7x 4Y¿>”L쯷08iG_FWxK\d]3ѕ}zrqoo|;m.SԽvC=6OPw黉\~"5tQod6&(ϵ,5.mT󑳽kXҞ ǓFٳ0⫨S3m\wƋ+d͠y_z7+@oefh@ og/_~^kX/N&}?{'W`7AoqmdN̒F-]thOâ4eW7iYjg8}7DC85/X~']$_ xb)OvbQ&Zw`QavČoGftŲޗljJwPVVs4WoxEkax %q4ĊJ']̇"0Cn0 `K| ӍKy;Հ$.E6 'yM5s4§/_ma! quao9aHAHhC``) #΂I.KШ.By6HIeE~7a!SJr./ 2j ًPZN'MuXc9Bŕ)_SZ4,7TñP2OHގ8'=qWªbP "Bza;fwxu/AT)䁪4C&}ylFAc!b}6&8 "TA!#h| HBև ќ$ M/>o"ZZ /yn֣hx΅5#G1?(bc h+"E&cK%â0 EpwO&,.{&, Kh%/#GgoK\E ,BGI9Lqu@Ay 2GPp֜&.a:x\2xYFFAVKq-(*XypɆK% ldk>oSi(~Bo* ȯFs;µ˘$t~6i=Li0o8z)g(WgWFt 5wJ@yf$68 k o>)hFo}1gA/,N 8-`Jzt҈LޜzT0Jte7HuЄLGN8<ᄅD =*=m\ \ߪ&cin\w-#gm>` ڏ=\8 1$@P|6aUNZmd,9K l8A5\2( ld%:rsspɆK6\7d Dc\@A0b$?~O5?,. 3^JYzKw`>Kb }N |zXSkg5F9mO#2 Y -X19db턂y (|/ ɑ;R#xĻǾQ"/4[Sb-JvsJv&0=`=|+?r @hNA =DV(u/JօJΕAx]rvR~}xF>$ R9U|+ZMaw(p [ڀˠb`҇d{ۧܦV`: M [LT4SI D47y{nyD47^JwK0RiUP;w]Ge!D1%릟:=rΥ-.byw؄Jݝ$#P<dR"c|gf5RvRLTI1&enF;rE5iFk]۷M ʈfq9SQeψ y ֶ_2ݍRpedr,kKPSuar`o9ZV>G ",r|] )E F:Ҿ >uAD%&0&̚@ۃ0*wj u1O d'j0?ZHpj ahu9S7yv+krh{Yi)ƶ. 3T.nF &Zg_TTkW ϥ߶ݟ}ps3_CcJ+t^B!R1V}8 @FJZ} 1e usJ cxDĖ( !B&Ha cXB%Q3w$/GUFG ny XhѬN).!J< }pD} o#}{$kL膂NDj+OkAH-r1I/^QnmT3z'9z:1[Z*K뿎- ^.ݬDs !JW'KW:n$JazcLoښQ?ML:9v>,F5xffvvԇ)vI%`(kZ[C:ٳ6xүRb"QY3tFlD,s*Z;[7:gm]Jl1v=E@fwxq{/8۟[۟[yn?}GG{RXӎ=pÔ2),P\"JD8о߶7.M`BzKզ[e_m ɓQ!֊$X!vq2ȼ4_+MZTgڱ*HJU.(,R҅QI!G!·+EPbۑ70Q`\~F#U%/1*Hs&vWr* E0Kvkh ca)1mG~{eCGXQTQG\'Tr`1pCBD`AI1<ʖ"!30n(XR0f gIC{I+i\L$5''UIDo`>QIQ)A!'4ځ|iX"S?t1LM@Q/9Jؘ`G='J++ۗi] ZBYl(ļh(OO߰bn N)vMG ӊ&4g-lɒݧGlF9bH`C/]!6M}eFEvgY,_ 'wXvl6w_60&i+-G#9q߷83Z3={g4:vYa* Y)'5O[DYV$Թ1+x< ci˖N72iok\~W>IհFYMP09"oot #Z&8ؖxy1JYbúiTrP4&2ʺ$UJ"tpJX@Vg!V8NZI"(jxdi,LϯˁO Q*}x̊[ @ox)bboZ~/Oe_~w?CBFfE9%8}r_8g3]a|Q4'B.v?/Sj?soW%2Fhrx/ΌNr1qSYL[&̻ܼ]KIwƎi)QG-uD2^K6$A"JawRZ*ѶNrиbe+"O$kFxr^3EB0*GW^1Hآ;O P#Y}YfǍ1`ۑ V}~}ZC9f]af 5ct=Y^n䪫q{ޏTK KLj.ݡFn8JhI Z֥~fpU*'Fxs#;`ZbUъUъؘTPC8X8C5qaLrؘtm óՏ-;<\Q=>˧iLUgGЁZ vr8;oqV0 S( 2$;y>+^v>^w1k @t%z&Ys-LYe+?WM_Ϋ|MDi'׳WK1gKڸ[[f!16cdh`tt$2dOciKfF@EAB0Pw-1Zv5 aD&.ݣ3О(b]ٟ<~t/k;5y풻b~]m俞_g1=EvWoW?FcWҖ[}r慠2x?{tvL*zө;crHzD8Fi9QH--8zzM1ru'⎚}Q\('KDm8yUv<͏W%]}wO0{؍,C9j%;^ Mg( PPPeҎSp8W8#k ~}]6iV(ܝ&Mͨ)w.I} uzq~~9~{`zͻPLUdܿNVBNjO8˟_OC)oh>(;/fոUG !ּc-2W;!Tkłg4xu lxD빯|9p=9q\xp[ϲ4dgU1q;9'&)pIwx"wCR1_܎i8cL/|q(&g}/ئRQ}bře`b|A3S3鋧=g'`_ҺNc>7D(i@<@!$!;K6sLQ(e  "/Ɯy0DfkP+&3%HuEj1}ePLbRO c IW>hܬgF nm0'foi\ & +,!J +@hnS:Hu2N\.g 5#lP*E]C)fs^[ .Vu¯V&H~!YZPB+eEJƺ4?lnb8E}ytt a,]V\h$NB%03i=ȕn;b5kq ⇊+{בCͧC  3-C=_6lC|31Ë.{;tV~fU֩窱S~}Un~ !X3I*-C0Is"sr# LTRn3L]kr@J@$chbh5Q9ҲΗ1Ko"D*oKP!Q/Ed*q@)rgsT6Վ={Mo?j{X0>t oICQy.\|tr6?YSSvF|~∩ Zr 8C]"I*wיm؃ 0S=pq!|IJK EEmVΡS-G1KYE I"[Yiz\{;Xϔmy/ X[ދuځԓ‘"ͤ,(Y7Az&LN|)XpԨӼBYIAgH{$7"x%]s*KUp_?-X*&,R'NIvi%3*>X)⃔Of6]9Cѷ%uq -!zyǬ_pHr8~&nown c vK} vƬN/sɮMhHkQ*mX֭b4>Ĉ)']9u 71[vNjr/*Ŏ~ZOa\h#N6$'\S([66lBS98FڬZҬSj/ R!g1Oq`9D `*1"-(FoM ?gB`Fs]*S{PE O d . [EFeAg^lmn[@#s|#)5M誰eejUO/[k;>u7@xkXz2IF*ka G'oP\#NGP|+/|hɱ$'|-@@^zҰ,٩XH  |ԒMJB~4@DMen =Tn}ֶWwXZfVH;nVS^^^ 1xҁdm$ 6PγfN#A9dd V699h"``I"x+ܢ nՁb1d jV @`_A*[] iJV\XS0%%gZ>Bqn4S1۾1%{Κu ;>Պ=tptEިZ=;+)dJfˈ!jC~d8hD(sK@tn !r1$2-ՌE1/, 9sŁsI6i)$0YiQ<&vZIE>fK)|#21.V/.JkIK{#3mOX|]`r ց= QKۯjRFlDfP:`s3v=ko#7E=\|? Af68lrd0_ H6ikW<; HɶڒZf7ُq flXUG$1ؗ,TLiV1C 91k5(44،yԷo})$Q/ҧ^6wŽ)nt\yg'5=1<[NEr-?n|9-s-dpTYnD[NGJ)yڄ6ށDDnL95_m:y?Wcfo*ˋ RpHsV.ӻo 8R! @ed+n aT*2;} _ק& p7E@[FI kgmX "˽ <!BjCiT 1vBیkDAP"T܁BHL S{; C 0B"2.:*my`);;2{n'i&B`-HP2D !sSne~b0/t ]Id%gjʴkDu '|hݫ Փ.e:2@QO9B:ܜc5Mc0ܜHٹWI_LwEuzK ](s׀nzUoɮwr|Kjdj履lϖ`DhHh?ή'?] /sK؞?]NʩMf1y!0di*͆*˛"*ck pdG%)씢@iAK"Ke3FU=xtɚ`sS. }ޑȄp)`զ}]G`yqϭҍZhFz/wfIā}`!J )]dN)MH8/19Z%F;/`@F?BN"22ύ$qHI*!%$fuU|;1yCI3*'qJO8(2%i)_)`kKGHf x/v 6V<Մ[02k7yA{) W.ڮ}#SW`kyiFNAg^WAP-911ͼW`PiQ @!OqB^Gj G4{ZlđtB&gD D0Dg謦xLC:BDr߮/qSOfR46 }*O"pC`i (, PciSPٔaT2*tAO]2!G i2 N:W7QS=~)ra #`%D eywhsȖl@Jۼ)q.lrAH}AGSJoBzwdWUt[o&MPFraa 7QT]ʭi ucy/˹Ï1ZRV֞*6|ۘ^w_0)N=(MyW.mwv+1 +ñ2?m8v->j̐]&H@Pb:i9MnJ_ĎIdh܁r`RƬV[D ާ]/HJ$MB*IϽezo ֩^x70%gaBYjkqhC-8Drv+񂋚w<7<<%9\TDxgDŀ1՚dH*NjFH]ORkm\P}X1nq66n5IOC%B .$b=6Ef֐iKD޹x3%'!'[y:TՍ84CBeO#9DRmsפ(#!b jI`꛷osKǍW~ݏT[6]xH 'dEgD/aQU9dڏlEE&*Cdi*N gG{$f;)DT"CYl, 8bcy FRV +Ri3fg]z8lR AVm]a.D ,(9 b^umU:4hQDq<>]զa8 +xGXG'k.ñG+Kb15!`DVm?̙%!v-?*vޮ?,{zUiR%q%e:C^ l$3[^CO(Źe5䔃3Jy5bp_a {x >ˮgsry_1kh$bxU-fLrtADfW NaAD׫e$ՔQ@) dD~m@0?x:!tq+>M;*L[LI9mܨgo.>l/_}.bpq1t̓|\)cyB8ό|\/SV1oQਕ!ka&#/ \/',ͤ'&/Ofoq8t?w}m'N䗷<ؖ}RQj5)׿q q\nL??~?_o{[+a7mgTؘg0޺:|sVDCTR3Ep+JbbBlҕTo8tҝBb'1FLC˝RIr%ŜڒS]b"VVrAkaYzuʔ'$Y?M62+0Ǖ98aՉA ܀9z9]ۥ]m{ک_jSZn@0ni'w7gٵws98gR/ #˩Mf1_~/ }V~Gn]b =/I!>g#GmK7H͹9i%=dWw3ϡ{DzydY>Ү|Qcc4*ҋ? ˔fRhdS#J4sl)$xlj$}nBj&EcDmރy%sL|m{]<بEq3x`)6JV';Ӝ)nG"|Ng^AR?'5#"/rZeHKcao%zioC:K!s CEҒwmJaN]KZTڹ3y~; 0Q,JrT6H٢^)P`*%@htM2 _]cd.(-SdI2'dM]4|*0s:QT`eh(+&d|{\ mYtH)k'' BVi*YFj M_]g׍8u#Kq&speg$_KKfa㠛D_qzω5VfO>!,'eH_ppwB(9IpDnz/f;^ dozN`RY'JyRDcITM^l t$N_HjbbmѦ"0>$:6u;vOi|5A|(_u$z.D2ɹb ;!\ R}B\+`uS2yJP&B#|bR6i|BB5/iNF0A(zVCS[pb: qr[JV^Q-Q`_ T('>пa[_ g~|<1{Q܊?ݱ wsjkw"Joț׏_(ϽWA/"H :E__^]td:yȷĮ/[tʿ34w`0ORiA$/~swW`A%]~NRqr**CTcQ.G v0uSx:DW' 3Ը4e XIRiwNjۅ+޷ s~0p:-" JDY؟e1ݧ7nƖ=h#+0rOd="G iw +^Ϛj/1ۏsO%N~B `Mw) ;mWaZgd[JSDU }EED)SۙwwѿucǩKAr%b>Ol5Af$U`CT IrYvTκPXtEq)AJnviz[SE׳i\c%bcti! $xb%kͨ>{r[d%fXkuX+cL b[ä٫u(#&(Oʸ@󃾺Hh^ad ô|B\qua`޹%;B&d`4Y/|RNzJ:tӾXIU#S[I\Hy~~#r<ެPQs+Y_G~r fnnb|sSd30ӡs7snukiFsw/C(gdDgotm4;s?~<27q}o7c{#xmһ7os)/{ݏ]~7pC{^{zӼ:/f of8{w`~E>Ç)("hOs.:¦D G4KTDTf`6aҌZIs^`x=NX8 [w&}]{rfѬ?/ E`KUH ,UÄQC\~oy+eM0ܙe^ûϪiT|W]ʼn'@{P)E{J9u84֬.wE~7yjRMBmޮϝ]4"⌟ЋW(T`X6ո}W NHpUnOې !~x݆GI|"|B)جbN>ݺ~$3hM1q#%$pRI2p?Ef0rXʘ *4`d) YKxtttt`ؕQPFc7fC9=:ebO{YPd\< RnQWM_|tZ8{:{*Y=jT$, wuTV+w )֮jY?m馦,ngѳkճ~7EGJAsah}DR<: g(H0~EG Iq8 " H~OÝ4ˇpwZ 3+щQ@9BNKaXLYvg8=g:N\-L\HRJIDeT+VHC\qN.(2U\P,NO⪝ Ao܂ע7FAGL-[p01h!ʨA2g`7fut|"*܆dU~cte HTUU]zI+AõkR,bRlu~0k.weoQYB3㏣\ď0N&wiΓՙZw.P7[0+MZun-^\vm`3x2*0?1Jd+[gNX8Ӥ/r ¬ӳkr F ݮoP'GWme7S}_Τ!8?sP vVQ="@s!*jI3巐ԉKI$_$DZX!G-K7x2_@> "h<+V?;Lw&+' qk(T\mqm(ZƵ3ZەD#vmd(}lcmW$Z?uWG`nï؝oxb'V0ѿӭR#Q&w]-"ѿ;- I".bݘo7ۍvO3n)D-6}cr1;§-bvZPGory0P( N N&XRKww ww>mw~զVԢ3vBT=VOE0$wB~5W>%$JiWGjGS[r}Xk%B(X#pApp`H0W_ю$hFzw3I4D™X8%3%T1Jen1V '!3kOK%kw.kV*IpۈQ~WF1}n@Rq$Gȸvnһ8* IR::Mp{-(N^lڿךr:Cˆ=b2L8hbө6Д$'NqL$K2Db |(>i(? ׊RR2/Mwh6+`E)M[}ZTúuVu~NC[\vQ?1lzEW%BX\۱A=L S R~yd_NL]Aw|qY K=Ҵr>7!_8E5QQ-*FBp SѬ[U[hNIv1ɺuc[(UT':֭8s Zպ5!_.֭,ekѺbPEuBۨc[nVnMhWA:8׬\u ŠꄶQǺn nVnMhWA:[7Ob-*FpZJwh,hUք|*SN]cԡ`жhedzhdmERUH` 3Y5"WCgpO/ykQ4H켜4osUhU1R=ʩT3I&zrx2=Q}ƻا>`ѕoo~{UʫN;&|`h'>gHg[gn_s޼*4+]yO37rC3>Ӱ 6+#r3+- -->2Uݏ" /{~ʮtlI7t*!Ae\bǂKނt`lL*+O`_KTfޝGk?%{R{r=I ӧ}˫ gW7=|;κ~]/U>}G86}mX},CITQ̝@)Rq⸠Y[:On0tq|P{}]\=&k6$6"MWdFՇU8:KpB"jR*6\B4IbXdz(%KqJ]4))dF8D|]ZgW.-ۉw~ @.Us")!Xŭco$aXa$MvAkDs;mPvZ>S2AK 㹝:k5~/X-~Kr[fo˜Sև @/m}3_ %X#aJa9,֙\s0X`e 4GiM89 u`1IJI@)E1>j>>CC8o?>E?dphYEV\8=2j'NڨsR ¢ܸ\%InIEI>U'II>$Q!Z$Q.Y|f -h; <0w1R c cS>eVܰF/O~/߹ϧnyR9k`2Gu71v]dA00w6ǖpdR'&ԛ5t BŨeHUxs4)S紴J5w%c048X?փ:9st8Hӏab&C.(q0C L,1;ғlŧ`` 3Q.>JyAݩ{aN':Uq\E[)ʸ翰Djɯ+;0m$FIbF9]>SB{%0 ŒKoTd\a-} Om uq$W|ڞi9yàwјB_/ۍ#SR(`mɖDVug3eY" Fő+ P5" )fGrc_P9Zx4 qJňp5CTǙ A b5ٹ0}y 9&ct"J 99b21/m=?PAgr2JtHl Ìk .$b$o-Šz̢Wn$LLWWQKe2>9aBD󁖌؈kyRS:\nJg1'(!RG2*;[-燩yւ?K@/Jb^( Ӆ+R8ov7{|PN$U' #{ 4Vӏ2 >nocyABBPR)iY(AA? |6{6 ~i`re ɝb+k-}tQbܺ3S,|Xݘ|o7hڧ1R&4/Yd&a '/=hzB" ($˨~c)Ө d}G MWziċ5MydLbj2PƼ/%/-* eq& +3ix(eB\SmT% q˅%G)pR{9E@ߞvz1{6[koy^8qLe;S̭2*`0Hgzr:T|Ly9/qǑi0)g@1y&("0 az}\; Sr Y/J?DSz0&Qh~20/.<)>hE)xӅJ8]-&q^i퍾6idEUr.q+joYm}~&jrd+W ?ܧM~K!IYSr#1$,@9f=w9Qx՚Ŭ p(, Ɓ&C;, fe0T@u*OQH<U\ROcD{>eS`Be`3|?8+sW%SRLMoJbKӥ,Z|[ͺX%dHS0;tླ’)VJ_`n{Ysd+ȠMx+GcagZp6_q> ɐ>oGaЎ2 @]̉uvft, gE/.!:hdW걶v\҇hAp2_PP@ρsoKEso3!?uE(`Qz8=- 쇲 A'k;'&R-ʍt(XM},b-LB?ܲ TR?Rϲbz e:|;oڶkbP:qZͳjZr}O[3M̻ޘU SKʐjm'@r hn7 @NwR! ܌OBHm^!$^\tlZ4/f':iu޸ pk̦>:˙ pdQ[6 iI,1V̎AddX-a\:QKh 6"g>oـ(e&B́ݔ.㜛<_y62&o^\#FPCv^w`asݛ9QX`C:L DM{sNo\R# lɺ5ˌ" Itdsu@L2[lٙFjtpxKI;eߐdWѡ 9y0pү!Ē~-ې-nL`LFz~;Vl_4eey@|ݓp[=ipqjք}8 c"&&}?W_ C^n qwM,ݴF}[ yy%#}ImGG|\7S??l2gURa 6"J뚽}{xwDxEɳ ?zf)d|ʅcL = 7[jYI<0wD*)3IX'D4̃E\*@!%uo|<B%1Yƨre}/) ~_fij'J-{:jؿ_&dJ2ќFJU*PJAo/e̢>72135gťMVsԨ؈IUh>$SݡF鬏XLƉr A;Ū E8nu,}(nc_?5 {rnMƃ3@3%wqʡR4{O%[*(D{ `@eh\a̘#@;ZjcOsA.pU哧jfhQ3{9Ԁ$8w[a{=f)!Uadg:jT];߲S"StfzMځ<*((q>o"KʺW3^1<>ógWv;5=e/E#!t_*u]wϖ7N$wo2{conv:Ѫmv]O9JLVVkV&˯/ $*mvA@M圳C:E20\a*}fxDpI1~BR1S傸{3t2z] q5 y9r?]D,x*Y]\L=|eCK(H;lDzo_ b#)HQ{Vw1$xh`X ޟ$aZ^.m5hmh3% سԬA 9pc"0ݔ=`BȫtFq㦐JK.e5#rp[o1K5sGpfn,,9_ly0[Y*V29C$44 l7k;A0ۭ\L-WgVNk 'CRuNX H@H NSCUbPE T5aqfguہ&9KZ7m>ӷ=^5qQ>\$YV=Ya*#qEMehXǧtFXvy -e,eOz|(kxL t>.7"瑯h~oǓ&~OqDdt^$(A_f0 oPl䣧U>׏ [ Jؗ3 !\ dBxaD3A(b%MA7.8fv@ ĐH7D_؝+lc۲|Ot20 MyL16")j)R Mk#ftWW0!!V`%QZp}?hI!dx. ؞B-f7+XFtn;uK Ui(+Í Lu|aI_BòO.xuuᆋ*V֯{f\Kg&Zu~dPuT6nW#MyE 1I%Ir brFώy2A3WKӔCjHLkmLcxPצB#Pb-e9asH܄-JS1=缵O% _e=^L{yRD|ޕ6n,Bʽ/ Yq&y@Ҵu"Kق栗Bm&)d%U_UuUuutڔKO?”<#:C M:36 O'Ai*P|޺%)Ij骠*2Tl u'fW N9TJ̒M8Y~4ҧ_,:*FD]dEJ/A5}Ƈw2lQx~c?ϚpjVa'DzqJ)S B F<picf-$ 7zc࡯6m#4!t(eF2VZ0Vܗ1{.ynw[VFFyo;uwBl33׀|Q}S]V0=29bzO)k([ իoA? l#sY)$ΥKNphG߅T%qlD>3-R q !FDء?rZU'sX6Dy$(g{X{tnnpz _u{р{Y`7 q8:"KC^h0z?P(мFCSzZvMw|Rf Uy[ؾ1!Nqrt#wmQ48zlqR @iÀ,^YSUa"#AտEyhqۻΪz;-`:4|"' ;/ ŬO(瑢~ >Cp<@<0KOl>.CX/O҈ÙŢ%x=i׌׳zGIL+V헿T;AͧDkvV7zꋈ$w1}PN\MCP+;cXHlzG3>)ڙD+xCX@8+{7 jF/b(\]Bh>~yc7^6͵z5I]zٺYsK]V^#&u<&mNP ָTRM.mNILk¬-,N;ښS m!M)DwKڪ#J["G ރfm,@ 98ffM웸XrgM8C!*Mm4 !5 2rBAֆYUfD@kdCM:UMւ#֯kmOŨ&Pk=QImVx*Dx%Z&p~ n=lUm"Nh7RX:[VU9]$Z$EQA:؎tcŞ?Oq0ڦgm0{\ &WmuuV^f;j뗮PuZ:,;2uĝ XQ_Mꯎ3t0(\dr_d㤦0Rr߯uh?cJXqm H5<:#{IP xAD&DK  ٫Qln0. 73y) T4Y\n_-0ϚIz_zӱ]ɚ;Z<&k_˭ۅLT|DV+/BݟSm &~f5L%!&ʠO X댣}?n5j{D}u^o`O:A% b}A,A ΦY?@!ςcEZ+Gth٣#ikpQWJɚ}]s.s7Z&?~՗\1L4X΂ɺ% ={4qn-*z vrfC͡KsL~ p _YxRS;w۳4wW"J Ly5Y{|]0uoK7xf2LƑVL|$i4&i?G f6M粖oyςfM tm_4|OgVVV3̧tu kPt,΀ܤίǦs G;΄~Τ'%}0L/\DdJK|IYm``cqPsLӆh%-v9MWkaj`kco4?BxK~x6۔<Ӊho_/0q :`PxWVيMH+Ҫeaw?,<=p(?wlfcWtRfUox tvdQ g"~xâhDpAsr=Gn6ГK@OPx[3sK1y߮dd뻼,z;uw.r)4tG?O&_EnYֳ%`=[ֳ%`| ]Xrq buL8eH0p^h!DRj"3T[jS0gRbj u[ć ckG۾o?v|K:XAf]Ld\qJ6\$Ʋr"4*R"6 f;D48 "C 3\N"P)iMgLY=d:M@ㄅF(`RwÌ@h-Hzx P:>J2ס< s8$TqBX% H(Ȅ누c|DW]~vSfj00@_C0nk 33^V6~nG/~`bƦT*! GF 2@r %fŝN,˜$ۈ; qr7wQѝH.aUB7%Da$$( C8/JTDX$Tf SU)8. +Gq"H%M,CgRkO$jeUDC&> p+_M9ӘX}d!AHedD A@C#j(T c% ,Fq@Cni): .n餍_1R,b1Ak+wVƑsPh|eW18XծoqhrbXC8 (@Q3c1QN@tj"':9"g 6@ܵh;DSɱdc>ѾLAʖq%""2{| 9:v#˳ƒR,~Kk=Z;NBl>0E\KBFc4_F֖l[JlDyż}>5O+f/rUؒދ4 Xjж_D.~^AR]~p}ĢX(n/U(.P^ -MdshAU,@×ֶrc3s@(Pni>~X۲7ٸI/*A{x6$s5TE^ fnxW*F`ɚCsaDqD ]GOQPmfUo+wj[ y" ~,HO9WB,*DueJlT*b/g~var"a9BHxG- `%5 F2 1D0 a ST;C^,8̚\U!U8XϒAW{*+^WPczT]9 K:4S$9Ē1%IdIe~)4`G6Tu4Uͻԕn{h<(su2m 1h.phUH.N2Ol輴>kZIrqO/*k%,* X1,*J)nlQq .*UH.:o I~ɧB(0J 7(oG 3z֢V $ZWgQBje=?iP'K4⍫E33MUʇ i7Ug_+(| mmm̡ c*1^l>Epx /VENށ^7m"I!x`Q쐝R". F,@='~2qD"&7LuN*+ NsB:T&8Qആ)trZA"@?-ڱݧU-<TBH+A~pPIeG O:H8ɻl :Sxncg.Wic06Fugs<0_)6:}s_GOxI4';i4lƩMf.3H48׋*( H6-1At㒿:z2qeCh\˔h\fZL* \d[!±zq@a0 #(ݨdy$&ĉ.puVj)?Ұ0ofO>ZY12~ b31mc#ێ9Btp*\"CMpV;'/+K#xg? $pu¦3X"v բXs蝾U\H7{ mMF^!e<0Ɨ7 ՅrS|Rq+헤xh y1,brNvѦ7_Ir>o}7SH RK{XEǪ.z`T*Tm ^^Qj*AHqM驵"{3{32U@DU R,U,)Zi::~dWj }@jh^-܁`!ر*m2USH2}~g;䚙'jӲy@VUҦ@FsFi"XID֢9O@f8qT @s?m%L-A hvP#-bR-Svt*>EKMCMĒ؝Y /|q===ƽ/^M# *GFHF@\DD 84 c\(±A DRss>_{w gcs#:еWfjU_G݋;~3䡋#w.HS|Λvtuj"֞4v!B!M$6 B'Ĉ(h 払Pͩcښ۸_ae+3WVRvv%`$coƐ)bᘦ˕Xp_h4qouqev3T1fve%c!+ݷŢ.{oWUó_=Hvg&f[.[4 )\,9Q2 c(aX9Q\ #1' ɕbyr ̖& D%6>9&D \ 4IKR(k:ePV+rۋSMuOf#,toǹo͛+^Gcdg}P!=mUGz?Z=F#^S{*3˟֝\3]K…(?G%*{bl d岿u~F)iEQ"hRZJ􂳤Az||ZRO5j#gʷ{cG"uVa"(y4Ąa1Tɵ qLY~a|7.tX38 s *P$WryΧ@vcFVVlΑL( _`V)UAd)fK/%i#//J, 7JlQ F* + |qs!E6la&$lbs4K7.=i1m^0 X4~z6Jq?&"2UWs,8; K<=O9|Χ[;n|DbT8JMʘd~) ehQd4Ĉa,!9Cu;~`G";*vƺ}r0EOXiaƨ FO*M(mz'ilBDL'bSĤQiFrK%\#lTDdL0u߶\y|D7Joe=1G!SU9tmwgEcs8.oXݕ/rX]![|0dހA*of̤+úi\_&]:1= 'OJ~fǣxrQ( Lۜ:Y*9&VK5!PGSpb'dxsÜ"˩rdE Lp d Ve@KX 4'[)Euwqjz00`.c(4Wh;iikjC1*w:As+3nV:Z܅ly'4"ƻu2V.FvU:t_^Ȅ Θ"a3Ԫ%'Ǫ|xlZ,vg'іsFC[z>F. #^ې('457fqzPf=K J )DҦA6OfiꮩzuF ߁cBvt('#TQB.='_WdMJBwt-b CÎvSBҴ7*&uAcBҥ!jJ"`2`wJ>V ۯz+q 湆H[wO^.TuB zP)&Wac[ATpn 8hO^ U=͌̆O}w%)< )ۋ{o}鰗6?w7}"W(3~NlBds?eG/oV>=/~;S+ljůZ,R]zc~R̝ <( V߁XiOrg%(Wiw>:URɺyTѴu[ȉ]%[hXw&ۡV. qեnak19QuONp rOIbrZU}ZT>TQP@5Iz`=ڽD{9%=z眒ӭsJlR-uU*u)ӺFN8 u> Ք0+Ub'liJl- !߹z/Զ*ͿmS& m,+٢E͊F٧'LnT\"VI9c74pPdc"+F;WqCd3`s22MNO@ެ 7Ϩ)rnK'Fl;F?K!s MNE~;6U_ qw߃/ûyyR9 WMS/8Tby=8zKxpP$*?)cC9ۃ44Z QLsJ!DohJN4)~{0e4bJ>$+#ٵgG%tiVei//URB2?ڞЏuuDm֎_xx m5`mԊ4,IeR5M۟U^uJ'A=]9KcMZ+…h' _*Aڦ:k:|-69BEPDqY~/q e"A<rq^G粝 =U F 'C`x'9<0N# ѷp߲S u5iȀP&J{oW([€؊ 4Ӊb4,\DGI2Lsf8LIW<1 g\(C˔AiJS&*GIqz_5Põ?r6SsUw2\SO'ZKyuIǚZy _E]Rg>IL)aK :&f٤;dr1L".8 XJƔ?LH73#EB PX3 C 7e⒠8E"IfNN$}FM faڞLV67(b][[c~]c[gU`Fd&ZzR7AC,O"ӥ#zuzJ{>|owFb:Z=L&%pmDYGr }[w>q;9c RTڳ4TUFRGDL\UVCR/ \6 8;TFo8nXο\m UQh6vP"i5sQ0$l !ƪۥ z4*^`FskF`}6"3.4RG1ʘ [AH|0#tBZ5)h . 'N,)aa9ZFkx'f6 a aJmq1)H,qĒXE:&,2Y#P"cuŰ84/tW69p4v0rޢLa?,ncl%ʆsoMYɪ r3ʼn|yh?gDTЗ.[kIhG¬)wtI55.FmP®-\ _xSE.FRMz4]uQ_\0(Ja0\Ba!"Tz) D.J|.5є91=f6˭'wn!0qb]5ERR(UY5UAnHtqWv߯eeBiZ"UEٯa5"P[TatT/ ]c>Vۯ"<.oU )EyCVAywoVdM4Zر}n@XH"zhVѠSB@͐!X"Kx2V_pr] (h37/ӹny@D Pos;6j;sX`x{kb1a9==/7y)e)U/xJ=.pa/el(g{y໋sK.$jL;خ}L>6g* uWW<7Ml)r`Pju~_j|EeS[a#{uNUO}FEs>p _H4ddy i}>=TOj mE&DL(cq n5 $5J]y Ai_@ödMfhA`FYGyT0Bp#t Z^摶$>vANjXÕBF$yA5WܝkY!#STQ]Ys\+S|JFľ,ٕU*N3CY)ۘ8gĉtMX˜ݓ6neO;k9M:Jldu6L1XS7eZvM9v", ~DŖ9e$؏ScԚ'4 gc>YnRow c}gFn4o_"ߥGtјe=۵|.K\oVQ5j%k|Fyf ͮ']cm}qF˚f#giͻ2ϋQb2̋O8iz֐/RLa0_TLj$_BSvEpC+vEkDs<^T Eεn$z5@QY0H=v^\6,QK;y wg.|%c+{[8^ySv݀_!)>;T5Q|1a% ۡrc|ofkI%Yךߤ585D|8p+pD{a iCN < \XctLx:.Eqty:/GտOR7f8ZeYEhv/uUW$2b z߫{&:%?ד\V&yo'G?%g1HI|׊1z| ?3瓥l?~ N٧{7TL}MWI)B4;XQE /1 {J3Jh2< !`HVx!B@E㕢{,BN9oQcH[eY̅%XEŒ)&R#!#>ZÌS!ȹ$\'O1j{X$*D zsOnO]Kt97&Rm˟ꛧ'fqꊭV?޶sBV}ǜ"rۿUXswf0/5<k)"{jWp矾;u 69'GW<]3Awbr8Fo./=}[ IMTW.8!y=T\b)p`K5NoTD%#ﳿZ2tmLiq;5QDLLLW׋D;&&=W R˨%g'Gǟ|f7MOVUM%o[l rFw;I6\*^ 3-,Ӛ6,zwj%v"ul;\--R,f+ GqdC2E ,`IL0Hpu՝BM8[,_Vۈ;)|6K^x\x7FsXBlPā 34ހTZgGkA߭<#0-TT|LJ`"࿺G&0=ծ2S 6TBRH!L3\Ѭ!Fjd@kI92 +3܉ۃĄn`RNS֒k>տGK# ?ds5M5wES0,pnx5qABexpAru +ϰX /)6hAPqRD鵲$ cCҒn4s 5-7m^XnV:e`3lyo׈I"f ax|"[@ ܮxnlE0Q];0咑/Pa BhGqLv>fcI?s׫vRO矗5fNq;rQg8 s):Syl9{:QC^~ƈY"{S|y~V#kchc"d-?|gMD)1&=GG61MYg#gW/:w5<_T;|h|M1Qv[<~KvM`;CG{X{VC Pv6;7Vͻjù;n(COgjoө^!dA̍֝VYZy|P^1)eJTY:'+lR]?ݻIwcRУsRjᰤ8WY^L2ZyrnY# =eO6[ݑߍؒ e_Rn `Qxnʎ3BShO73W'$I_)O>ϖ]pJT Z&A4n/v^^0_AfRTYB̤U X:+>I y+p?e(A 1a,S!\g[8ƺ`Pj>\\} wn z͙`<+,o rNE7$}g)k" 02ƹ Jz5qIe>21ňBҘ, %`&"Fe:_l%# .^/05%J&9W ċs#1)HDpG>#mSV?tXٸuHhX+e䊼 `Ȃq`JU^DS 2*r+*$S4y&AnT 5>56[26R"M)x mOZF`َ 4PO2uv{cUIT;S3rJyVXGbmxl''F*,R`~Q Q0 ?F >(s:`j"LDү>7R2T"|Db,b4c<[xc Q3@tߥ Gw\ojx ƕ)5DZ i Ni\'<"eSB{Z26t%TCux~+4:3k%zLPhgTyꆁ/yvH]*YzqS0ѺrzةrD2baw$k*!NЂ2xO5ԩBZzO>ӝBP,S7FיL}m4Bnֱm+v U>UI_]L6'r!u 44 ڥ)>K1p𻻺9eJbrz߬C6=ON/Wa(|-##ަژ>As~\ߤfjͯ PFW%*]g oS˽gLIGa7J5 :maVB][B9D0Uቱn[̀V蔶QV][B9D0D3v+ҼRc*ɐCZT(!6zL7dU9FDvuf+$ i$+7 ,]`a#$HRLGZ¯td~ wa*5Uuwe3 #+[/:șJۨhӿT{*}鿴gL Of7̘EdT9)m ݁M݊e6!Z)1]1[rSFvnl= nm CS7=ߢok1kӝ\n@AFHpLX"_FdXbg5BP6Z.!”bK1G|@jv^KB>{2Xb3T9l[F`@;YS26!Z`je7MGqT9)m .M݊e6!ZJzE2T]8`aBí""`N١!Z)w+NYּךqQVim]PFN:鮤,1N[Tga6ELF:H:aZQڑv1A丫L)sh2G^sRviu4TL~Ey-'zsqPY.iaVMTbAWM- CSMcB X!tGwiZYv5^Xjb%4DXi͹DYEb Q"wD@%J҆<#DZ.Q4Bu(Jo;'~ +hW\tGxkQa=RL3m$}yLO [1l郟 U^'3?>M) rM%?Ȩ؎y *> mOF- W1sІLT)>c缡 9LgՑ|T7 8@q FoޔVՉѾ,ETp\a53+62Mk(x[I0IsvM (+uQfʆ'McJYRk/4J"11IK#+3R1 rH߫%?{ƍv»骶loj뒍+_!q#Z~!%g83$%;q98t(ulA,"nŢTC읉5bbicB(mz`JiR(|u҆6U…>EisR6DRiluA5ഩ\"/C8̄VZdA ?|.KKF*B)>բJqUpS;\^ƍS-#+0l=:&.^db2 XP܈ 5.L߾S6RK*I$l]H$z7wk tj2¯i6oYLnlQnוh(xfy|[wpwʇ~w7{%XT)ҹXoͿD+h/D\Oqܬ4}"٦^+~&AH߬LU`ё Rhr:Uyd`e9|OհUޱkozLaDnpʚ?N{#*']ɍAU(˛LG"i H4h]O+Y,5*-lt2m-c-QbpAϙU?hmScuKDG^ 14j!Dy6ZFW,Zf K7CTJ! ;!i;zf:/mG2c(hT7ޜ|6{$ۢ=;-L||A1o[XA)o؛wKLWBhG y}ϠR_^{q8+ptqOiVrZ)Ng/v4hKaL ۽?p.)ˆ~˟5`|i:1Rm:12 hSwAՁ7oKeRRe})+g^MOWYESb0"w?f-_d㝋t\"'"-^&q\sS! :`3p=>KyW|9IE죜}@E b29MX.r]7|٬%0P/.`Rg&ЪYIc#.(Y\&H.M7ꯃ9c.bС+Y~{/aiƅ.h.R1m`{ vhKxBHN :ɴR٫g ZnBq|R 4FK 1~xMmM6u: R':9Af:0Zu[vtWiwʫo2tӑ{ya!, $+icD+ b H_>%ɞߔWM}l9bM~JAJS_R\M}#a ; ,HɁh)׻89p7/Ӽcwc)2NJܜyHΛVκEX,jJdlR0v,Kmsl9׌jǒMP~\Wi/Q^D-(jy FpteEj  9 Pm}es#: k n^MɒN'wyRpeW.&#@pc]:,%Da<U ;6'(\srV@l6gJebn= 8r 5ALd8Il.Y9ZFC@MLK`ɴ8|PT.RHcS{\&!7ڇ ,F}`HЎX#mࣷ<8jL@nc# eXRcp & ,hK'TEEcJ2 g:R#Y8ۗaTaTBS鹡;"Bt!O%W!edC0H.%M Q)n4ElI o. ^4@=KZv;PdgK%^yȜP4U/GӘ__I@x$ .t5Xa:CPXkmFjA{Ai6*b{[, ~4<] 75,P>T)%PH'1U+/lPkƴX4AmqNrAYxv;Q0KwvEF;bZX`n\VV#\$J… HV^G #خ{` X^bSkBRri`f[&G"v8X[Hun] ) Te=7{\(1Fm TebAg"@>CqIt9x==-W<}| ^^::6܋ɑ=뿻~0t$(.ɂjvONLWB 0 nKC]O%!2BpԣβˆOg<^VJă|RU*f%s*r *ZL侭Æm+ 5b@`.R(y`4h&wsHɵu#C:zk*klX3#BhPBe&}dDqN+[luYEF,(,'hY>i{=ffױfILfqTKBZ>8fcexp.^rS  64<d.sfsV񥬌')RX;:jT4h æT=Dq [;v4` 0JJq”TRYHAfm]v̜V5,/{rWJk!L\h}Y[!?ݢ5Mpsє#[]؃۫ۋh_jnF3ۮkb͚0j:5če{_98"btPۋٹ )Z[E8èf=ᯯe7WWٴhUk3Cb_@o:ȏ \;0Z8a7]+1"2Vg2, `q.$ <90х4_~Ql) |de7q̐C؝մe6tf&m,pa lڶۙ8uH9÷WfF04 E8P M1YCd<ܣj(u-V|ݍ"u3)OvLr=J-[{mAzdb{oc{vQf }.2 OM-X#]vFyt]CS"]A<|rMB]>Zppxu 4%t7U S 3rI_)k7 9sQh˅c,zk{K4RCxVk^!.6z1S&S@z٠,).\WpzCIfӯC8GD󹮜@%u*7\^'kLAY䣝\ ' Jd1NI2܀ p'ƙN01L2f$X62@J-y孠^WC?.5C6Iyt `{{8_'9 {i:^>5AǓe9p)A G.ogv؝ IRGnW{U$~7.X+Hxbsy8YL`>%I6[?e%KíWϴbbb^ 1 3/f? nJ6x[[vXz{'"$6`-z .tqT!#5DJAJ1FCϽi>KjR¿u"l!"*k)BY4^p%B$b:hC: de͢0^%viD/?DaێZԭ>UZ{Iτ9fJ 6P .?`\e`UΉT'U-8Vdts&}ݰ F“AI\LZFA7'q6Igf+RX.I:E`(ʣnDK>#dMԷ7' [sUtU}v@&-$8@`pڒS'-G!Lցj5:QY*J^FkBšVP,pEhf: UV =^x$SYO Z`5 HN9j!/s) : l|+p 'DvTOWUsL[!G>Am!@,>+\Ksd WQT<oR' 3s$/A nT]I VSH,N*ͻ{_*V'[w+Jэ`YH\ɗ]{)yFH6jL6$D"s׊Qp( ЊmAr||I̷2䷴H$6o1 ԧ4[ jT:bTϟ^MtdTүɌR㝼+@Nj> MSȘǗtv8h,uɃ-$HDyWuJ..zM'H CvZH0]Q4Ygܒ5Yrƍ!肵5#6DB1ODI 4?HdT %i`@ΔBH$L%4Uڣ3:H Z#' Kmov~};g?l3dvOO[ 04qϟ~o4 m请 aִHN_so֏;ƃ|;+dRkI<w)V ${ssg֏sx2]4HH bOixP;P^68w< ,!)9S>`D2,2szy;bӻ;eqwǢ;[2cQdn"Ť&E璙8/A ~[bz;ggr\5OuBF @ThTSݱ`A{=XOB4JFi@cXIƆOVҢ܊$ʩޛ31ȍKz3 9&6MWw&M6O~uw髍3<߄GoIbM0/L:) 3h5~nl59޳}1;j)kU 9s4Kpʙ3R Ox 0ں>KPif8cI-sMj{5f5#zxeQ~@0K?߽2 40/@$Sʤ_xv;u Axj+Hk1Wt@[..ѐ֗]B2@M~"_,IpɐE|RQ!Z'\R+A>L\TGSnD:0[QA&^ƾkK's0҂B*ODϵw2t4X|ЁpQdv"p<\GD@P%N$Y8,>-ӥ;r1e' GַO?u.OK.o$^QQ(=xd}RYPX~n%h-*ۗƈ]I\:-}xSw'XJt)U!*_o:pzYӤ"%#cad=?i7Td$pf"U%A G?.b C \R1 L TTC!՛4SJׂ,xQiQȋ]Yk{. dWl\IS襗ERִh \RYZ }%ez/ʢaaJTU?| ~qeцU(VT-XH F0ѣҺ=D}L`k=.'PٻD^j]s՛  a}NݸJOa,}E% ȸzPQ(h`QA"!d1qGIS""te|Ac]sݻD ȑ&5=4 Ga`]b | (4D+'a[[oOP&GF5:c&"Z;lBNS6p~GifbH)f_׈2F9R]->ZHFaX^hU)!rk)%t& US}5ǥ(MM/Vp\Ss1gS'P"g:!t 6a+g2cZu~E|.#40􁁹Hp.L('mbfQҜ:;V)gz}}{y~ty6O62zOxHG}^"zO"Ocѐ(0:ޫ)ĽEB|./!>f!}% #Ҏ5E,q4;fG!|";rhbtN .TbaVjW)"_-eHA+/旗ږ15>U79޻ɹD8N(3FkJMZH_2Csg̸֘ _}A*VP2JԱʂkYئHlz+T1/k|E`|l/u hR븡X^\yWjˑsNV]6{xj?{LοLV_`JVShMT=d5%V-NYv0zMq"oyyK<3gn_9+A= OS~SprJ>R>xI}Ճ0!)B,U! gK R'E ccC @x67/n)_6pp6nM7[ԏ_7n`6泺Jwm5=ڥ3qm6nmVYNtY溤ikL|4xDs`Io7(K3<rvY*z|U_ml#nv}]Mcz$K${5-NZ_|1hv_na1oPi;}S*>p6vV|@9Y]2A}SOv}V~ɦsuETϟn@fD H#+OHUZן%iCQ9[:T5dl:&0(,Č(`HҙQfAFĤ J,v5199JZGJqY_Nd=6֗Z!zRv]kK0UyW E$^ HHQ4ʅX,DXRS\Oj)N-gSK#<@d):jIpz!Ԡ2& y.Wb֔q%$eÉASib` i8kH-5]7cpCB#?!z B 8E䑫= Mp0PQ!]5a9ϔuwijK4B# x% Ew>z<+5ƥ Ӎlq~g m.)[E3 0Ɨ.i‡8{OO}{φ>4wKSkcW2~~<'\iAg%O :J7y?Ό@g7;j$%&զ3 \m: uA?s^=1gϜJi x%UNk!/8G|[] P3Vn%JVnM0SlGXwA}*m{3IWns[ yũu68uϻ ޭ.USxX&r1V?Tѻ5NZrgy7JE?[nu1BXŻm/tZ)j7wk!/u8El=V*ԩU2IënZջ5NZ*Y{MXbP:uw^DimVU[ y)wwXwA}*mUS ݚ` hNI$vJw9VJ5Ҽ+/%"*nhy)ZS{,MҮl(R VZW%U 6a }[ۦ"ue]EY^ uG]Q ݷ5jUt*ZƒۺnZ"U5tU Zb 2]AKNiwO0鲥]QKPLݞ& cZYkpy>r_2k seֺZ J+\%ư&ˬuF-A5R3R]ˬu,u玶Y{ceֺZ ܻKۓu.֤%Pev.֨%%/tUY2k Xfd2kL.e֚_[̬qFPY2kMZ昷Y.ee&wG2k/"&ʧCMfM(eֺZ߿̚$ˬuF-A3kZ/ ׷9Vh07_ӓ{O7󜚋ADv,;?ه__8O n6F=BGof2~0L>UMh'!@%o]]1q@SS˾s{.@|IAKV<5Њ2yr$gcrA{dy8,|{ D7ͦaqČ3gF mϷL6 _|~^R6;qbV {ڸO^旳dh߮qpNeY{KKy,V;s: Pkm nUy.b LG|*N@(@z\>ra"-zx7^NnP-k ~4+is|ܖ},c/hYxk.|y CW\"8} O0s _i-=?'Vs[ 0ٜ"b4}N ɴմp{۽T1sU侂Gʡ'VV*'Z17m15MZD?zth(G0DK(C9pMf@;?Sg jՏ- ڝ7_AλޑF?x/ѻܿ{?>/:r.-CK>|g59{OϪ } gRbzPJ&)dB\`dK \WG(^1{ODP"ڍYVT'RT+6^+O߳[NET~br$d۾ i5_y_aE-DQ)b2!-An0/0:P d/`?BAQiin}Q~W&bS>uk?|}S*Kk6/҅a*Ui-pqQ70G襀7 wfmpP4~gEPK8;S$9*%56doOuªD zAɥ-흳+✼\Wf9rH2IRRϗ:!9T*nzĝs'yI9(䉦`q>y,Hyz8ĵ hTnOfqdm^ szO CE:/1"NS ވ`u$q G*0*! cK,(%x %L xf=1!H Ad\hB**aهX@(㼧e1*^RQۺ0D԰6Q~Io|Z:ڈ6/p0[HKWY+C<|H)= LͥUvƑ޸9t\2T=¨S%9C%\)[q@r1ǖYx]B$`\"yC%W>>Ow~2SQ9 vIUpZƫtzWRS9fNgmK[6+._d._ U#Zf\[FNx/++^#bs< [D<+cB8~C+qRUK0ecK%(wTvej3%GO&5ˊ&`ןu3Q.y?myCo7?%#![~Hݛ+/)Fp5|b?O&}3J`ۃ7B S3Ht^ &zpLE\ɨN %KW5q?ơҲ A$lw8LL8ܲOcnxgL"w2s71M,7Z ^W*P܈O$9L)nWU^G^]нKRzgN.jZ/G,ut+.nw6W:@uJUw׭zrqnoC;v.L}T [mo@#ODJ J]9jbL(V@Η+v͉Vɇg&*0R0KXh EŁ96FGٹbtt(h 8PtIa~b\ t Of2!bRbԜ6'٣ P-P8.8$ag@e,M4 T2*.FV*J Qr$;ד td8NR9}о|$Hom?w 1K̿K}y;OD'UYaś3fWu[t0z ^uOx^alf' N0|?ڠCJ`RdDx 1? V>8(os% w-- [B:A0 uR-}K6*R/N Bb 8-a%Or&[Z>3&' HJWF(sT֟AzKOLnT,D%߹Dō dTzW^鋈BQ+>%: V=:hd [b1%B|s5'Gn|Pqb~e)-.!Db(珲%TCkQy:$hy%iLlkn_W[Ѫ%`p Jn绌,P Iqjd]!#$Ȫ=.ѱ1ʙnǨCHTvvnzD@͋΢)g6|6ï! Zh)ࠩ+v#+R3fٻ1q$K7^Ae!x`r+~.O ؝7 싦 {4奺؜M{Du3q#JQG:7* Z[$7A1x¢ 6+Z6)q% "U^r|j\.э2,mVQ6$]QUK;s#ܟ~VkH_v)kO&Z8t眢d"^[V* ,8 EDw,%ǖsj$/  ۔7NAOf#_ބ:ƣ d^|_;=e%vHp3IxB .51 #=)M:td5oDqӼp|<^hDPpD7@NRAw!:K KCgnCva6O"c{ B 1?|V2x8D~J1f8N8~G~;XA1 *(Æj80.p>%ryXNt3- [#X(1Fz{_)VڧqS;  ,xfSbĘ"r-Uu {H^RJ&/ "Rį7EFh(MirѴv6-ܳ(u! )ՎHmz%A)Xi=zv6@mԝݲKEwq5lQ$X9X\Ҥ`jO~lD^d B)0ԅ4I7cF~(Տ1r#ȘuA;B" ZDPPQT s"4-u'%u'?]%uYyY!+`4B 9"c]Gg>J`dQ%lE Aڊh<Ր9x󜵬W\(dBσBatngo_!xa ߓONQJQ9'"10耬eQY9ǒcGHY($u.4=YwZ{'5qг6ȗb&)b-f̛Q2oFɼ͛[#\P(!9oM)JxL" iOq`$4[NTKN#$Ss@>Kzl eQCL#q5_7y!*_DD V?WJkh5˕txZs{+xwH_1Pc/_E>83OiH Q-Hpk+E_#mzQw;Kۉ>D^ ?B)fvu~:ɱZW7i2"W?6+c ^s _yp` 3KB3n< ٌpV/pެ> ~U׮ׯ^a1Lia= g V1CA$+!jLcY9w$ߐސ 4AAiH뗎u6 (C|ի 7Fc]ozʝ`O7;~ т\2]eR=ѾzOOl=θmjqF=q>9/q^fnq,.n_D ʧ'~2sT;X,ﰃ8 у{ɓn'kZ{lN:[Hæ6; 4/Һ")Y"9lRFBC1N( C!d(HPv p:1SdQ0 kLW̯-LFu7_#C/$Fj:LQ{0u9Ҍ\01|bB6'd2/\(t$*AZ?xdb5՜\gH9`ak(ln!r!oa):5W*ɫ:Wh9m!Bep~~9a/,`.ϒRz˚2w E4I?=x[*1.xvTTckj!$ $"-Φvӕ[*1.^N[Dch\2.͵7Azf!b}p:&Bݙ7g~5WW٫_xBT$!H;hg!ivds긦QmuIJo8CxQ3>E&fZGDSLpX)rnFR> FRV[\N`N1r. ,:#)TF~`5"5/ds.԰Л  Rjh7[p0 [7 \ ^N/4hIf1!;og8B(EBu^vϘ} WhZ=.C3,&@*0K`PR)KFRVa/ohEDiʚI$(F:KT\s@kK5LͰ 3N PN ^uG$T)=Y/w;\y;Ґcƨ!=LC9%[Z>MxSۄc[K2Jl"\b:xN礋[ Fq`FB;н?(B@}kɳKݧvD& =OdӉ`[YIpdjڠ#Jú:P^6sYS.67J{ILcr挕B0 X`#%D$9j(x@]oϹvs3\qdP HjZB&h4VKKND3 o?zsi21WZuq,ÜM=Gq/}S!Lco`=_e"\눎>%-o`El-%q03mV ޱhiZqpd[LDqvOyp[dfҊH64 ZncYsօ֋ 񄪽t`=jـ zG,@卿)U|Lpo޽ރ^_})ni M(:(K:F,mte 0> qwliRN^Y) W4csh2T-fJ~ 2&[XE-wS…:&S4◫OUPp)ŤOg#Ǫ<#XWܽ }/ё[HF1m 7i^#W*=ѯO!\ zStu9$foǟ{yʜ&G1LH=Q8Ȑ[pH܆b`85{+[,'dVQlj-T=:zGb͞ o/)2j/S1B/r~1,Q803n>>A p ŸD j 40 Q!lO!lUpydrU#W)Xg>^ Vn{y^Y!ڏ[޷b?>e A./rsE|%_Y jT.r a8ܬZm̆f|_p5MnH}F=Cv:_77}K~yoa&3%FdMՐ7F2YdRDj+AQh&߆*^ gY +ʤ˔ޠy e( P( R% nv2"[\^bq㵶9$;'{-/yMrAxSf9 p).y ZڱmVBWv9ݣgĚ-$l5)$T0!A9&j`P1BKĄꖭVvϘa1gOiAqI3,Pэ 2 \HPaq>ggڀ8enN l9WVq 7ꗵisﲜ)14׈9e8y녌> .!y+#d2hn<`&ڧ,l*G#SƵ5[j.\+5b 2AmJDi!4&j4G#i5E$"dVLl&3&ۄȺX.('F+kR'.adwc.u -rf$VOicRci r$&Ê FU&PZFۄb(LBW)X\FW)p`,+é`1:``o̢%n`,XX Y,܁#aIUAA&\ƽW r9خ[1+mX5%Xpݷ~`,?"bou."Uv GR-$xT`m09iИI^F%5aPGލJLI>OB~T'g#J8A5W+IxEiŘͽ=^CH.)2$ (=햊A褾v;uxZKSkEj!$ tS8M y -I}GvL/# xKtuWL]زpny C.rUw;2TUut}Sz~t56 *e&AE=ܖUysO>ԉaWnvŽ=}mW{Q*i0a,lp.RЩ=N=4,܅Ks,2]~&:M|zrn'c4.eKyKz94w>ܡbY"cEvErP$Ӆex0~&l Ƃ9R rTXp*ehhpWNv蘃Q|%9I@j$LZ)ڡ8$E(҈b(7YeYY˥"(#ea2^fUNB&3v:́1#QDQ}T bSWSb{E}!$ 2%jCjJn~'o+:RT/8e=h~.e=[hIY䊵edM_w[ګy^seo{N]4>w6  8NzbU[(J_u"ߔef߼{=c^_})mIP2$[Q)rՏx(ĊJ#Sy|kC'Dcd)T'է*W(g{Ĕb'H%&yG@ =& sDq:=bGU]o6W|9cwȇbw{8<קn Q6n;PcّcʔdobhG83Ùl(ſ0|&B| 4¥dY?TͯyMPJ=X'ec)8=[ ^. u *s0º:A-N-Cd{Wt%D?ԽtjѦ%l.*,hK\1xpq0SxPy8 pIaהk0/{SL8 ÄњW$^N .k} F[vN8$Z']Lxr3>ֈlO:Ʊb,O:2&j$' W4Z^ZWQjL%=C1XՔJVҎ&!C"ԓ&KeDOC s^v?R ]<i5P?Kdy1T2[Oic[[}bza1ESf'1z7Z}}fdԛ' gzS/f*?v}br)be>ujyW^P!ĕ,{MB"r⏁}QdS8a>2ji!TOn͈͗Ub?n=bhʌk`ϙEdJ8RmT}z yaX TcAU9iXh9|cA쏭|IŤs9j<H\vh'G5n)Rqcpa!^nT;1P 9GT`u;`d^J$Bj'|-˘N%aNT;\dIT@,348\gWX`x[m^xw(|?"xh(oO}'ܨSG,l_*n9Q@~҈NB$N>acsm]x5Z@҄E3fMxILe\$ғ2ϬZ,/b5FssWCF,*1 PŬdinŸUNCS1 q4#YJ[>?٧IgthĒA)GtBGh.Q&фx}x3Ǵ6_ڦ岜z_MUd1҈in'Ўs:c`:Ïo“'ʾI y9|)"O'[e,^u +[ݞlpɶy s>/_s7( NZB;FRbq{Bk\mwQ wѳ]"(TxE _J_Rro:pFMNMugqK@(u$щ" 8W Zj": U7a*U^a~JK"iE.ܾa@%/˜HŲ;pM((uEQKa݃1VI%T08~*[o,\WItK)^%I6\^-X]jN@5`+bMk)VGɮ)QRn)ۥHF!q4YRܘ牀RPZܯr?[vXݰmqP8pڼhN?1(hK2-ɤIp/"qF%  `6 䪡͡hIӌn !(y*ݔKKOS#5Y >LX\ S)p- p†ٌ+"MX^nr ,!д7ܝcac"WiDZXT:FDHJ&N'8AJ/#?#luX;j-&9-*I' Ԡϲ- Lee"a^NZcXϐF2qh?dv:Hא ?.Moz+V<  {σ?w{?Z|!z˯mdHƕn&^̪i<#WEgRxcNz-k2GwnYF>Fڏ0.⏝+cA":Z:[4[L﫭uLSΣ* ך |ݹ'iP\K"~15}{׵jA(\_GW\s}SJx10t_˚ko:c?65 RC#[w^s_BwNop F(ݙo-vr(uYKi'jZslCwzgf?A?wYCPa;J;#|sєGgv^,w t"9FqbUME-hӌϦu3M&eйCĩzĻkfb!hiHM*`1Kg)\o;wr/MsĬ pPRMqGG:~FpyFX;z(J2jȞڕllO~3^۝|c T"P$/+yY=Bs{NopIt培Y([FƴW4|K.6^+>}&q&nV|&l~6;bJ+}o|إ7v|ftrw";H @KYG>y ]$䕋hLRO",[,BD'u/J[|@քr)%ޢ|sn<wn'x=`AN.@` SkJ**UxD:ugP#7!0>*5(1rig'5~k)؆~Z#RA";\O>SAvUYDKRZQ멠+TӝX=FRA! 9\ )uj2N>-f;'|ɁNM֘b:6 ]vlimcCGww٠]'Ա &M5Ln|) qjwp2J |qV.ydRG:A$b凞Np1UjD@ H+(:Rp~8,RA#ygZS]T4 H+go?nJ|'n<wn''53T~jm6{W`Z XBzKCV~\dV8 5Fs7{{wZ{&l66|ئ'rjh0씶S3;(zo & 'ؾ0hTM*/qI$l8kTl@)>DGSs2D8IŸddF@Ԅ$.M#ܦLTf_d?<,I;\]JV~ idiXs.\ɼy\?t#]!0lB?igTUߝL~[)X VneoֱۗKv@Co|^)n\P65R e909nJ%JK{pt/d E3?]Ehw讗_W@ 8 q_)4 $$O~pȡt%?ԽQ-C ZE#mvQbljeqX_0k34E;QF5,BkߌG4.&++}k6zyPuR-4}x=w҇}ߐ^!UōdO`6Bk.Eۑ>vnO޹;?۫pϨkRf$a2_sÔ23 %"sJPrC۽0b_t[vYa^f/?Z%P"\HʢK6jKu*J~X2z+,OV#}TU_ 4(uۢqhbe43PTfi$f#F~"&f%Bjrc4#!LQ4Yj\&d, Ѹ47M@$[i[Z1BxZU0ELpTHsBU%G #"(NQPrDH pP+UK::}qʎ% ُeӽlJ)f C7[(LeM05wZLƁ#/)Mo"gu^xDZq>&A㾞ͨbVw땁i6>MF(0zK¹t3Њf1R%Y:O7Gya+9tpVuUbAds!%B?;>tg<kN]9T33f;]0fyKFwt!! qV0Pģ ,SM*\"%j&險4v@KۿڛVmQR/)4PD1d(7+)QJP &a& P?!o P^SPIJ=yx' jc1lp9Y&~fZ"7>>/C.uߏ|$srD ܏z4KtE~._8G-lrzfU~>gv:蒹>Qa )@0~g? q:;\r i?9$V38c/0E@[5ԆT{ƶ4%NL>.Ar_s}=i]6:dvfj8q iAPg9OA:Ђ +uV@O6Z iFNhJL! >R2ꧭ~f i_ !nO(w -h5iߪ q]us}[+5ipdjK+>pKks/~3XW:$JqQR W6T ް-J71F:-zUըGyu*Rub7}OOFhUš=&xu%dOU>nyQ0ʅWb*BIAQ7"lOwe.p 'o|3ߓ;1?Xwk*^:9jkílNе[U?5ؒJ*Fv[1ҎFwR X}+ID4pYd!1\J!\+KXdbEOsAc[Q^O7vr֛!ޔ|'AӁ>.є1)=KԢ״hXY4F{E>{QHn4Myeg9Xs cnQҧX߂s?fMO KIfC~ʵ#OWO]k4[*a:e6@7T ]7vʍ3R=k)[ic5{8!\#k 26\(c NEço3ҍ2Tp`i[pfi&C@7Qafw&PyťaƙNE'3$$I)(g|ڻ;FOVN`O3LXCf͒SBE}U+`Pc "s4<{ICZ/ٽd2:Mb'4lyTT>˲1)e@緮xz{,jV(XMȬ߰75'~@ Gŀgv%j3 ]B~ÕWg)#B^ֿߙڛ?,\\--3P\bܵe* h0񪓝o]cPɌ-7SO)PƸvKz5f݊6 zJ]hi⠔ڸD}rκ@HWccRR%-3M%0ǃ[$)hM\]=@"`MT+EV Dwx0D!fTXJ# :j=w<1;Hqq yKH"tÌ2*ڗaLq Do['T*j@zK, IJ>x8q"QiHev q-jpgdCvPӂn7~d1Y-O&۹AOՆ V&2@m|S%<"0z*ϜO9exTQZP5#KJoXˆo+`+4)8ymivI-SMpLbx8P=S'PA e L+O-A-z3t{M0SlWv-I4U(ݽ$z4U.IN)yB%0t8Rg HF mbt!O-AZ-d 5(pقBDw"3žlIb#|),R9u@G~ 8rr3)x}eGm& KAlQE~f1Ԓk 5,̛@Ѯ!hOHP ``͵WdcmLa@{Js $JTBg ^Q%9Dΰ"`y0127]'LjBSOaGp*!Q,(b܎HO^3˂PNy@J[oN"eOd1(JĥH7=F9K:9%Cw7V{@_P5zIKʼn:r 0¬pAH794fC4tB 4C fFA&J􊃭3M&?$nL'1I|QxQ($4bbApt` JeslG;5}J%Br`v]!Ŕ{6q)*!= |d׭tBqVA+e[sJ0RȶGڧdPŴ Z %lMXZ~nL=6^}h2L3^UfV{gW9=Œͪ3s~7fܪn=d_וm}lع̧[KDxSVZCg{L۱u"O.O|N4:v=w@ā!O3gაQYB°K+T>Oj;U(.y]Pp&p5@x(qBȁ;qQd_38hI0'ZȰ-ˣtCbbLb1 *$x=P%#@XhLA\:hAcCuɦ`7͈+bmhT6׉ad*aS 4y<|Qx%n@Hy9M9 1uKo;o˓p7VpRav:8Wi<r}q _tۇDDm:oO/r+3+ފx?t#c*x$P*h@ Jcdf,h{|{G;{1WdwoN^N^A0 ;o9Ä'Ke:YҸy((!zeUr_-U%U/dWU\;4ޡ8?nӛ埬bf>+߉+u!_ċսu3Po"77//{R9իE齚}ibW| &=ܧ<~ %eH7fũ_RИً6 >^|zA\q,KK iHީ+N N0bb5r(Ͷ5/߹, k|tFm?ظG;\g^zb:sgWT7rR'3#8yE;qwaZk1OΓWqѬ|;)BvCYQڼ -<+yV Meh (jrj&$ Y*45+OY66j.n7E u}eYOlgMS(p&Nr4bՄhdqHe<~j{=13Ҁdu˒Z*{ܽ_2$P1+*e>HE8FKJ+$k}gw#֡pgܨȴSS: J{Lt 5[75䍫 Ւ'w0kHz7DZ'"mI-5ɸ3NE> QF9 7yҒYJPЍ})LjƈVˠ1S,:K he x<~⮚kEv9F]=H>zJ4]|Zm|{Y^S5EV^\zGDwtb-n ŧ'gf0d9GǶGV6oNίf7$]6YԭDzgV\I QЛ8#Q/zƸAߌ:x%g_ _&FȉB̲.QԫDž̩ibWV{J9cW Kȼ_E&;u"0@ޙ ėĿtՓ#Otp/5?iwڡ(INՈtL+t3f<;5AC#Iye ޖ)0c(}qN~(KtAkNq(mNh5b >8d ddG^ 5c$fqGQɠ#":&N^$$n ,`]4uM=g\ ?lpoǖa*Z֜m-ι<^e j&PFjƖ$k03`Tժllق9.9MIS:-[ϒXz*Aymz=t;U&ibE_.>K? )ƈᮆn` ~!}e{.G Dg@ IXox-(XW T孜o~ :Ͽ NaS%KeՇo-,[Ia8hq`mYrO LO괵2jia*޺i&E}X$Ҁjt:1P ~6/0 2ZWJ@@6 :1xj"fASHXGI)ng-Y#nM\ AaČa2H!q*)QZo۫k.w @D"hAB'yEv|yD"W[UYX0+'b<]acK+ұcl3wonB>\zNLMO췮?I$iK.{,P8Å:p>w!IJY6RlUn/d+ug1$ (uG .-eF;9|9ց5ɢp,WQHLE/JſZS=Ph8i #H3$H3^{MW> Φp7P,*E`*aDm%w""NqRhDpw7 d_=J` .[28Av[ 2 .ltx Ϧwy>cw;C;Sfmq5"Ls"&B0 :nx`:ɯa+XމKI5&hIaܿ]J'xRrU)wW ƛ| hwSrE` b^]X 9dfDv*0>NKi1ST-Wlr':d3rLU2cWׁjn˴l'1,+<!pz~%ǣWaMRFϳgγe $ݓ`ypxW:gFm< 0]PhMhRoø@$)\%1t\91)g$-hr)h(l) 6|56ZZmX4'aVS!% < [BI ⇏2xFwҘd5|hr])DLhdR[2H3 L#uQJA.BL,OWDAf3:=҄IGshKC 3Q"gl-#UpФNͥ4mgPykXS@Yt:T)I98^WgT bN95v:Pfn9P%@Ϳz'cxtaBYKAFϳg=2zޖvRỲ`=.Ĩx0sv>moj:Z.}xt\: ~\E5zqDIoIvӴŔ|Nec'B}vp? -rWޅEa lVyږwoԠ؛Fi.G&)u?%m3v+*%~BOiLwgw7l:@D? nl]RXл:;zm`~tqy Tyog>كCYݖ{Ň#}Dg0iw!lT'%{';2GEYm\}d! 2sd2R1i!z=܁游t[9n t|q-k_~×Iƾg}Hg ȨQ0ݐ!:db Ndlw מW o>G>%}ٜNZ$A+@u!+90k+1T~gIwrv^|kxқL |f (vb陘!җGgޯBP>9 W7 ۉ[j|aeYl"xor)ntF?g{n>]d\.ROVϽiV#?[lh^w&7/%k~~< Wu_G}V Qs%z}XM&iF]abe@2Ķ]>LdT4dg9y_x I%"@ ,T!I(KvcArImHq!&xնͮ$i|EݕכׇWK|ljw:(B a*։B0WcQ2DSĨ8k.5j&E1y]-X0 B g !n!hl#ƺҊmvu'ʮKUQ+z&s~;!S*~:B-FQxmҡ_Ps_N$QMUQ3t63ĺ5pڀM*Ėf i>Y2ԇ h$=XtE:IBU{QmkYen_}x[rݗ oUxۍI(J?K#G&ǶV6oNίf7 ;tQGwc: tTaģG& 7qDgawaZƧoܯ:zc9 }Վ( =Zd{Ϋ./畔N!X]_\KH_EB4dqC"UN |)^#f\Kn00C̀AVȡXýߪ$x&$ZwqKƊTϷ[E詚?~Lj=Q7! P C/:4vV XՖBc"-YxoՖ5^鼋pk =)G$>qsbn*qGb.nE_xg.r.b,Cɤ籫y/Jrl  EЪNDɩ KʆLxb֤j + f u7a#Qso,ZJ^Eқ U=YzRGh-8Ǹ0y}.,I.]|eb[L2Ӗmɶ$?ڋKʖ, 5Uݗ*,G#񈌰ži e:_Uj''uR&LWxoN.%|Sn7ߧ~l._TyC/_ }3ח=Fד:mܷ??XRӶMMY._kxw 47ˡ%;D/TMe(WDVX["8X#K0:TK5̗|"ROAO@te|xd%4r*=~!|Z΄ ؞t*!ulӵز鋦%2d/˙1\51]H;R^B\ӗ;H3f4F*ޢY//L2a?/΂XE4˅~|[}JN^/rRɄ6(z@. nw^ J?U{%;M?cs,ř]AtkSqs?Y9-@ro(~9װx+#;CU:AYЩhvJm%3ssl5n>i 0Wm.IA)3Gn:q #9e"'0DI Bk3=+UJ-THĊC^eբ6g}e~}x:rO-YG+ ZJJZ UȄehƬ\jEYGZkӷZϻjl_0 LC/--Q=ބ+IM)R WjOP*ÅfX_ Tp[O#''8 }]N)>zp"QR8IL<ͯ9ǩKh<}'O뾈)xOtʼnהG୧1H ^+FjLMOz8Q5RFo:H(pFikӶT|!0*Jig:r=Jʁ?Q,c"^<@MGA?|ç͟dd3/Čݺ':e:Uu% U6 ?ZTB:kjø*vVT(?q֑ulh5(vT]>B7:҄| ZxЍwni;idLNSe0.Ǣ6~bQNBDa H<-um 9壨蔒4mRUĮFT\\~x4M$EQ5*oQ3e׫9 .l#k}'j(i}&(vy&bYCm'<̶MRT<힋pp*:ɀ?$EKm^EM%IpdGH=EUo*%ZƒpB{hݑ!!*+/Uyqss8j'ۤ54_ HDޓ'x.nS{ ӵ%O|BP\wlԆ[^g}dIX=JUrAt(5 zЌ;9uަa2щBԸbD~e`M9>|v)<)cp=i-R:&>V4̟"rEd] R"pXYݪclk[(ۘ2#^ɕ\Pe=VU@$`xp^JqA!dzApRd:m[HSJV8M$A+¹PKVO齎!{)1#g}I38R F_JGLUDiYɞ16w{Jjp3#~.)KQ?C a-(cp&V fRQh+_)#fp1g`N~^#V!PLj^%;*+Cjup5Oqi[Y LkE[yc&0•G`5@- a~UNw9\IqD=T/@~a [m Qn6GfߡTitl$tl& NZ>kl؀`TRip*0^ozzFW i r?5 a7X!/i#dˎ6?gMh =ȴ&>dFR44\>9|X EXA7/;#RF *"EKn NGQb 1R$az9EP%Wu/AjrIHb|!id DMOg0U ', C+l<ڠeF7/`NDB[d=I8Q"`vlQNިgnofW2h8A.7BT/icln.&Cj‘9c?PƗ/MNp~\x+aۂۇ(@z[ɪv`U zc?vZy=FR(. 쬾WR8K *k+/[uQ(˴dz cp༆ lOX4@.Y Y, FZCV51Le=cQRn޴7P @3*hD sLH>aB\cD\&bsL>1z}Mp{_&&'em~=$]m/\y;Mo;s%TzC~ߗ-pkb~cxy2 44]-X-fU86 M'r/)83KP'_'fL)G0w/xRy?O&?ŷ˔lw96c$ߓe𕶗Aoj4,!ZdވL\i8FOxlǦŢU@+k`Z-qWරҺe/pma6l3v{]*Z>Y,ђTSTE)O3~TLC>qǮXEW?$O1GwOjwBRGn/ $| 1L|Aɂ=Hw'N]u&afU;Q)NSBG&nw8k7a(ܨ&Mab^ND瘁 (H'}goi/LkTRT-E[ȩH_~P(*yi`i 2γ2::'QE 4 B.tN%^6P'^:XAuGъ\N?iXk)*wEaג,q76u ݞi801h',3_-iGjOXU0"E)&HL8HU~:Ɔnq^p$<;&{Bòi$t)ndzl7F󍣕÷|"Xy3b̟ob|e|3,uGa]gkR+=&׋emI* s.@y-4Ch`>0MdZ2^P 62u5&Ecj4Fe]Qh>\lb5K+{ ӴɼY`Б6PVHBh\LKHEa0(vӖ_zZ*{MqGMsKF Q%o<I%)ۧ_Nk;ӹK-uV؋ {WTh b[q&sOf2>mHK/|_~r~QMnW_o_.Vo_2Ϩ}3׋^]MN[o?pQٻڦ6eWT:'uEu}C*TRͫX " lL +a`;g cfY!BN blc7t~iO?=8;su:]743@G # X:3Ѐ3#,Y'~^ `$f#CS},$'Bӧ?l0Qdk>y}tZeq"5*[+OQ:6{:Uu+J S kudq\VcXpT#yG7V[4<:B`ܒÍM@v$ r`W"`xR].V*s \(_S+F(B[V@iU=E˕z|r*$Ҵu-zo.C>/)MܾR ʷ;E5ёJ&t$5ёhxw7Ծ`\0/A_?}Tt~ЬY 9j~: C+Z h:Jor 15zt<軰zjdpPԣ^,Fafu!lH Ә 4O*T*qI8w,QzzM}N9_[uŀC,! o5`a^Lm&CяVnsQRo kFJTJ(Z#44f50GHbw|@ i̱pCE B0  AhㄌDX1!$(j"6Ddi̩P-̩xuDFUJU7U'c+rJ8G ;[;3YdۛnF (3&N $!/ԉv||i&>*LtFXB  g8P5V f5QAщ!8Nb&[/˗X㜵"±3c<.0jLb,8b s"J82:a4 (P\ J̣0|Q*҄|+=Y{4diɊsOEzݷ(JNDD4$i:d(S{9` ʨ=3:,;r)(< JS]5U,'H)_܌|W^\.@&F n$Xؔ.`JB a4!pS(vr*՘(g dhpMF[0) 'e,HP[6xmPF.k(\`p#R6P=@t\ 8WS!V,wkR!K,b|z5X1勵چ8(l40 ) j)n8!av. A&0{8 4JsPJ7ͪ"eaLΈ3dYR-R 8ੰHgn'v$eq051$h#ьKn9ILai3j l=?Q_BDj2PAP1*$JeaV:1ud 9&G,+쓼V뗦.;I*RnvdUbjok2UFM8^θo3T0أ;K,Sq(gx:c&%ª#5thj+>zF%D_GaTz3{VW}SݽT1dKLzI/U2+7Uj[#4s>`mZ5kp`XB3J8ሰ- TQ{2T&R^LҖAV"h?ZXAz4hIH# ( Ł[nj6ToMk *nu]5hXaIi){*ƜĢ+Ci2y5VH+wRdk LM$k1^:^Z:^Ag{t/U ӽbvhB[YH]3aK;дU&XGL`HÌ)Ӊk+uu ˜Xެ7J^a5Hh{s䓶Bl=k_bAr'o|؟\|?n#K&\ tpt[4&-tO#3:KLn/L 5Ѵ<aҍy3i%/.ϋrNL8Yw2: fF>Œ£5{hxl!|fn}:E -u)O}w5Qxq&'yfi799jpxM}>GI 6w14O&w:)??G= w~[xvR(/#c?!ɱrSϾnM ׅ;^RHl0|7vO!v|5 !iwW2 @8n^OW枷M/{~ ]S[z?:sǠN6ƿE^\ S,3WSuxM#up!!,NSf499TB Sp|IwξwdHej!^R,yIY򒲢.:oOӄ-삛MY-B(ɽ _vpI:jn%Q!ƩJF Q9 Hj dR_5!|Toٯ   jWB@¹L )q ˬV߅nE9!NJ#|/x/漖Sdt @M0 !(xB@`%,1PD[5}Lhb?X5]5]v^k[)K[)+ny@'G 39nÝTp5#q}M;ߴ]o3quܿ7}qD%Eĝ pnPx-0>VܚjP%!cn/HGBD%ȓ(cJ"%(<ŅCૃQ[־|L`O׾\xbUOJGZ{w|Jgp;K;KqdoY݂]vCT0,Ovu)BT"e$=)T#tЁDq, BJ炽a#$f2@%фJb Qm֡W_RB7S!עҋYJhUWT{TC`^ϸRZ^g+3.j9;LlPc2YAyq 'RI-P--x9ڹ" bc@"d噫8'AH慏!Н f4%)̆u$D:ư1<чO:ư1כх[.TW^,9zYrw'L$v9gJ Ea"ЈBxΝNah1TD`b9n]-_MP+,} 2%&C7tBnvI|p(^˭ݍs~WOO7~Ov_@lG'48M嘘y31O1k-^Mrim%ZK.ZmZ*CK,e<- JH񝉭,KlsrǴC>&kbg%\Bsӊ1)4CswK,ԇ|(U¼2DAh{Hti8 VhzxvɌӫW**u 4^:J+ ,h\ ZB N"-F6F' 5"o4*pF//@zu~Iz%'.RrDiiṊc?nlԈ8xŇDһ%E l$ 8%l\acWRPUSURUKeJem̚*i ]w,`TvbLS>@1>sjzMG+ksф U=q\̱2Ek_Ō0 W=(\iP1Mnr#@U&A7]*^#ōT]2n&y΍O^a\9Wb y`ѐXu<3W`ze6Ȍ㘬4@w.? @5w~yҸR=Iq.2J)~6{?*a 0`?eqwPY]`3 3S]R~)}RJJ.TMʚfQ7 ɴԦ42Dn}]u(&Iuܧ_>y_5fm[^=]%\r6>RՅ% [ԡ.BŚQ?9[]we ]]y`5X_֌-iZ$ ]Gmga<Cx`)^wKb <}8E^fTRޟ<m2zvfc簢M``C+;,9n#7u\|lܙ9;z6g`G}hYؾs>9Z@i{sڭ_}p) {℔u>,^?m 7۶m8|q6"{@Lϴna6ʯo.?HO.DtMoЉ<# ![ o+QaQt Q)s-QV N(\YFnﷶcgX)o#Z@!Y' >|lJכ~)hL8<P8{Ȧ,|;3qn [K sE8>Ǚ1Y/ͫWecmw쯋xۜ?xe.ɹ/5=8vxVҫWWEl7Yaa]_>UJXy@Fi v/Y3FJRbLFa0UdjVrv2*nRʆS`+IpzGzGnX*\R\n``h{T}6Q蜼wY_~9(yÁW5$4^hx3?/SUkq6bfG #ItǹAG%BM W jZ '7υyb˂oZxj\g4`8+u أc蘢yDS(9PTWFS%Jb>2|MDk ~{].k#mРڧkҘJӠZ)eii1W[/JW9'`73^.˷3wL..hy`S\^ ~+%JGqks ~۫H"zwYGǴ?])dg=o?}n V׏돭oYVLS7[נ}^S_UH]Qy#~;dgx̐U/)jE|>y ҢQ Ǐ Zx^[bRL 5B?~z<$ obI_HHV9iљ!u|slƜz>-5WU\Ė*mSLMkˤ>- B*ѐ~c(|˜q-krcwy[%(3,3:cL` /E&Z*خ,cKE!8\v } Z} 66abYޠh%v'|`L.X1 I{Iy>NJʊ9i7bf ˎȁY> Ln5Ԭb%b<ݩԠD0񝿷A-H(N!Dy?(oU;Hk螒 GNvMKyciDdqGЅeႂj.Ц i.P0@uN(-)VAJSfRFPqM, -Fd); ,VNn8g+)&Yc0T:MÇHqNS ^- ًAr!d4چvRwdVEFzɺ\ pNHbƺIX;P>8\P,x*1m{ uapÜ6CW@iY%J)7uAxhJ| g8%vNe]n~W$A1RzUW؅I>I,JOW0LdгU*׃I$` Q!yɇ;'rT3U,xP̩R4(5RYOJƸE(a0dy7*;P1rTF{At¡:&Q]>QH{rSQZPVg)]h ]ӖRBl7ۜ~)* {AISmc˼D{/:KnӦ d[2IXj"ʖvӤ x=28{XCr;]^4ΓOoàFiTdd\F)ms`R[;L:[ ]Lz" J̤ޒFr'XY:W٪fI4&z] <}]nP5d`2a3hr1zI=T{ YsYf;PUڙ; )7'|2=9Z'%8m0^pi(xfqoM/0O<0Fۇy< O0OSfqɇr|kh9(](Ɠ-?lF7@#a9#dQ$wn NQbdWm^d~RC1 d "s=L/?Lͫq dV7Y7L-aǩ&ۆL j-V:Fgyɇ*4Y#3Xdwo~3\g->_4]o9P/v[hdS0[$;%]y63:Dll֏ (D;,,wu~Va{Y0yd֑܍$;Z[V[mfr^6cl% 担mW˹_<@() :'h0u7Y.-11]rCmY"uIBd4`8J62hhކF\h7yy 1y:`hؓ[\Hߋb ,҉Et<)f1AJBz;V7vs7(*/jyICrKozh:w8gs09)]{f/~~sP" az` -ՈտE8Wf&(9o 7!4*v+(c.mpkSL߂;GW'k/_/c~"תWOǗP٣|uTZﶡY{y'_JVwk??{ȍ/ ,ܼ^ ,vgl6Aɶrl%%*Yb*nKSGˣ̞޿D\Axt~{y0Yć(+o?hG矟ܺ/ P)M#&[Lh}| @u5g?>}|#0`.Xl/ Jx)oLxv>Eؿi`⅙w]Pӄ NZ76^W-~o 1]\#q9AfoOf~yHc0j—,/VX.*5ҔY`+Ð bK d0Bk(9X˃wle⿉Wn52]S@6v:҄3NčIDva$2pp1qv/- XwB%s!bu~8-Y\4r^Yvj+qp_oάG+u!>=)Hn{>ݮ 5V@US0Xa_{qoRoL %p' t$ R >BpfE–9* f Ik]uXM A%㍋G8]l[OeEO6rN te6y̾*IU%P3ϖK(lĢX_b󩎑TCr+'25>Wzp TUܕl8+vg;č+.?qj]Q%N@ai]MfϏN?.fR~J6Ish\4 98+yq7w&>iz[1(__-h^G:ZU`H("68*hgY'G9ƕG5Zk6[òStuGuӁ>L XTha|1d~6㷶xfyg|ɯpҝUN[0{J')eS'`0ZFHK-g)M.ƒHa#dz%S 6ݵUA@q5&@T4bQ^[/R /XGbdX3QFz,)94bKC(&6HY_R=#>_Ŏ8^0 h.Y%;k >ĥ6%Q a)Uv RHU¼m1OK;sbT*L|0\D᷸X7AwvFNZ.(#Q5n[<]p\3,%{NÿRG۰bB Y5(A~*61sC*Bt؆BIQ4HĔ/U꠬^xxA8 裝1T|1\om7V]&Iq T8 H0.&9⊅c vFd8 <l^Rr,LZj0)umɞpNݢJ1Va{LQPS`J3EcO0Į)8 *:Sb]) S]h9K7LAX9]^h( PRyauwOt WqgO{K( &|te?_g!bHYR9mLE;)4Ӭi$ gW*~JD<{WH&MAD]je N(n/ǘ pΦ1gUe9Kϼ*b41-v-Iel @EQr55,S/=} 'NV%+ׁEʒ FmI(.ޞ9\!#GhOD2X<9{؛kr&CT j F39G3{8QAgK){ڋ7`Q hl: L;h#F+6Xo. nƨ#+tV0"=qO]Ң KCL `vfEtA?fx0wT6$ǤR#ͳՔKȔU!2G@wDY5qѶĖ\\Ĵõ0r뉫a=Bbd|obWutp^Gutp^WaaDD1&püǥLBXSBQcA^ QNY.FXӁk:ɫ+ {k)c Eg3J ܢBˑ+ƭ7{_ܰ_Ȉg]7e?]"\~Q_k\!0"s*pae (gsHV/:~1| *}NpP)\  \*Aho'/H5E9K%Xoac70yAE [G|aOdCR;sjQtPC$T)b۹yR_OO]ӯu[pP ,Md%IŊ$de~ + ai&[n$*?]_n<[!Dx.(FboSm*~b<7[w{7)i,)7MEnaeIƆ#a}EC"@vxgZ5}I?z@(Lϰ^:F{^ǚC"uHtVi?;ċ 1._Y> l狩)-o{%KuFώזis#ώaogmىۓ<vʵo 2RuYkY=y(S(Um<lͲ$&~Jh>"˺ 9UG}5Wh-$G R_1XRX<6Y3^g~u݃x65>^n]Yَ"=((+r7HO.܇uOv)]|Ƀ߄﾿*?<:̫+'g㵦-me0C<nCNvMWB ۸C49Ţ~WJnɐD 'SUBN-t@]!v5oMU&(Hp2{&BVAnn%+4 ,yS $Ё-GyE)/3Ix,̠NȐ(v2!@pWf"B^mh]G:tUZzŀ!ndIIL)`@)!OZ˄pPuIK4Q镦>sf 'Qc%%P`LSR("@ @tOe`Pt)>S@-%=K_}JzW{EOAn?뱡HK Ð5xT{{MKS%T[$CIճd 8P܇KQAZo9isޏcHh/Ľ֧0⊈$_^#'HzCi׃ $}q zę^uH{pBh^C.))] mu1Y,2ư蠙_>9f2nA$ aF 9U`l[đQ!YYӥVH@ζYROg= *lGd 0GTuBZKBsp؁iYM8SJj y6,r_\ #{I:[rKp3̂aٝ [-ʰل8uO٧a!'EirR})>Sznh #t[TO9oLA7s,X" bDryTք[$HHa VaPaz)ו2ۏ\gmMȪ:*c Ż$esNt&ұAq!Lzx>}*&Lu>ZAbâ|k:c  -!W ]\t߁7 b0phdmArKNlr\r|tЂףnka1 i,ܶ쐷m4\õ&*(p';Ie"W"rѵXƈӖG99K^ض \Լ kJTpX %E$_J*ZkQF85+zeg z1^~QaJhtZj-Sg{I8,oyC?&6ahD6wA]`b 44ia(=>:WOSD{+)2;pvV٦z ͭKST9 aIN9R8Dr~H0KO3RCPukk=^9~u"Z21,}20#B/o&1BuC-"hd+姛 +j+v[KVסu{ U/_Bphm@l2ޜVkKWl4sRKJ) hJ]P2 p'*at"Ju%VO%8Tlod5>)ub\ɫOIyŔ2X+Y(m| |J޹^1k| a JٞPuxSJh/en4+fH8980m0+ug{=;x@^i *un´mf;SqeٝE]1YKn'`n[AXZn`qA"ZNS/w %Qݬ 9jg BƻE]X[ Aڭԍn~9.ݎ!-G:FHSLhxB0m5Hwvz9,"zt7ttA7̩ns)qm#`j[\Km马dSDݺ}Z Y5UKSŃ\8-X[PD0C}*?գK-l\1\; >\\¹Y|]ߪz84wL7#"\bF2jro@| ĥV_ b.kR 4 åg|qJ *9r< DZ}D2,̨.޴FL+l]4̓хMQquFeu5'̴WבiץmXs(:C}h3Ff;*TBu&fc!Z.|ھ v(7 k ݶ{*o2a?Ϗ8^"a^|dS:r7<iK cg9K,*?vWTwlGtQ0VVn#qʉKXqG%Ak% ˜n )LW'.+BJ~5Gň//7u$jћkI+\ "z{$~ 8}!$ Eʵ=_SsG2/n&u,қ#]6yޕ`RX˃nF/dr_0jd?}̯wDY&Mk>=7|am? `1=I2p/{bVd|?z`"lt6KPC߷1 }hBmq@nB?@#ّU{ܠs'w~6p(ExMQZ.2euXU Ͼ~C])f> [p s< Pl>.,vQ"$Hv 럄ah !Ts>l7tGŌn~<*G{B0x FKOaWQJ] jx9h#etH]ag@Fkف qxG=/9QcmO&%ԃP s?|.$=ԉgkk ߺ󽥃h`}\ò<iFiktM.? 𛅻C/M fKa l#[iO Xn081lu¨o޻_|+qA[%g Nd\E">ioQV.m͵PY-7zn ;k]̎܃RS/f!"B[0>ث=ef"ݾS@pnVY7o֝A1 ͹2aHRZ޻(V_FnrN1hr^$.x7)f/#09fS`b!&RP2A|qG@bZ#`juuvHk+_oBk32ˆ3}?:=yy[5dQԛq ֪}P@GdAT88H!kBfp?N6s?e~D"HG 1DBf?EW2 XB+B"hĤP _ *#_D aH#$ӝK9 9&菱:ܦBˠ@Y9>T)\/t`%!xvP◓7&4ʹᛌ IMe*Ea'QD>(@ i{M=EJƐÛ_kGܥ9( eX`QJ(ò:QRTL)T@x <hF5 q a(!<0h/`bH~0_xExxE+:^Q= B"$/‘, MB:p-#n F\D?XPeW&rʻxu/U-P,[Ì:O%" m0$ hIFpJ#C%YZA ۀ,_ߠ@Ns2C0KxBucc8]R*:VQ= }=}) )}\vlwBۧ5BT(;*ѳEUYEMn/'}T*FX: XEm_EY7QN ̀2aDcfAHI  nYT 3?a=̬}(`a^({&cw{WsX\N;ȸ_+l)8|a#<1rһsǿ %4px)u qSWlkІP1Jr$RrwkD"]G02_UW❳~s3EԇK0»haAvfR $j0?zyxq1N2g:ovc2*k_,&6e"uL-SFarٿ60V4P Ń*6P' Q-3mJZo+Ǥ1-7@^QG!㾐w%S& qR 1Q$2!4-de-x@GOUBtƿjGJJQ)B$\KԠ%'mKWg-|t!DAh#ZYRrjr_jmj\_E-yʀҕDD mb j)P z#%H[{0"dK:&$gԱj`|T4ZU2B*Pm*'}ĴBi8l ګʶ/cc9o1a7];dN-~vAlXz F Qdk9ڠ`j0A%ɔ)BrٙiOl8 L+[ ҍ[F6m6l}. &g=6;t9ҡoJQYtjo!p`tzY3J%\P \e/N"LL91 % VZҹNktd'LbHAydVQ@R;*pL}ᢳښ(yAm<e.ֲ_v_x޿{¹ȸlT^^Pw%l%TkyDzx_ߞta%[4Niy7ʸ +%yGOt0] +wNsQ!wO77LL2enKJ$/c- X:!=!$\ R*E ;^+& bq.WxKGxQE>$wL+0DeͰ{qjm+p!RT zanA[V(,ni^1R Y!G]EK%1GdiJEwK[8Cz;K2͆6 5;]jvtV"9M[E.ACJ8*ΰ8AMd@tݦA5#³ v74!}JvVYwq`La=b;iWWsd+_մs~ߘpc7-/ëIavY+hw1!Wu>֝^t>nWZt>n嶝b FPLoҥk'p.CpE. ޱk(F7yY\?oYy9gN]K2 ;h),'1Jx\64t>HuﮩvĻO6Ym̫)^?|Jp`+=`Djg^?\yz_]~!,L.f/Rs m{ 'g/P*^;I@.{-<Ȋ7!$8irrV 6ߝ>6Z9 ({9h6p ⺵Ev(D$ CuGKqO'|3ڧ!T]R$ku6qTv莽.P-%RЪ="[:im7i7=m6v2 @_y t4 T P6kuU<=D8OgT2Tb1JI5ZM?Y3v}pɑ lWiH%5~!#58ʓ@ qy^I30gvb\[pnf@B-H,hp+#\QRev>˫Zϔی}eyYTitI$/d+lx+?JC6j_G51U39zոr>bʑ(몦|]gXVUxS^xЮN[U9L"]\IKFt`AjitkIV5b~^ 8^_Z}1 4~_:Ѓ6ox!CnF hz8mzs({\NP|II:#oSFZ+V=_d4Kc5s pFHy(?aOMO@sN yg#e@viopN4yhQew~>#=|kCRTϚQ9=vS"@-868 mfq&6,7;cgvȑ2 7~YewЍ靗md R}o0K%eTR]aYR3x" 2sI|X cLWSDYYM>%c̜=7?IxTDr2`41EEd^b$l4&Y&߱>`7=siSQ)H PHAzC)&c3%,ٌ\c[` 2JtZJS[ ({0uf7J RʬT.ZOd۬7gSYBcz5O44C eQ -rj@e F̈́ڇ( Уq=z~Z,M@>Y2%?&aU Lp 5'1k@zgQ#0LAA.ФP |BUZ"M4 Y/ i{/?JЋ 436I SiFɜFHڌH R))T)(h̫>afMAk=BkLSRUD!򬩹 B XkE(&Z0S@_k55}5P,557l,Zf^4"ȶqw kUB.rM=env*%4ǧ'lї`sl<=i}l~H/@qU1^џO|FCa8|uuygC3憞!ɘW[e 4uwōaZӨ {St sPWFU0F]u⨪|4|%oS=WI:Jco_oӏYE ox!CF gYU6vbӏle#Fål31b(ר8($ٟ&$e>0%=κ44+ *bC3㨼X| 5wg&IN1 ߝ\]7~+CI=t Ԩo` aؼST4nh DAI&;R~5 %/|2yғHi VsBQ!Qf.Fg=D"pMM 쨌X`&n^z9ZTBZZMlJ[Y CcVrPp5͓-[8k4pOz!d*˹GZɄ^o96B Z|VY`H2xdb}4[F4۬"%Sи9`Vs6jHJ͊ y."aGIǸULS(97ƣ5y1@DThP Q*潫5m%Fؚ'a zzzg̊{9@qh XԪ.M)*H!8ԞHhf] -Zd)ZFi8ٽEF(QtJ.ua.SZZ7"e-]ARkh7D&|%1mV )Y`d|5c 3v߳X_GBph۸'@k:Β  ;Z)3#7: ŕ&x4 ,5N䓣k\feg#ûQF+S,5a ʁSe9R%-lH+rW`Nq;#w5P(+8^\ټשӐ}O?=<\~֭% I|{{!q 7=}.(J|sw߶/}_૦/uv侇qH$ \nVVuf1zAP9- T*p :PWlaخ-xR185UJ]mxס|VFszD;Dr'ӈT84QitsFkO#V@dW5 꿖D_"PM dfg9_r~-4fl ?\]LJiLT/5^dkxѶV \HVPFjlrFp9S%o gQj}Bz6ɿr~l|;pԇIlL!y.6o#˴yd)^2~}BS TjlVS^,M%2cLX;om@ʹv }~#x25ƯdȺ!2#BA0_6jYvLץeɀ— i27!4jMc.M`Sz8ih"$:SP z޾g1nU -ֻ;v8f;l7}ťH#+Td(Muwꘓ*y2D]4mYl|w֠7 _TEAZ֞9;Y"f[2,ǐqDCF&|o{N5* hQY:Z(d=om6gfޑTh!#IJԺ.!/v*Tu&qu %t/lі|KS%ܢ+NG2h`hv U zG$ƆRւ=%qOfqSW,M9Es+z(D)[Z\rM Q]*tW2땵n~caG4U.[4ؿ\.t>r/ V]:tB 3U–RvBB9Hnnd^ypNw]T2.yx^wPd PZAb t-@'jԨ0MWO76#GT#1^-sGLTzO:^ R̐ %=u!bRP|a3Xfry5%8|&bzz#tPȣmuˁ3tzcJi$r1L:#Əopn;V.Br^ zX7Y<ó7Z.4ݻ>kWI0BfF`2io%D#v@XqMc] \͹ Fs#D7XqMQi҉ŋwnoY:}zG/qw}mlJ<s=s˜U ys{I2#Pu~bۊ"^MPh| )Ҩ k{c}%pZ[%fYi`[Zt i~]q$\OW6YG3f /b?Ȭ7k?$@#qZ␁K#]vt)4Ub﫭zIuMǠˌ|ܔڪ_{TLk&F:zs>ǵ[cǂ^MX0͈VXd:jDdE*-M+slel_c&oC~ 91N'([o'O<=DzϏ9׻x'ys?t)}tKor;49qdp>CGg1a G\ǚ?&OS˼/|d<:G1VXZ1+5C1Id?g;Y3jٞ:4ZM ٖGf5jcx[um忑͑6g-\OTa&|S8輹iVF,|ȇ OtZ9npk}귰Vs r3/JU/N+UBR3(/P  ٞrBYT"zj-4j@e F̈́ڇ( =kݠ҉I7R5;ݥrCSD!KIs:A`/̪H̶ou2kLz?|@/i.RTF:m]KS[&KSY T==JN-P'Ç|%&jm=;>,cll^o]lPv.}6(+#GECʼG~Ytw13(5-j{}ɬ*9&,ڰRd 2d^ rR%?b%NqFq9ص^1@{O[t7(^注Own"~GsN㳿?o\xPcr|~Zhoף?3!nyk\i 6#%%+XH[aCl2yRo6 8][m1OTۻE TQ=/Ùbcn6KЊ8n?|) =}Lr* ՔOr#]fY%ؔ/Ɠ& _,Hף+s} kK$;aK!A$G?#27)Ql~pl.MA6 |]Gzpz#Rߩ.GN0m}]). pNK#[qb/0OxubFt}d^b∔*'μYv&,nt3Dr{oX//=m݊[Z:NOB  |Us^'yy;r %݉F1Q`cri/b"ǫ}9or`#̞XtI\H0:BS#ƹu~,r;($>| b.1L Aik [V X6FE+ K;^K.ږJq2"S PO4x b":ʳ`c1Ô"kݬYiY*cm^ &K/{8?VQPn{ݣ^$mz`zcz}-Z#1~,1a Cy},HL!M&Cɍ +EoʁwaqFh4JOH+*Ks{F'шeZ9LY\ygܬG*Q!Ŋ8d.'gϵ~_vW\FI\Kl q?Pk7Iˉbt,6ȋGUWw*E30=zK280Up؝XnkZi5Q B[aZ'Fިz d$(%P0,Zq-`dD\$i 2:*'-Κe @8 I}b!ŕU`س+ `t Fe[XSOďOh?^uwO?-L/^',Z&΁qO޲Z*@㯟#wl^cZ GscOfU*=_swoOqUBOwOw=a /I]E}PA#J7 8czESv+v5*D.FK ||>WfV $AN:ݜ%wsYr7gMwIh0w~? (wi 5`P$`L* hq3ʭNݵNթvzFKў?|x3':,>QL6$=9!Xl#rUANEwMl@ZtѥA^^G1S̜ܟ7SlTQq(M# [/PyLxUFR8?a.wHMIH_fjO߆¢ !E=&w0a Fa'q&)9`驵Q%6s !ErtVYPTLQܺz$cJ"s2g6V$SМ[捐ԄL9::n%@SciO+"7֧NX`g:gzBy<1QP p!b,u@NhȔDcFdr Rߵ)'f X/ƃWкhL+ .ŧmzEq̂.hFh,L H+x]o,^keS];>յSd#<,Ef1vJ\1Jg8Jr+)50υWFO} D[_DypMylg[-Nѻ>WGnؼ=OE0/A|rwYj9+q/OfW}ퟷrO9U]WY]x}n1Z?k1u}^>9O+{oqIPCQ^LV&Wm5qOeP!5Er?jp#j=/}rTˑ:K#ыZz<@ψ_l>9zFO9۰\\p3;ܗKtLS %KFP`&vѼb9YTD*PoT"Vs9L04Y,HSdY ц6"(.A.9CtU%cOws5>(别qt`J~7!#n aI{#Fi 9E!!!$ 6Y@gi t@i !S /"D-WAkP>MIi];6ص֓5N ̬%v2(z!5sX@s=V<xw9x1 VKXKa@񪥞rP^a:B] DLpnw,{6Aqg€lCqa<c. XF1Uj 0qi¬BAD$55_bTcH7J nF_naw1FWS_QVz6nHp Kcؠ`|T`.%DƮBr"9W[R`EYFH[}`Dhu77}t;;iػ ƞHX GɾJ}][x5c[&,PaqJ3eȫW@'жU]KO)򎊚P=Vnzb(zMFcqO"1To)J)?φ˽$Dx@sHW [Uvx!}oY8%(8q`,FGgF5.Ui9FDGd$[ݘ)tU H|r}#cq+B&&V6G-D#1\6Z4\vclT~7+xI3 Y850//S"$PSWyendnaF@>^[ixO9qI}kx{7E`X#5T8:!nڛdᵗ 8`8C %q1hz.nSW̤2ἰڱjOqahl97QA>9K7靗Z(&t l5Qlerۊ)*hr+PcKAǶr=69u^vh,pTDWE](?c5eÂBŘ|7$+k|v {8b{^A1'b8usMiU+MŃ+`uNJCDY 5fQN_,L}5E{,o˕},|ʙpj7x_5b_r~>='zJl_}z] {!eX ģ4u~녜Z[G\ί4׿__e64.wKB>.˻-_k&đD uo+q+9B?\mo< qֈ^{Z+ޟUi-0BI:ҫ pNG#79T՞>VGNM%m"%$B{9c@ bq);+IIVrOdH EBbv%>hK9*lT*!$F9Av#9\Y{Y4^Zg)?KAY ϚAf%2i>ZQL-h@ LBCH4>zŽ3sk/uS3,f)5K!Y3՜Fès(`EWP Ca+Qnp ѵNkm^mΌC/q3}F<8Uϣ0[{?oU% l}r{pCկݟǵ)%2Z TebFַ+y_]NWY)rsPW4]o"\Y1BsTh [;>,U<9_@vZuedm*f6yⰋ2'4~*dDsn W2j6g `#O,MΈvJq[8[lzC^(aSk=1.nqQ\=_6_%~r##ď"`%5}yx?cYXcstEk3jtҠpbɲ?ٞQ}muҰ*4 i瓶 joU3_/3̭s3bQx%fW?]kʙ,ǂOLOvkgkmS(OSJ[xte-=(^ay++$+LH{(vmNxX 6^kL87~, 8*kmڽ9LNT ^،&DzR:L2x3>˹.$!p5/r1UP %3ڀr"D`(F4 1a R|xH.j f*%[3+9M܎e۳qmB3Ǘg+yNUq}zptG;ʛ>ӥBpNj*yեD Xͥ$ "+ BbSBsr$YK0j .Ү8ISyuJׁ+P^ N{} X?nPFGWsJ LuV`h(׎U(j*fXՙ([D(sQ;Ir)wdvߖ٤\.J1fb('sQZ%*Bj(՜+e01ZC4M[` %2ACfą-KjU%$q@`n HNP/[T[ craꆅ*[S+QJ tw3ysоmNxI:%p A_:Vұ <}"x yV0j ԒԒ.ƺS7-*2^=ϮwbƯ.6ssts}#P$oK-oWqfi7Ed<ZZ2f5(0"E-mG5WY !r_ j$ p^^bڮP3@~1 Z heKaz>ZKUuЦgGQ7QGCD$|?ƅ+!۵)B!y^ A% Re ZtR3hC~4 yf ?P: 5Tߛh9.\ՎhrQcW4dIz>w\ wW+Lae_95.3%h)&eMMm2/kjf?DB մ۳x>F4m<&5[,j "Z+fn腽U>o 2Y;$-YF΋05۔ !! ."$OՕ^Q׈LK"S`-tۋ:;r~P&o7PfįU)UTdQv..%qǡ+}i \)4+Lyg T!!R2,ў:X!t-HmxZ/<\˹8ӳ;r"Љ_Gynr8Hz9~hG_?}< 'Q;|='YχWݍRNի}njzxjQ0cgŹ}/?;}?|T3n%Og4\d۫("&Y~FߛkNSun ^?g .졽Bt"-ڝ\?׺ᛩ ǣ{3i_Lg`\ItefdY4E]hȿn+42m[~HF(ɋk@wp1:S29eT=6I?҅'>,` 1wdl-.3M>5!_CĔ>]4* gޚx Ӗ @fh2^)f;m#keH5TEtk ;vM"6?]h@{?޼+ d˿}g &Tŧ&ŠW`iR9k8LoqQM 9<,҉1/v$ 3 S I"&)h=a|0|{.V58fR4autnd6M KBt3 &NR "|ѱAN1[nM#A ͌_XmrC0n,V ~y#U\<44!%'8ܬxx1AZli^:p(BrǀQ0Wur93B*)͝kHi3l)N\៻bŦE4̞a'E:h)JXhյ֢\r-p3N4fQlBqEru魣qQDbsT]Doq\h.F8ǻNo5cu&Lz]^3kz=LjSN˫zk[Qtȭa逗rO+0)rF;^CFGj1G#k3![ ϓYrn)>S8*̈́m. ]/Q__R8xMCNo{se}*;ٛO^Y$6J6O"$،=iAҵԘE!Di_*DK F5 $v4?^qv, W3Ąd/|ؕ\23&>?w22^^Z#$0=~ n| "Q,kr_z> '3,F I݆z7iC d sU/dٙ\BGAﯞ/\%V;暋ߛ@!.ڵUHZm'>B&Aޘ֒GWC;'qbƯ.n7e1Y %oJjQXzDHV%ə= 9A5(ӜK5Nir@(c8=;Wjga 6ͪ5i9 atoCjW1^BT~ʜ3L̀`󎦓tyqYnR14]Su:/&xr1_ƶX O*򯅲$ *4bAj|˹fG\{WFBfJƒ>t~^21j[[z$%7ET kQ/dY #FРciD1T?b:JЧ#BwboqA9|_{RԼ{g+ήF5)*N^/F|DTY%5I$g1GqI&q;4U2ϩp[IܼF;"q'$Xa)8ɭl:ŕ py/)lrd WnEĦdj#qٚȠptrRhZfu2q?7JИhIA_ `[W}DgK,:E`2YU ~[eP_d5VRbEtd1%(\ Q$C>WIGC I$=T. J8 N"9I& ڦ=H5sс:F򉓷> Kϼ&W*$~<`5Z~"͌!D#wBD.K!AP[?"piA8'u11%'"8DRK'"Z'JZ,kZi*tXH&;UeO4ӜO2Jt+^J<93A;4b"LjAhCV;BRH JJ~pJc`K}948(9*q5Iuh\ @20q!F'Xl1t!y/Gj$*H =2g%<(H} 0U]lh7DnkkM(kٷpM(ٷsbըcCfޫכtcC"`>L'Y+MU-čLH}_;wMD`^]Rk\(~ ֫H7&?pZk<}a E%U,/iFDq"TbqTps'{wO+իYU'yp7``]]Hg_tL{]ml JcndCAF9^J;uw(+-0kEwz~pu#ۙeu,,|,dbD-3 CGw06pFr2bʽN :bY.rӭ=9{P Gf?;6W-e$wŝGwZ6Wg.9nI1GG6qN9:=f2Qh?H9:=ιCp:f3s4EKo5]WkQRؠkӒ9ZצekMaBۯhMh{ r.y'KhA^6r8P.Gdwm|W֤g >?d!ߕY:ޯ v+@5$}mPp86ӠٻѐKSEv":f,ǽE_JKG}]fPV/ KǨ [è/ $赎Aa~cPH+csW%B1?,>=La0~SLT2f452=p0B9}2($./ލ ts/!ll 9?<`A*t? qc_ۛM׻[E!ʆ`0 tO8(u8yDt3gaTR6gÅZ=fF z:Cϻ{юlnOot9/d˛/>Bmg˔]PNoWN8!*FHu~v@+/l|s- [] C~œ>iRdLd)7sIfܝDLf{*0Ydr-ݲ|ݣt~/^J. ]vH W!2:J BJcm qԃPvNiMO 0.2%Qkn4Hs^鴊Ũ=CpbD")/)+3'f@ſsjKJt5,()ۚ_dO e 6!d'4Hk&0DJK|nˍz), R6A@4'K E`~. WGJ[{?w9Xs~߁` 5 BXN[,0C2:“#cJ9s{7Ey`7,/Rʳܐ(iDQ= F(J *tۨMT u'ƱrfZK$('<.FGC4~Aڊc1vG<5d uе"Zx."!{ E,WJM!+'MNH(Kq΂8XE8c8Q CWUw݆B!,1l&?Th6\x!BmpWqם4 %~@iڕkW _K1C"=;' BܚEn_~MwtH'.z3r1׈8b>}}s2zJQ(i2u}+/7 hxs$e"Fb ^i/[j.TNR3C8be*>h IgF⑪hgX$N +,'5S8NcDGW*Djc2UD6^SP8(!F_-MdbS_>+5'ƭ$Q&%IFprW $KAY-|鹓3FU咋4^%NQp:dR^8E4'f6p39/;0_0 h4E 5߈ȵtqA(P#Js^2Өz#|I'ZJ;,$!iN RRV-˙Aȷ(&[]`fzm6#OޣUįY`iFӺ w JLx}Q{t&7˞SNN.-nUrvxU؛aLH\s~ TCDg:rsKVXT",je Ǣ-eEp}uaHO2ZB櫣Q]v s*7F-e*ul``}x%jum`Ւ]v:E[%1{a R*D4A X-K Rial&+{DV>L_|VO360Eꄗ AK;d-;I=/ 龝/٪UzUG@+~mf{cj^!5h";= ݑyk;0OMKzld~5N$O4,XUr6 V+V3݆>~D_Wdrʠ?_ܔst~ș!S!sD^08K[rߝNy6T utM4"Z,5;ʴ54J9(U)#ůWA& l'v8ؘ8L|x.'B(CRi''P2F6 dVp9+na{1igN~0k2G3Og=}f7BWo kS&bFTt~7աclyɫpb ;QC[q$9Abش^ec;]f)Y6|z0͐_V${W+0Xwf5:iHY*89y=)m3zJOsܰN"Ha[Z\zLh31ly WjYDEȟG_Cm$^qD iF1ɤQU%%Veu5#S޿NPu9d ΰ</wwYLޟ? b=glٲcz|n=؅21:Ytf<]rl< %^orsD7:-‡-^[P` &]ӦewU3]=nQw Pʴ~|$#q2j(yA}ѷk ٻ涍dWX|ٓ=eySV\qrR,``1(|R  ąqb_tO ̣n]~-Tf{s'"Q3⽳ ox?܃SAsekY@6{1nIyw`Vwl?aFwe|/FUih@:7}_]qss}0si1̲OGq0\\Yv9Nbv;}̽"?)4@5W]RiЩEt>qyeUb$ɕAFY3mQj(#k:VERcBGY1qtH6E=;0ྐྵF= ?9 ֮FLX@,E6`Obv㟓CCPkʼ].2 dDp!yĒE)w!zWMV<u]y׷V>kLfapg}%jQ?FE_Z|G_\/>d-xmN?a^@ե"{Ⱥjq p UE|ŋ0lX[l'26bdy#H۟u~=j )uEM<3?QsRUs)n쵀ꃉf~R#})ɥn'h_fJ~<~/#13ϑId(ʼnJXMT]i3r5]<;SҏSn[+gNPxOVpF$8舥ZF&IHƩCjØ :W0&Q{Sl/Dêp׀;`Ϫ7&kS!ff ىU݇vpAW2]vgF5~8 MH1)Bt~ahwpS]>B bܔчCCo3#  d nQ"Y׺:;= [QתSrîfV(ȣ7y}ck5ܴzP\;8+j-}7 ko\(@Cfin 0f"bC-'S0>2~pk3H8R҈IFR@ʐ MYp!{M $D*$MQb^ED"qbsyW+d8\dC IHi-6_~G $(J3(%ΤZhnfC@Ms9PBə}gyPAK"7qj+X;ƴcv naKve+tSG\/hJVgfY}tԱ0HXYUqR{9AҷC+r  Я9wg)2"`su-eR bM3N9SwΜ +'oP꬜C7=3MMۊo'ZL ;j+Č'FE }=|$ӈ .qBeuVExg'(STA]1];X0n>b? VRklTb@iՅ}o+0}D)¯ZŜsGDci }@/u9E< (i)-q˅sE!҉ /?Ff+ѹs*K5cj%2 #K&vUY$(Kޱҡ3\ʾtn'DdY{x]8ۣP,xWu*a3͓S>hmEogT;q)S* [7'hi]Y 5MgPnx/Jux?#{M1?T39S~b3:(*l>d'baGy@ʌìj}/n.29f-AN2$64y3>!9tRk΢JLgiVݏo}\y?\_dD D8,D8Uÿ zj_A5{ K'~alqT(SNa*Rm/"#$ŘJ#4N$$q(ѠBZ- +ۜ P_׌5ҙ)'27FQ=IW Lrh/|:nQbOx_W:ׄ!̦|UN5܁NBH+rZs@(k r騚t6J5KӈYepYΥNj!vyUXh VQtRV5 A j% Ư>o[4wG;9{!H..sFi]YfڊuPq`-&(D8JU9zx-[לdDUӿ,,\桟' _E6[]|%L'Bg*_X>J5qNA)S|u3p~gd=6a& BMjIM,iJEP 9n҄g ^UTiW}x!X܏}܏}8[p'2RIRlG`9M O#"b+1XGU?X ufMQ*X :죌th;}02F#م8gXY!2+%1r)\_g2 2H' /2 ^hUTiU}b0aY xU`mG y<ʚA]*[]ʠ 'ۏA!J&MO`žB%]RJaF'J)It}T m^յZ=C//I!D$4Վ Ntb~%1` I%lN Ɍ59)ت`#Ԩ劌] T4Fk ÈL`Jp+pL=)9P1ߒjT)l&a M(Z,tdөгKL%(+bNTDؖT: %Xʒ8)x0`IЩLkӢ Z XQZq i[Un6O)C Fu2o|[x`V2q̟L@5-SLlͽd@2ň1$͝pFd4dW߯~h?gIa'zO7_ѱ0+nSԀ&"Ykn ^IUݗm.P`Rtz9yvZMHB'G_^1:.!ql)y J؟bb c̈?P mN5fSXD֍|C3 ٭q ^ Lm`gΗ΋aN*f5򧳊ֲSEΞd;E)4h|Ke,o5y>[REJ/vOځTqI0/ rzidǔYz}wPAQ?&pYp x EIiSmb6}cdLt?UɯϦd)3c7L?Z@є8F$V8(D"V V#|EvagWHƈ!rcEP8Q3f+jc&B1Pnw{| Gc'mAyqx>H`2'9>Kb~vSx?svwd_nш`9B#_s:xПK@@ Hh$V") SQL<Qˌ 12Gƅ|32򍖐e嬽v$e7+@.xF!T.3A &KV޼u19.g27ջܴ \ewST[܁hwqCZ!TmtV=Z= UZ^t_H.qhc+-q]$maIVfG-Z9}xb*36mWW|2ji#XTi;9LY.m_nu{ꮮ9Ӟ24B?e:EAu}:'EkZWOfEE3fE g@WI)_O)  eImEn@/i2VNaN3&68 v)Rm-W\b_{5!e1Ҫ8? +qv|4l6bxֶL|\q*&:L2ZMoW杊@H*B ;D- 7@ <#Iv>9wL~rc[YBSFdU @ͳITn|nrV.=`]y{j>|z Q@k7kmXz1}?9['$mQ4ArYREy%%%yi>$%DQ{*D) ?C*-*z $A OE ⳗI&(r3Nғ(TNބt $K;UKUQrlZF[|0#'UAh. wu +"j ffGgNF4dE:͈%A0hjF,Tlv [aٖVch=FVW]z:_f;RHPy/Ny`cㅠ{HM/2y6{FM.}}L3._Vk1/B!#dA%k>֪6T5`H`J'nǍi,lЦt6%+;?7/g/zS|hެYQl[hEkW)G+rKjh=>;CBխTVcIfa{bS(!5Gs k#aTpV$7l~ !%A R],r[DXI( JtH•2*" &eQ*J8Cz:/CPex%Y04#kNrG&(Lyd&ĝwLCכIh쟧&A|ןN&SE݅^ o*k8߯I${2/<CmF$w{x\ `~4wga2;f>-5F>rIz4)^H% Kf_} Ur2>Y[A-!-gUpnSdw:#ٽY0sjg>Y~}.cw+w??.…&5/CQIO; ?~*b g5LγUoKn 5:^Z8upUyYuY¾r1j WUQ= ue[ru@*" ZaCAP$#{FᣂI:f\Wz̬\Ĝr( Qp]L ~H\%Q<ř7Sٻ5R3y(qI $A!3 +˄MlDc''30Q /&U.V+#\],e8E>$dׯ_|l)pإUs+rĦkPb !yDBǃr%j줨 #)B(BP4!F80^PE:J8jQC<>}~;toIw0A`^vzvfCa>|4^ž>S(ׄ7f¢ 5b4l.^kv'SFdzx,d}Qa׬kX 22c9դ/gBٶxz:^][ -{VO-ᤶl|Ә+%XYK$К1 T ,l-;tQ+aRd+ ˈ꠶̻{rxV +5M46Qf1DDs)I#HZ%22ZQB3|oti6rXvX5nXsê.5!!TSbV$LŊ'^m dd1|ly aw]J; e =LvTl_h(Bh."&|G-Emu8F.>7\TwQ-͹CHo]S33|"FGk^ox>Cx;Y S\ bH2`Sb5tȷH(}t{d#{砐#P}á$$ˁܾi;(ڎ3_sZ63+D`PX6i pƙh \#y#tG^}pbuӟQ5㪥GӠ+BFTM:.َf!F?!F[6K?~UeK9n$PY.ԃQ-< dI@eEaƞp47! oM3t YIl0$dPSg fMoo1 ǗG2{2J,@ң ^S~Oq[_7xJ񢮵s${Ӳb]G$eg-{66tpU=Tt l޻N$; U/;ܱD*ĉ>#P,ZpnLW:8)\gc%<0Dxv^[ĖkNܾ7ơ5B!C$UBFLJlDcȝL58IBG .(`ݖBNb|a#1@&&Q$៘$F&Tq4L<Q5 I'J6EEC€,%(àSdfph(qZcØFb1J 4" kƨCdK6r27Wn$3բn^X"K`v_84cT.zzC$<* 쯃P/+3! 7!w-kG.'1Kt޻]桊 Qƛl:C4T"׷2A;HnD{(pL>C, nhSg#TY{T32M!P[`ҔkR]PhlJU).Olh12}w{ Nj)f\ a_w%&Ἐ˭Zy 7i`?p? *#Ū+Ug+.R㬴(}*a;dT*V_]Y>ZvuxŠ4yIbf*bE˺ŞK?r^ 'C 6~I04*!&SL?CD& F& "4"&(146ь7R[pyZGp)]쀷+iEІw| T=<{̐@-cwyQL0>DO??/ۣdfsS=.C6V^['t%U$IkVO#9ÔǑs>0xAQCu qT[P׼x)ǵHGDžRX8ZCq]a01!q%)}Ν7PVɫgq飝zz8R~V7-f-+Di3щӂkL8j%?iKi"D4+W]ProCmɦ &yU9򵏜uuHs ׽( :i$AN{X\n"|CD]NbN*̆w|Tvd55K }Bͻ IˋWBfd/F|fAxzR>z8 IRPgl;,Zښ=Zexas4 z/о.;ߘ#7Q_|pZsјq%s#)}N_ڌƥ\%W7[VBN)){`svo)жpO'\>7`TAq sxyαHsNCz^xoxB au1Qҳ|a~[f¼ o^mix gTa1iFkxϬ|摕oVzXdY"Puh'CJUKx+ 2&;=nSUp~yЩ ii&hI}$!07p~Ҵ\b¼)q~$L7pAg;,&SYδm G„mZ *Kȉ5! eGe1U%Z˱1(ؘG:aV$ h!$55_]*ALՙl, !fQcwHSMŮNnkUkWr^+#9H^6FFr $J#3XYƥ9 IB! JƂ3^62k6BLJX&=?L9 ?{W6nʔVT]n+MdE=YydSi3$(/̘C6n4 M稓 k(n<=ް6,t`h[ @# /Ո1nzaܵaN@R7brJ=ĮLj.29Qqg7͙h=yn=<a7jKg's Or-W?~Yж}؂ʃȊd?fc=IkU(Ȧ\mv.ecdwdu/0Mo-f>SIFdb&pRBk)옴u.v1 le$c&7Y* BevCs/L|dMBh>$Omd:s$zA%"%`2I%8;yf,J\fSNa`%oG(K[Ьiddt6A9wKlzzrh6N'A9|duA1uX]Oؐmy/g]:ye\7^L/![VnjTX8'`v>5!jjE0d"P$س?J#ѡγV.yʜLx2=̙2 %=.s9@P%tZZ9{<%g*v[9/܂<-Z\j3gg >3;uw_9K+:vlֈX^\tWMM,`@4ٽϩ˰+~4N($1| 8K9s|I BH&YR_'dHJUH1^+*MڒJ\YJ4K+^I-* D8%!|!![:ץ4D+]Pf%yUH'F张Jpcs'eI5(J*c(&UYTZhrA# 9S9c:bO8SkM29a,H 0S] PQFeAmsD OiXgů# BCQ4~up1PToG"4n{  ݟqnqE5ɣ3Wnlٝ/׋{S~̊r)+0MdS8΅#\8΅#p fu(-U#+^)G eUeU`*(6RҊrl5wsy׫7r`w]^ݽݲݹ°rkaC}͇{# &L^¦"=N,(Q2Jk#"B*A[VB9LX(Q+\…옎V8` [ [J D,9Wܲ $dr !,ı>euf4 mى>2.-s FYUbJН=ZhRr 3(F(PTUY" =)q^LyX~}SU #Y[ \hTg!vl1,~Fz"dCC h)ZUI)b$"Ђ_cً+ )y;8 6vZSт#Oe ,a`0&`oPcagX`  hYY@4ԋrcgrYol\1K^*Qid s! \ZUQQV.ʊ"sT66ʎBqzVuf[u;Rl3!}"A B݋9ϖ&:?Sо1[15}-XGqoX3hZ/pEFl1HCl*g>I2zސc kqiHya!~V`y)AeQJ?<i9Ŵ񼀜ԓE.%Ea~#+9TJSY㜰WDXm@!<@]H4,rٔ8hCe.RRP_^1O8V,O9~'`9 %chRt& /)LHx s4N-{ p0WuL pv86Ti؀J7ևivbvۚd ~ rҗ^tm$䯼+$a{솩T#lfڇ1"$Ib?m~* cTqf!f׎[~*>=_ #]\M s>î = z:|xg7p2ֲ5BrLO-jUӇ@IC 9\pp,FO]hL4BU|'O5 m{H%iMh Ka$z}ODwE-#I1Ѹ.ݩ)Q}_2p0A]`A)KϏ_p> ?=~'&U/>ɂoƘ!$ao2JuJuJ뫻glrgT*R/] ǝqǝqá ZR4*Jj\HG Sn*.vY4LYp8=9tg36 ||NG߿Qup_ZD !oA|hBxw^R&_~!Gw)w4ze!NK~)?Ӫ:sTsد~:)C&Z*IV̈́\Sy+^*aՍ0qD:¿ۉXDRI{wS,̑J7Yv\y±zQѶ~ ~.^JQH!ه֋>g$|s"Om>}~dN6onwق&n%8TX~;x[>j0 =(#IM.ŀ⟪B{=c'z,'(hu | j jӄwnQ9Z*&T~^co.(g ثb;ޑ c@cS~=F"ƔA녝vLiA~0tBQ}*Bp!ǣZP?MAO]>_ݢxWa^߼0jEޥ \|p&^R:FPrp913]g_W~YD,x*OŽXjkMayKI?^/&Q~KЉTу_:oNz.D[S%M"4\:/\VZ}bɸ\.[}{͟W l6Y~^=v_hUūz]/\.";Uz3&ӓ,>FMW '[+pL|~xzjSr{p{U5=Y37(򮲌>ػQbb:nWs:%^cjggnQ6bRH5Té* mҟn$"ϫY@4I+IQ5fZЗXuc])WicwV$jzuCT|քEе]SPrq*q-n :#n43eڙF2%נzaHb?|a9: u,䙛6ŐؒlΟ/QY.D+ADJ4Sy@'Edg%pFB\|% r&y,Θ9< ΗhF.Qm-7^-7?|j>f`=<(3Vy? N P;+`^A.)9~mZ1IzzN/֪gPv)[T2+zbe?*FxSNfLs_N[gEt4gxS!}i+ ng޽|Qco%ۓe=fױbq_+@nƈ!kZmMv^*EoJװ:NuqB޹:NiG9[ lc.oZH.朷*fpzIڎڱCkNSG]Y3ab/L[=04Tȥڇt7CƵ۾J~OY`͗O k#Ͱ0cX vcᴿFV#ޥ43j1"Rdme>xa;tcFT}cN`V3Pr2)rMFU0*^lK|F) zHvOwl@X&LWOW#Nk7H)nLN/ Ӄit2$5xkW ߯+cH7 bY۲?/ΰJ1TUs4cfSOȳ3@UsPMPsqYB,e̜t*3Ig, Gss-IQ@bJH~^h{}^B0bc!@yn4 =y-PyJ-369͘$+b璖cieq_A vTa6Xef6r7F 9{O%җ.H_ч,>dVt֧`P"D@9ZJ;R8X!S Н6p [HJ+챵|o{!o=V{ңB˹ճ[ZubpE?} mi@e]v)éWc2A`,F_4_HVm-֫S)`a!>& Up=U#RkmH`3r߫ۀ,'Nvcd}2`L*$e; ߷z(Q#^L-sn]]MiK;< BO[<Pz}]&Xk]=5v~Z #JW+#+oXeV4_ߡ6 ozs—tanwjۀn&D9m\ ͂4.ͭOiTkާS!]rIBeTP%$9J*:O_ۅQFݼ/U-^,m{t1w9/Ton~<1`OsMz)8cjUjcVx8Dڪ;.F{32ߞ6D5V㈕bs{ 3oQ/7"!W4NvyfnSe% 7@ݴSzԝ w ܠSLxs}2NAXGg*]Rۼ) Ư<K^͏e/~}.pY_~xEAL_eF)p2XH?)&DFmUQaY&)xvww!Cx"siٲ@o07akRY_GdpbGM]1;+J'R y"^-͟,jMOV?M Rx}a_TT&<rg|`8,)bED(5F BqW"h-C"k=Q4}Aⵊ"j+Ť@+&cHcWE =P[g-_숱k,ͰS!I'&gF J&ɅP%Jǣ⠈7Ůx ˶C=[ՒQu. lUPQnҟ_7;Z lgQϯӿG'0vrXumOvs\P뎂}ӓ+N@6pΕ_[ VBv~W7`pXOl:@N;0KWoDG똟^=NQNZ[VZivݳ=ʙv„xQWRGݱ6I#$]P j=o%a}֣ S`ƱO?d!Z?2ŠD.zjۓj[np;1 q\'%q_x|hE0s4, qw?~PT1o" ~9ֹEAw<#tG`2=SY4RyfĹ{9xLWYޕ}6AILr4VNf׎6)u Muڭ)uD;6n@ L5ڐF2U)Z>|A곝?6f;4*9M8ۄfY|z'e n.n:D烠BۓAe%MjD2#Dg)x񷭂~%D]^W=1%U칬$Q2Kq]F;tAQuX-5fP8tzQ*,DI4.9YQ:O_* .瓔ZY񎯁Uko6mVE& / &Mny\QEE[e| ˾lW]kfyH86ϵlsu3< IQLLE.I_Z̈]۰#j)gqLjI5960G5U`A1k͗%Gepn֒G yRS1. [o5]Lq̳F6.zrTF%2IZYcgJqz]N+#$ AV#fVt݂b`YvկVc]#Da6{f֎E'ocQ@va$;kyΥF89O(qQWgO6Ka ):K3>^;3@C|md!p ©3L9g_@D3FhB+̈́n{@]Q}Гs򦔮ϛ-'ɅІr6j57NtW UO(im!PZlnhb/CLp=+]1%ˉ! R{? ]#Ҳ p{#s"Ҋg{_KM m*}H 7ٺ}:H\k9hXH!Arg}HS}q0֧3L=+pٻO*~umn}{o>?g./){R!4sBr CLF;pɎ_\܋W~Q /݀AJR [^V8UxtB|'2x#|"`1s (TryB+"UYOͿ߼Oq83 (\gw^90V>Px*i Cco)7eKeA;4$ .$G E5*0)x^HD+ vfU&.DpVQySp :ƌ;0rs$ʆq}I"Wh{/ZY9*. $N$zD&]ء;s]_dҿ;Ჿ (8SŔҢ|KOHA,>q^ Q{:ҩl:tIrH9h2q6Ji S)⩁ļCYm"\e3!6uUgp40 X:pr= oI>Fj71tuׂN9Weׅnc}9mBǒA ?"=# Ębvkހ`AhE#|]nw>m>Ht@1}w^սp;K ldҒ{/Yh4lk~IW]SARV< NO_^{i0L`Ң#ezϵL !pgцl` 6'˿SQh}S``6tX&94S`-9JZ=Ks5%g㒇rF(z{J}9SNm _u0Ut=kQwa̽ƈR$'7W$x&*-x,)o29muhԵBXBBR1 єyp'´>Ti tJփ1#pbH `^fʉ4%s<(t0vׁZ#T/"_ʜk! eٸPuAeDWt{)] ٸ\x2YLU3C2FK Tn*=8.yb 8OԔ)M1c7vXWL0yE@ .G[I. fO^&ao9ht`E1ܲB=u^be@Ѓ4)Q7M+z1Z0p:3g(tVOpZnPG/|H PNI"`%(B@9óeLe٥Z%w˝r(]dIj6RN ))Xp0#(.YZc^PBj4Րw3%Y4_"Ψt3f]P46zE-E. '/@;it=L؞aCPe?^D ѲN|yߌL$?NDkPkoU`g{qfD#>K$H^T8ILҎX7z~ѕxP^xyD)jES?c^ PHQePD ap@ |-%%O5m|$ YYr)-"? @5=ɨ4K[0: $̔θ݆7i޼#7ӇW׹R|_ozM@0:9KoSVE;&+qȐ!h˂|Ѓu]tK<\Ӿ[r{58n_»wγ>Ǥ@[fV"?Y}-LTS*ʋ"g P(diͭd`MЂ-XQhrTȘq>IkY@5tq!&wgB"x@hBD;g@̚gm$RZز-8v SKe7Vc>ϧ,JKjdHaq#~L_!2dGʥdj)5?.WjdC` ^Nx!r/$&) GH,U%&Q(t?ǏOqNuӦ T'A ^ z\^q.2F\NP0J&dQtX >oUn@5\DB!8gs ktv)l$(7mדK(بלst1%2~7)X%7bD ]*J(hPq. Ugѥ S,kӥ_UV1B^-1ZZ^Ʊן^?HH6wkoG&dq=Ke33ZL؍q!Cq,_ao9JXG`Վ==a'>3 @cX{14i1{4%& Cv:>\x7moV} cG%5Wbv~UbvCN%f"KHI,ϝԜt IjKLF̊ |S{M.7-~bc;'5k!h'R`f.NPg޾X#{/?qw3cV.Ź !K0B,2,˅?}w΅_^];bT _QlKηi^I{Nd uOJc!\ywXA8Z|o5yDǤ*~*#H $n:!"RrP&@rJK1+K=]rmRԹ)o6{~&F[41;/XgE<.*UmSq|h~TsĂQ \T"ʧMrkDU @cu2gq&h_^:f-lǛ{5VKs\_s7[aJVvR`TdKQ|xͲ8 oǶGrg/?]V ,lFN HTܐYˮV.%hD4/ ^6,޺%Co81BHQw{xK3clhO$,=} k(> l$xxd72B^^..Dɉ]41!TieJdk?h8j@C3' /iYq]KëxD*U?ůfn $80½.x 0n#|oexmAh1n8CJqH€ RhȢHEYw_o[1R[.Zlg\>?]%O fYT+a4Io<3P2]̾$kNVƗIA4կo+&$縳 "8UMፊ~YΤcڲ/Ǚ H[}}.q^0a]>SeTD@BUy@˔PH$5M`WN0I;(aV[OY)~) tINkQ;&>77Uwo8od>F{oz)um (לޞ<>ϡ>,uCUIPİh(.Q0(/ \c>ۯ6lX\ w:ȉ](U 1$뭉|jCYcH.;>wfO6]xyFשPvz^yAx1 |!9/KK~JxvGl6KRpؚ|◻Ih9/O9b}Lcg·vR7V٦Q]}󚄮 tQh:6<;Qvv%t3J0Zm_χ_MA2gwbcq8I)Id~@$ 8 քƊTb!+F}R)P00CBu@%THm83kPʳ$B֘[xsh8t݂9G?^\ACog2dBV*nU+\ XH5j WO_txk>23\Fm~{~R2RS D uÊH|9oB|;^TDFƈSy^Bt!zVv,c8F6㵄 ;fUu!jk)㝢$8_q.@[Qrx_K`KO=R3Oؿwǹq .%lODpomzc8"Ko1o[NY-_lbdH\S. 7` "Jo#'<E眱Fg_UlkMHSQS_*Q:!DQgsr8 E(BB֞<%D'aj=щX@W' !g6(28/| $9X{fN\<[FʙFh+(WѝYwa| ޻l<»7o12q6ql#kiȼ}fֳh:|pv}e1hc[^8۴r5p4:Z֯0v<D\1 p02PGse ;y%\,!jHsY0s̾v28D3*3J@m+EK)[?9G=宓}7Kk-C)4EkT5k##Aeqd<Oӧ[4c̙_IL*I3>WTfGů7)Fus,v贈,)FUQ0 01(dP2Sb-KL8,4 " jrH1̟MuI|\&'߳r+l0/4)v%^c*ZL6͜/ lX<J)8ڌT`S9h=_<[ ~gc,WVȬ[?ub1  ! daPB+)A`RAuH0:s jn:, FmT]  ^,҈z43~/9Cͺ?K?QL)Jnf۶siU䮗[+=W==WFP5Z&Da(M$U̠ 58J8 XDP!O*m\ӱ"!x$Xs5\ 'q`"R=M=> =* 2VhkL!`@:Z`r) ifR01 )  1&fkrha=U^oLӥIjFZlPv"=0E|qQ1͓5"u1oXFW:+T#$V#H;NfH!HI|zA J{n\FF*`gn>s/=*dx\AvvDAF6/={QeBbbI4c  * erť8hdmrx" F""0"mF3Ҝ*h jFR$ƌVf[f*Zζ" 'Gob}^zV}l/Ճq[t酖%`PJeл._K@t~7c<dW)o.4Wzbd}^;)%xBcH;wЈ0'! *Ȱ:YAxlne+vBbc(3?z1 4~75 Ѯt=$m BE3}-Wݞdc4,6l\yFi"n i׉`j59 ȶU}# <פq$vy`TXu}B))}^M]>zH7*3t̓AXy[#\BhM` 0D kY(Җ/ 4M@*IЄB<|} Lf)|o=r]cTY`$$ HA2TLH@Dwm.w:=03؁\N'̃_[vDmLx3))(6sH[s܂|^,ږT~= 6-;R}#8xNQ"/4bJ`iZf ؒg . R ܖ4s]8ĹZTL&@WǪʸ}[`YMV(;QM.Hk*#m> [t}ƞln99tPm[v59nD5G4_)o-;@<">MH~nwg+CU+x)@EXmPmmH9վHl\<",Q^ν!H]s pXs&^֎Ex x5vp*ˉЉ\ųzokqAph] v ֎ `I#)-'ucGByF]׷_1#[;GXNJqrqAxHB='-O!>xVSE#5κJK ҇Ŝ: 8x9"+Oq{,"QM3+xO*%H\\cӿw"(6T -d +qzZU\+iq6t©XL@MdK,`XJ-ۓv`; si4<M *`#\IgcZN={}.],hh^m=5uV5Y\ku{ynm{nçxZ7Am_ƾ/|'y6)ɤ$zϵ ӣ]lAJx$oSyM#"UH#X$9n:f g8p8o+;_%-/-Ogͼ_TJ؈/bU̾WiEzTQk#GJz2Fc&NڊY#t弧 LUK/~S0Xn>YW`ݝr6p=8~5Y1׍?~+0ƨ!״_ܦA>N V8B?dP/%W ̫%yr]:q9vL*,dn2F뛣EAV<68wyk:fBy+-vkOήO'"[d'^d'Yd1y4o4#[}<8V$Id/%x2ѝ_V9]=}<ά,W?M+< {bpz? )``npЎ^yE;9+ DDaeAzϴ;ә;$\ B9Bp^Hm쬬#o\8E\0fءvhqtVhJT:Rpr@&RVN 0ޤ7[\a-4s^dY}Y%QD;qXc. UT!@(k*iF++ *E;yEhri ,)bNW봁yR&a$-gnsV!X ^  a=R"1LJL( ms Xdwvqs{ic"9#v@y`wˀ{ŹS ;8(N_ө!F[;%N4KVN<`\}GI\U)sW̭?lo׋ŝ/fn1C^jeU'{_|ӥ6|8'ѣ]|3ԋ'5vw1Y1sS;\qH 4X r哅1 fk,<+zнf~3θm-AAnm[O-NX:b=Ȩg1i9ю9S=~ Ɏ]>=8*^G%Bj'_{$ "PD[\)bMY쬮D[Nbkd%ZJ|-TT'CRd,HXS*%CA*^acQ,w0]F x[v(}FEJA+09b!2+P"Ã!Fa/B޻.eP :{<2L*͔ }J) oNs@@ G")ǰR@P1.Wba%ޙ 3E0H9#sՈ FBR-H&=N1-ZPgg嶩’`\'y #ܯ.akZֈ M/eJR! t;Ƚgv ^ 7"".WTy~^5db۴2ϨS-aUʻ |fp^:<4t\ڽ i{w{4 Bi^kw^oPZv=b9BGn =wv*GD8-Yp{*廋;,{ &=_1@l,l7fUss+s1?_-{,Yt%qBo?ԿqҘ|kʮ\t4-K>̝sO L+w\quAO~rHBr-)FnI:VA蔎FΣy`ڭZvBBr-)5bO [)9S:Fv:7TiR=oVݺ\DKdc:vcL`4 VAJG8+Vnw?z<<>~|. LJH0efB$mˎctӪZ!w߾ xsѧXǑ BǤ/5 ?*P6Yb*<1L1]NO@KY6L';Y|YtIO3>-Mh:yIRtga(E2=jLQ1@' H 7ޮ2Tob ~3(H+(Z:UQnx%I` gJO%'L۬vs RrgKto+X,C:e-X/NTĦk!NT9+E()xM? WXnܷ`Ycl8FP¾xbHď pٳـ|Cp\3?NJ>Ax@Ӧ][滕+ljPTbՈ_ږT|5`֨3:?Wk%+Yo&;`c{C"ѱ D}A[`}R/(u Q~]s+*OKct^Q}ö!*J į 9&kjMNs6%%գsm$O<ÇLR%zR&|:**k1ŖQ+"d"MfNUV3 k" ' ,$ CAU\#$rLQ=gJUvRKtC(=4sT݋0ʥn.Ռq= *)V~cr$DQr1+*>Pyu8Iz2\K8qEPt}ƂR]gAhJ(Pb @2\c*zE=p2z<'2 lXk jalzV>* p@I)*0RVR@;Fp@I1ZF-;k[i)8 0HJԊiN]ARDk{饄R)<"A[g.Y*UIBq]6Y2!/(oѤ ˇn];8~~r\E7Ut\E7UY48grjKGOd8m ϝqQ%T: |~~v!o3l[ѥ*0v`Do*B,>sŃ-Of>B%wv B>,{뷏iM:H=)2 yl&h&Cnvn7yG(,?Qϱg=nCotq9ȯg`3qtJeb9P#7nO:֊1!=}>8CPƵ:h7" _> L[kى8bpv 7m>u?]Ipe=b"F#54Q>=֯)12+3QjI:oX&Qr+@*@%ACPϿ3ZBrabYY%J̩Z:/"$[yPJT] k)8 9S,T"+fX Qʰe:1tTL2/9ULXrX#*0HDU"&mNsIi{CVN*rZNꀑ qUK0SIQA``U0ȘJth%]$|dg}x8Wg@nux>D"}l0d/@\R\VTqL4i~%((lKՐ"#ХBјj.Anmz)0oJ7 K-GE}w9N%v4.'lZAN'14{= <tVf]HW.E2c5F|K11hI׾[G(Su!!_ɔbP>J Iˠ3q<9{ {BJRHU#} +E,`!1E$񸸉I'h&RD*DKtUROtEe] LkmVŘm^ٗ$gqM<uƲ,߷(i֥%[j3EVU]bz] drdq+IBAP}xv7׾ l{7=o%ovXJW¦Yd@¿6gWӟ_xtbt/]̧ȏfQ|.7/o/ׄCYMTt"Jb*Q I"ڂ$UP:jvdd`Yy"-F 3g?UޅhyO7*/>a_{a;WhGF?F' ֿnX{Z^jHK?5k~/?8]}W٣zHHM̉'$tt-NH{Lp3 /== @awv|qf`Ь-) ^2)UiF2)/i=O5N6ҦO;K6SҲgsi8OIFMЗ=mhAVH߾5- օs"jOu(%57u-2[eݢ-@9ZgdH -ek4de2p8mQ;I@Xq Юuju͖j:**(&M,'ehZV'0T'OI`W+TP-2j[VWZAeQˣBȤY08"Fb4V :(l:i cpH |(0zITDhk4"`=x@ RGm;%'0 >: 逖Ҍ2yL~z=kÔ{68ԾAKv\[)7|-H)7^h!D>Yxȉ`ZzHsT_Sr%ٞ}z>4kͲ'"$MV /qt@.00|r+ʝ%PBpBtp@Qfw ޠρw$; ,$ˍt,b)9D q􅵡ݾ/o`ȶNp5)S:o~/;2W~ @JRġ"⻘;t1i|LCOPhm-xAeƈq a%J˸/G,x&F+aKƞ}Ik3 Sjl[Âq3\b/:jm&.`s"ѭzN62 =k9oYb?6S)lR$JgI0=g{y+IRʸvVni>Y%|4`bJȆwp^>\_fWOZLpVf_f$ҪR1"t`Wr.6Cc.YWO!5:) 荢O@y*`oMT)͍Vˎh),crԠE\.qinlDS"8;4)Ziaz=J`DK~Qkf&z1RQ8W% L @!,l _|뺺@1;Y XeQAP C!㕡49ZLk g_jP# ߲W3aU&b5EiE=>~$eEZtp̦ܼNO,{0PF68>o.~1?q-͠'k`F\/aOOVmԺ{}yWiCV`c#zg9BV`)!\<S%#r17t0_&aVXݹ^| 3.l%;`.>s`w0nc̐s,&C:ȝtQ +z XM0\@}׻8D8oT%/Lfx]og[Io?]߽]~6??3^FdpvڽkstGߺ4DLe.ۻ?Tzʛ)or`MqEjPo!ivsz~۳Js-ڿGLDHdB<'oۢdE?|[^tc&yyp'ӏ0O<'ĵޓ W.C~xfe9otg[?]1П!ͬ'j=Y;7N6zъwӴVx@:Gvpu*޼[-n}X;7%6%9T~ie+}v{-m1vQ^:,y3<'=OLrA=Qpc}E0qՍ17]IHĖbs9&EoQF$߿ +, 5$&sS'da+ǣ]e/o'07|:dr.CzWU(;*p~U#ۣ:m[?VǟK7w:@hfv\1Ұ4đӏ\J~C1K!p<0S cefi4py:䷃\@Վo jSA-H&I )a[GѓP)ʹǩ}ܸYۇߪpXِGicWqF[iJ8z>sEixQe/gNo/ ?{_x+A>]}>^\9(*뿘uQr{>kLǦ6"7^h!D>y(4JX+F=w^a5z%\˅K *ڛ+|( F#bH#{цdՆd롞]{n̵x[wR"`˕dONbFiK`mN[۠5~{p|sgA8gOn}=RS>c0|Zi2"{߀Kw *41:E 3*`n,D`txp.^6|&=b`6d*f^٭5hV]M^Z,aCÀk2nhF"p<?^G3T Fm2B@Y##s:s:ࢋ耧n~.lOk8BXpbv^N(bpn K52l ~Uqp9Zڠ•KB- F<~}-#839B9!J1л-ۡܽ=+^A݃?}>6{ˑDw0`6;حsM6E% 9"FIh#ePyTQgB"SϘ ZqiGvGF'"B†hEfBIh>Rb3uot)AQeSG:bd4ٙd{ph EG_tFO̡XvJR>vsq\S;e;k杺v<3I hHNB&e#*Nx>ET~mW ! s2[N-6)lWg肝戱C@#$IsWgP똴 )h[kzSNhg."bWwԊEb`E`x "`ԁ2*ʝ%S*TLH2Co -QSIꡤ\O^Ż1{6х6vsK%lG5 f&1IiX+%]hC{8׋:յvhaE!moMM.+ލJ2 ">sjp-:hq˅,䝛hBAhy6D eJNE0H1MbB){Cܤs K}g C8\p|խW:l{#uu?y2E |J 9\tE$\t "ɘt{1Gg`ʪDs'bzh?f>Te\5Piw\kr#JK8 Rg(XQg@ E%`H,'CG ``V@ @JNdHxA-2Cߵ__H/)-PX*լiPM0;w %@Tī.cT ^qq.18 =OE4m׎su!!\DSdJVfU'i7)*InĈNmې܂ӭHCB^.SXÿN甊|yKАi?-e_S>ߨ\Et-\?6TA 7Y 0STIG|co6U-WL}W0~z9 _ÌJ10cY4!N1j"2jBˀ4NT]Ɖ^߇UPbW&q#>'΁Vrxoߪ|eÖǤh%MtZ$<* @kBb4%bV8 hΙkEaE4Jn۞.Pj娀,lJPJqqxWy !Aeh jÙM_L^` T p<_"$"sfDY%X%.VJ?Y">ۉc*?<=Oht E4ERtgnr@햊A褎F hxқvK?DjEpkFx8.)4d`ʑeHZgђ_>%yX^'PÎJʡQ3NgF"Ќ% LRW8@MԮz=ZVl9wH,v6veůIS u:M %Ww&I(` P&B6~pY/v0 8VlJOz`Vqd644\`DžuC ?0/1bFƋac%M~[Qh!D-Aڿ!wix| {Wzx H')CZBhH.J>tGhH9ȮK sJ$#^Sx6˷ހ][|&3mR̾=b=E`rX2y+# '=72L^@>ښmoژo5޴VơVR`!װF(.%0bvk?d ~Ke9ԶÌ# Hn RwENs1W# (&8BG(HᄑeZ@ wg[|v=Y[NOBCsH}^U0V~2;uyX15Jm&K%o?~-^҂Yw>qzc7woϐ?4 }N~Mp}=]=b7߹Vccsꇇ̷ѶˆPۓsňQr6NaTb Er<&I&c6Cj/\e?rPZ+Y !/od"7waV(ZsoI~lf4>W5ۺ=)J47R+%L38qVK# '&%& CFFLiɅ77!eȲ>IdF  oŕ xwF 4 ,d޶+cv(C<=E∼F8SƆqtE'KJoR2i( JSsPR CH# -} :3ԟ9VKks[۶@~(Ϝmxbv{d?xm z0#Lm(\'BnhV[۩Yt =w5߾~h0뗽=FMyIle 5ѫ|k_t.[)@*1}ꗆ*A|/d-0 y,{]Yuֶ`V;)ztwW ٢U~Ph)%u m@0s`K- XI9@ȗco 313T'('CB L2JG{dȁ_\hyC3luP=ZTuΑSGG1ʈ 鼾d{ DAy&kp$9!+io: }{+ygഋgun|l)^,ɋ_+ #q!B_Wza}3zQFd :P>0C"23tkӕċsx'E d&$tЊB%+DGO:@4EmR;M-ɂTlm)*N=]:3\i;Lr>ڟ# DȲ Y)-SpgRWv{F].V, 0\:$@ㄤeȠs %)R""RtK7>[Mr^o1H!c+m@X(cBI(ʹG]:)FBYì@HPжRfEUAuRV %T`D/ѥ1AhAdaS[YNR-nS7˽Y݀H Աp,J(;` P8ąvH A|X`VqT06q? [CTsTەpQŃH 0>T~~*T4LDO=?ŪCknw oy~Bp ?琉7a#X/$On1zªr!ƫ{?Fk&;?ܤxI&9^a~bo {]\u6MQ-e5|%}nT `xO$ t B)4c-mk?Y{UIJ*Y%A7C[P%k4ܥ+!pju'+4#xM~Z AHZ JіeB! #  I"=p>=$Zx9Kb< C*PI闸"`o%@@ ia= 2F!q=:2Ɂ ,J!ňWk'=g2,ah#!bV[:&2P". JKWRrh셜Ryv'=qYP^p׆TzC45Fy|JI/%@{]5wcV맴;PM\GG[qfॸ{qE!\5Fl! ɰO}$Qcqg:I1E(5Bv"VcD&ӷ=ѪۯzLG_{6&kq;l;vuld_gu[&q˟^3LcW 0og٭Y.G4$#'Qd7-zGb5r쇍o;xOd(ަk`C62Y8sLZM43(: eOFK6M~j"m2q0z7Q1?0 5BmH ;EDuw6^W }}. cy5M{` }0ԪV~N:=IMCI28C\ u KǤǠ@`V͏b#ܤo7m61!%q?ɬ`Slh oUD߿70g*r(Mvi"@n黇ϕk':N,lL e/p'lB|Ov1Q2nY!Ilp6'R ΂`Jyya.LF$N402OTοU9)؆aKƸ(DVf}0seP Ij7cDfR -ezBPGE64^+J;W`*CbF\V"dec?B J{1edE*M1fʳJ1Ʋ,QyVpJ?7VQ}~9Ѹ>F_#3ZkmXd䮾?NٍՆ3=Y]~pS7۲;hCkj+~8VYc!]\|{ۉB5cCޯ[Ȟ7mHf3تxI Jr+HD ɧ 0ƭ{ߤE}oCAB>S.UlƾwWxeyEjK|KU,]rBN<$.Co''B2T"T|R՞Ѭ4gH;GGۀyZJs񪒔V2VnS^mJ61spe.nNZ4nğ c2}or=I3@;^=zm\(4M+P#Kg5|UN8rcWkiNcR=YlAͳGWconW5Xq9D#-j}> Ĺ~T8bj:af7wo dnCwtNVQUt>HUmcŽ,lxdOK;ą9ĀXYqS ؛cQO;2Gw1| GQhu%I+-TLHhM-rmk- ݶֲ+` k8.Z51>s6Ӿq4aߛXy%GGMh crAkqu[n?9SyLD.w} Z>@PNY •)^8xxBIJ1Z)ńޟGphUk߸hz^[Hé ܁+ WcDv),]}Ű۲Eļ3wRA_-Insfc)<$#EG3=<4+Ҭކ##V8da>|{J ػeHp&aTGNA1!|=sZp8w\+V Ųn쯻 7 )*}&.HIWӲz!a Ow( ՖË>?Hߎ]Hfxi+ŨBPޛ YT(,w˸R~?okM ;ً;%Gu:1%9){4ڠHEp;/@ d67ө #=k딍M@v[F  2=1x;*s/2$iE0zqa8 09'agh+Є2Cw1nxq3Ndr3E~Ww˟ /Oo iuТaHwV>MS&qyyyy;'UW ?Oߧt܌HqeRQ|02d%y}jLӘgNU~07zg5= jwsi.t7 ˘^@6dXisbmV7W'gP?]}''Jڳ8f;U9O5 nA6ZS՚&s+. RɒaJkЂ7fd9ܺu V ,$ORd<@XԊh0C21[Th(5Ԑm"7r' Pm#dQPO1S]^YD|RY7c YdJbY9Pz]K, EPx _/SF߫(SSr;b 3(m0It޾i0 qqQ1 ^&$"F3?z~AKⷯfX1L<<^G'OoWyZkK6>>3̖S^U^0ɼ]ߌK)#׭ƸγDIL|ke=I3.2ekVՕ/[7֭+MDmӅgSQ;$kyM[EL*wkjݬ܋ح+MDm>[6n}H3.2U_|zź麟B&u6 HFn'Ժ!!\DȔjV6\5M`c>߃&z* sc׫&pnn/U'^5A<<<' 5'^58<2y1OS,Ea'X.'^5 L@N'~5AIsxyRcЫ&ZC xO8 }jRx' Ȍ. ז}ۢ]1&+2R%)РPFHI)WƸ-^2.brikG6T xcR@$TN?\0czo-.thL&2No.nG׬Ai O↮\(4(2䤴[jas]iVprD$2h4Tqd~>8P]GEМ,ߋVmy"ШV% .YqB?+Tj~K޻{;k}{( ?]w#. sry)w|^+ݘu,pPCs핊WJ [2+l.X]v:ST\g%xt6A-ӥl?a c;H OԂRc,}WQ@H2ӻZR8Us}z{#+\rϽ>22Jq*K(e4dR wIE*f\4Aׁ:ك?XBcS~@$ڊ_߅OZlGgc[Co UUb,e ;h3sZ+PpiNJýJfRpjF. A[ʃGSE\ \h-HQ\7Dc& '01yMdDe ET{ GG^]"~#'ۊQ LeW]MʆD!eH>WLK7#E,>:cs1 OF1@ B'v%H!\ir AU257.j YP\sFPkRe6Yˉ1\A1vg֥r&UI4hT%ѨF了材[{dg˩^ "hoa[=\jh1$AҠ)P^6A$cNFyA59!mYFaL7crqNγ{;1[WlnmM,@I[ \HDYE4&IQlpΠqdAH )fN`R%no,Ÿהj6_Wcw![LAmvK):?HOjL<>rf"NhVW)dk:xr=˛iFM/cPM ;s&_ϋ7r7]_\C*k0u% B^۝4k;jDPMIz]kn1;< i9ᒄ ^v,A'4niWsb{irW}srsçv ;k X4BmWg6V?UǰuJ8~aQ">%E|D,Z9%^DR$'0SlgmsaO(lT-Ng;ȎAFKi^qnc.xI Y#*0* A.Q܏j %~NEЋhxaQzߺiaL5:R̚$'*#@z"kf !%e,&NA#Fz&#:ݢ(uF(RCK^vg^wm=nz rxgXg,M6i+M$x 5RԒWive\+"Yel [%LQliٱh@{4&fܛ *QxEC.p*Oa<[p4Y O$O8s553\'c]'cj|(Sc6*F^C _= G}9ؒ'xz4=I+2P`qϹփX-FW0D}3zj?V@cڗ9eGq^J5wc~P {MgBMMpB,pp$[鑎ms}lelkuXwͤP`5 F{՞?#! =dU]uzi`|jTXP "> G; `~G|D3 طqo<_Qap >D*n*Ƀ_Z.CۭtQeRΚ+Jew0;;3Lʈ&{wo iVJ4}|`p⶛Zu'k0]0'+"3ˮ x.&_= '1O\y }Y88&|5dT`/qA3i3O5F1m,rLS`zhו|yUV{-Svxa*40c\IDK( -BK+@EBlJ:;G$;+O-:B!-+w<|fU:!32(4'4GH`9`3:~D)lxs?yF5aZ r&|DzqŊ.}C{ ¬\R-;Sj:T\uubƇKIp,Su(ԘOr>Q3+A H$Ƃ=,Bـ9N)9Nփj1qZQKa-4\cmRTumvkR5 .T %Қ]S׼` py҈r҈Ҩ- zX(>Km`黧YmxE[/c]' y3J5o 9j0ȟ~Q#&ޱg{GtOp3fb`98n n egVZ0D S6j)c޲YcE50%R`n`^0'ZB}AƽXg范x%loH ecU#tQS55Z gP&@OA`e#" Ϩ&( HQemjaTtT['p)ϳC0Ua}&˝"ekj;i#و=RHIS%cs%@+nGo@LL)Dd|̼ V#5BpM晐, 7 3NgThb!%2Qwt^?J [(NQ,2T: Ah)2)ZX$Zvʱ/I5T3ԋ_+s]dD: ZyqF#ʽ9B*3a2±WH2Ej\9f¹>+zc2{H3,h0:a*1mxm̞aZkB9F}ٽ&,la;,˔h8c, =aC+,2`1K3&[i]8 {i^tA;W֥ EXb G^&K>5֩aaL I+QVV jhi 4'?,?r7!bSڄ/UXGIŇKLWGӬ_faVcP5 JllĬ7&T=L\Jg]0? ޗ /PM{f7t_obۨƇoۮݨ][;\"2!%7Gzܜz<~<3xECNyoKQĔA(`+ d\$=nk`D![Y=mNyHkxK=5O(R]̴wЍݶܗݾ mZPz!>b_c%O%j1:(E&AwNGYABnZ}I >ɪh:8o2 M!䄬bn &+n|X놣0݆vƄJ$N0i2"R!¹eoƓt1MdI~'nĀpȇϿcSM3ޅ+BA=܀EDHџw㑟. 7-Gw7TI?] /d&GA7X{So(LLG5?>ã^Տ?о/~-I0>V= ind Beydyb*~n.gv7e9?0Zr~`B~R x[ͨrYpTAVo6~ARrw>:X/Ow;5DN ]&]5vr&U)R{,PWa^y*Z0jIY.+3@RV͈:*.zTL&A(*%hؗΣJERvJQTɞ=՚s*OFK Fn=GU J7 ~Vs:凑u+?H>_|-A9cKG rݝVg&HPRU1LSRHv) XJ[kuYbC.k= q]ʑTI(*Ά\N?sv}> ]8iFqp_Fs0'[qǓq8ezd}|Hbk,%W;r7Kt*WS5l-37_ǩ"JL'X吰p#YHY*(d1o׷ꥵ(!'tHIKQlLZ8yS"z9I GB{+Qa/?kT.h|M c _1C'D!ᅣHj5z q̇chR'Œx:NFపF3 @]8y_)q /xQg'#+HeXqO |ԧCPq2+ /8ᕍ;YJf"Oݯ&[x@b-21꧇Ô\'Ch 5Ww4|y7_1sH_ :kܧ6 Jii |na|0Mۼ"zQ e r 6(R<_喫pܥ3![?iu=!X}UӲ0 SY|:=.hzx\݌Vkd[%|vl㷣lF+0zmCǥ)'f}ir%[G!W[B3ec81jJGGaAq\'6DXŌ*Wԃ|4V"-O붔Ss]S d A>O#$u.#[ 5xҵN AB=Ų]TS1*F8pQ 3L:e6(A)P `0M #GҪ4>y=}0YZ7oI#o@(\y3 x霉rVՂPQ>VZK H@:E+G)Z[<.!4G~ȵG J[M@Tf\F%$ k*4ҧQ;0T>r&LaWip7qb9`-ږZ&J-'QNt"HXLi@dJo7s? )rގvLS &GE @5c& nb?͐W3.;DžcQ1KXeo`HL$Gj鯖D|LQ7:Y.:>*nF-Q;/rZ^{IRgVUBWѡre$)hkEP$)E;MC45hb:njRUpV| _,6J2UyM\C$tW. U 8t՟=A3TɹSh$3\"kRP`N"GIctP΍D|/ H 6.G}JJ2Xp0Ep/sGZU3HE  FN\`iYQpf* F@ 1AHJOMFmO-,U п]]+ KO\CV!u#*+V40޸ױ wPY3?,3WqU\f̺cieQ)UpuSOF0A  {䴴F/}~(5b1/o:+)F 1\.1b`]eh.lO̊0efAw 뭦AE` 6+SN4gQNfE$=l&)^[ 0Cde5%H0\{ /A&XPؾD&D ,U%(UqOZk&a=vusz!0!R uOX7>| { ѳkF`ɼ&\8ian3qm8 [>t2D5VݍFixPqjGٻé$k0 ~ KKuhF~ Ya[iwj~4Pl6_; U?R̓Y#Ź5MFf}WNofkV-ldB|`zzzwm=u}mIdsVQ)imkHhڵp*kU_C|DU+Ԝ>P};O?wDyFr8v)j5GU(tX*|JW3gPSkKz, K8z *`@-qmr*vEk&8g˹H H B?'B/g i6pDFIZD珒N"bTQ)H`,(9N#ca3id4K`2Yu!P1Ya = PD"Mt?H}MWh(ƴrha;Rʙ0^'c}YktY T;7,/-n݊5#F}|mn|sw7lfx#k6;w߬nܰ<Ћظ^0~>-:GVGɮ|'ve~ujTudך/=#5줟T‡hΘ>zc݈b-GvSuo-faukc|C4kLUn ȎuBgn >X3*ܛu/3h!_S>(k8GCn >X3@f'Z6F>D7Aic ʳ|fPSԟgX! E`«EG~f22"Y6ʞ}ޡԆs5R|7mmTLHNmկT+}cl@Y%dwv\g;P~c;]Mdޓ*elסCb%ƺ8uGUQWqD]UGF.'CEG"pF\&^[Rb δWO.ؕQ^OfKN[T[*mZʐ;'W+C~W*A{5 &jv4_ܾ+?~0O,߽|Y<뮟|_SnĆ=|w@wwGv땅b>s-Wj9B DF\5j8V[bD7Զ):n\en*]6Čƅts\؏PUh`>GXA9|hHJ-lF^N.,)oC וA;%aLյ+=0vM\NPqH+;C|j"9FD_7%WUIn|/!?X`"A"nc'#Y1o7vb>fxE^]UE^]UuYe=ғ߇X<2cόTTit3[Wܻۖh"O*E0p@+AJvDr#RZNMI#45V6|XT;P=>Eecf0+ĺ .j"#C]Px(Q_٫skgK`q>,>o$!c 3㧃t CkrT-Cs ?>^,ņ!p1$=Ghگ5E?$ bnp|(.`/D'&D$9"*xIZԛH(N{t‘/it %LxJ0Jݹ._:&KL1A*S$= iʺKl1eT.= UC @ƚKv_CFZ sq.=_& > 9wx݄>qC?Dlpl=`ĝq&_IlI^<,&ќąUk6|h)u4)bt /FKLM4hHs`aX=ӻꪝ=&<B[-NxPq|˜fy[ 1Ylݞt,;C pJ: Axf)0xG7*vSCAI$rSm%"ZS  Q"/"11眾س&nۥR?$D[~mw[_@1FҐ¸P J[JP#{/p& m@7ihewyLnq/Og֡:$SNIi/г*yzbywhzxEjXqVK[{KCqk5 ܊` W8݋l _7T?Y܈CV`[LDML=6V9W}n}s'#"j?" e)E.DnA*GeH@ǘYT4 Lf@%.d^HI!R%%R5 .(8 .j 6vIB\DCdVM8[- v.Tڛv @bHȁhLA| vc[(>:F6\p|ߵv /bHȁhL9S=nfHAb#:c4nKb֛v n1$E4D$wbKʩVK&W"oJX( W䝼qOr҃c*ݤ֔b,0k >Tc j 6 lڊ;?wF#o狴Ljqfl9gƮ K,Tux]U[| RoSE].ʹ/ R}1FLj>,xW s}@*ך. )3gDhSQݶؘ&D0^)EΊcB-7Ήt%o[t)ݣٽuȺG{ leW;V :C ZnU8'zιј׫q6DhDH8f5p ΣFF,X:  @B⣫ m{$}&YJpR d76F`~Ӳ/s FVWlz\4|$QWVD]Yr%jc0܍dNԅSiX\U)$/iB|e)ttaQ@ed# WP's]!SlRYOW=tIM˅ j?ZI*=bl0\, a(+3.A eP4bZ +^LmWmEK#aD VcBr,\%C#9e#kX.3eI4R$ȅ* k*E uiA" lt %W<ڸqg5yy#/^+ O܋ÊӈS$4`;AA yRhi~B@"`qYhQfKWIl= N$갟dkog1q ni@o=~{)CNHǯTל-26^n֦)_O/eF^6k[Y[2D IU ?Em؀CBRFSP.4q*zܒp"4w4gK:@#6WFBMPXO믦fRgx]b&7j*٫ @!l==9 P80\BL2Q #J՜)]rRbZpX%yqu\ @!^O-ʹ]}5[6k:.drN,BA8n_Hςuqc4-hvӰ,Cw"a|Xg7!&~b h*J/ŝE?X}} "Pߡ]])ۄ  =H@1$cR| f9- xRs3BmVUIђ"^Of\8niŒcs('im:ߑ/sZ'v[B"J!ɺ 1t2 ΩO7ǝ%py' M DJ,N(KIޮ]YyvʛQd99CZ]"{]j,d mT7#$S@ࢀ"/4D% 2+ J4B |C> Bf౏ < ̾T\)  &yIJl( b79p=-Tq#7JnF!'{df3,K6TM 'ZCTr"'Ȁʘ11T"7%dQLdN$1p@4CxKw=*Ɨ*1U@\~gPE4|8z;#bW1[Vvso߁]h|ɂөQNo3~f,}|?sD_yK` m (qƸ%fKNbO8ֽKN9ܼp5kOlļ`r!"D(F3xە P+gPN,XムiS+WiMl[9 v∷l%]#$U# B{UrH "|]!TNF]wT D9DQ6/^-:5Mr]--$@`@a;Tqv`\qD.-&aE{L [`8ŷ7]wsYҽ]' d,̑:w]tEwr1@l\^γ27`֔~N!sDZn͢9ѻ.foo&RILh]‰H]B8%Мj;&9Vb}./NnE|iG~ +O-Ԃ YHV \pańji9x&]^yQ:ӊk%Dtze<`Ѷ*U` SN>f;-g8E&]ns+Nެ s@)w}n(}ڞJj.A>Ço{dZ[p]ny#o 8rcm ;)ƹbpLdycJ̲(.BNrU˻YWp"s/xaU2>YVޝ# V" \_'|6/w>#$-ri,kZ{-W_wՖy~Bi/@5#Z몋-Y>xܠ˄JQu6J;hcD]oP[t_o#ƪɥ4:oU*DlE|I6Nj]>x a,b8va=j5DN F9zn Nbs)'teO* 3y}fb6hTg6~tp}6mq?=^oC7^Aoc<C] dAn !?yN2ɮo\}v>eNZlz\4b8i9]teӕ]NWrZûseVfD VLҶ@ Cd312?yFe~󹬰>.lVVs8 tl7Aneͺ.ߐ)̶6F^Vy{b^_O7/kųg,|On!ɹ 2)\"n 0/Q e$%XrPB{'B#^{V 8׀" s+J0T2%y0F"BKx% /⚊R%u.X.uiV%֐ ",9(lLi!(6O`]`WDKN)Z{f$4@9EQF*Z eܠWa(e -!<%h -cٻmdWTz9qsN*urLe&/r8][$~lS2)&)QGuht7 '>}~ٰXF&w!Csl]&8z!!m9-n0Ζ fN {{ T* DMt:q4M?m#:y}|5JNaWDPQlk7Mqgi@Y@$rPKD)妻ݯ-fŋ99f!MsdD<"W4Κ[ՕFLyk*I$;697{&X>e–m%s:Pcy8hAgr*ޮ-MkUzt;OhUhڡ{[+)[٭]T4$-V!&rwܴHw~rPi͎Z$lgG\-JS|Ud|6qeY.edRRkyPh#_{%^U(i/KX̕F+vc[g3xvD9Ei-{Ta;K20E Q{p_FtRt L6kMM"7bRaE,TbؓLzhS-Ř fu̚xH@e0h<9%ԢI*gȦW`NTHFqf63=+lEO?ۖy03C.'rIE ?c:7iyg dM{(Y-]Oi>>o3zqvYͲs(%S"&]X3s#4R3A3[OAױ _~tpAB7#>LIex_chCv.*mZ厖3HǘOQך_0BUAft 8FjsQ8VcJM':9Cu~>sf|a/٧5BAon#D<. MBeH2Q*}:W @KYѾL"{Oߝ.j[|Ө.ÿ.O_A|<<r@ys ^ZF71 8暙42ƥ&ބ#R@MI}: |Gr3Nqiq1\z#}s UqH&ΉT++usǚ븣P۵͆hy 儳.DtMݯ glK6N[naܭbS5<7*is!E =Hj418kz*c9G4JSB;):شLnHl tWooY,esEbg+2ܑE8&EU[^ˏ$|92cr6A"A&ȵ-8YOύlAT/!Lk^ߞ a2!,)'gۇ7A>LgK3t6wi2o@wx4%Xލ@`m=>C .yoDZdPj5ʗKP}Wt+t 5l`L{Ge2]b (h4^:0LMs-ھz\Qƾ\ J%iUQ6˩,L{Xܢ7PNҀ7 R#;_<|Z`A/s>?>S Y\`:5)Lf[-sýsqM(M͍g6UVV`Sjfӵ,h!JR}S#0ɜQ$1T=4';Ua@V0`̓s`7e9k=F b )cSr&pLe2GYrAuwnr爪M\)߆M/Fq+UG}qG{|ԇ+^qjʉd>Hͣ>VdPQ#j)B#)穷B~!jsT(1WfyFs,#D pԢ[-,衫:],[z ]?3v| 'kDE-ec\Lu'a f{Xw\wP0-S&250d`lPLQXٰѵ3I@E55xB nI:Nq(zF)M,D=Q8;p^>8FMBGsղ LQ N3&~F8iU{u[Qrs^y*yfu.Z8SGs4,2*[+1\d3 &ڏG C|&]Ϗ/dPsBP Bki:55X"fh2*쳠s16%A)%^CTe7șFBFH+9 HM9lPЊ@ Ӣ{9[ |C+ IQtp DY#3T$D?1#qPGдR)KeBsШ-h A (4I!sA[:[&)U"Dj s CE`BmZcrU.#a3RI=Ց0V& x*GlLAhj^.` md#jRuJn2NtKguzdV+LLI$Edv#ꙻT O^\c}l@N% un#h|u[NUgvѱuM+|˸JOsw3?YL;uFа $e!'y0qmsK +`2p~Qikx0qsnZu#ߥO(wNu م!&`LtXq]k+ՂAJH.)> =Օ1weXŶOf*OU2l) eNnh+\X%ɮ3T&2.Hb2*GL*R:cpcF.A*CQQcv>&AMAöǏz9Q^9Z4E=4)m4{Ӱ0|Y@>*g+ < +E:;%aO-w{ }+C5:5mxt=ao&;n*@:!b;ͼMkmFl/a1q8 f݇ e|[IN&X%KME3EuXUj^ܘWGfgDaZ|:E.orK[xzwSWxd]X'c!uaՍW7~^yuΫ.`HȵFx| UG3 !/qi\4J]6EIuaS_5>D5,rS!ShBˉ䥜 *F/gc"ږXfXs L)T%!B0J;/ :pݵvQTܹ64ݺ1ٱc(; }ݯٓJqC3u:+u?,,|TPiRk V"!%Ls`nUaŐ;XE EJjEK+2zfN2K]cpL^)%q qو+9g r t+=nHs+w /KN/yq*EU|9rw*s @Γ #4Zy+P32'fkE3[8:H=vcDZsw"v@G~1ug]_M㝠Q ׼ck'&ô_b+dŭո[鼊 5 QӶԴֽ+pYii;{ᴝ5m섚 eD_Si9iiA𸴥&kBٸLfgzALmp 0a咟p "0ZJK%+UX 4U#,XhQ(IcU+@ȧy6_|o_N]J@q(9iSqkzhoj%@Mx\z.Eкp3x^4,(Y\9KXfC1AŞHZJD|T8eQBLުx͢t\8KJEp#>Ĺ  x8w(Լ|rQ3 1bru'(7FL1 )G攳"="U땯9.gikj]#WF0A&6ڟʲFW5!)Ř4%ן޺5Ra7 ̱ڂƿTnOu90;5=qCIj4;]CHFE@xϡԦtlN1a,>]9E| REAEl)%!v4<$ځf: AM4PNHP%;5^C݀0V a}~- [*hTBXf,1(ęb%y55Є6 pw10o4+H6>K2&pLQړ[;ösQz5cfA. 'kBaZX>\y+BT:BCuz\x|ٺwOd}̛nMyn@o#08.meprYH T =I B ܱ5$=-߄7XE1ZZyِI 1Zm%ORS$Ȳ7t]s}aqQ81;:S1L3N{H^^XIĥ`DL4|z&7j`RrP}t_nU;RKaf,aK"h7?2ɭn,s%XS*aAm>LW40Fi.1sTJi -bGdS%M0Bar:'HGxi@+8nrC\g'0+("!ǭ[OD`%b2H%*bvgι6KSQ0J1q=`r ,U^`=̉Յ޸/pVe9gedX 0hfSJ ֺkgFqZXa#I~:3wŲjyrͩˢ ŻŃtu|wu#u0;Mܺ7}-?ś'7o _i>3zH&$0wSGאW5~Vv,K4sfS_GnAp ?yۄ]I$؏I- N'XId^#%Kzp(Jg،Na|ܯN1r6YZ a2 S&nmClm`A*Nh)V72@Eq`lP"\E0oj aHsi@kn^5tmKxݦ߻ϧbIVkaz TȠ%NnIĂeezMS`}l"vBW)B`JhfcABdXVo8֒;5u 1'q&!:,ko&M_[kVA4Ogo}x|?/W^)A{/d#9',PfQR1x.'t@wO_&kqam<_w w2> ,Ǫ__c?xp~I:x:N۴6I9/4S?/\`*2FW>DSZn1n w4YD:W [Ѥ1ȥNx_3)O.:EuOxOdDŽC8e%CYZb5Oaiqi\Lꃝ}ÈᦐntaE 폠v6י~A]<;ȢBpFa]1CNUAv6-}/e~h|~jĢ nˉ@[(뉢|Rg쏵ǼBȞS?Q9v pZhjzS҉$G~&[^ Y jK?i:bd֪rY:z!ȬMW>k:9@sdS8 G1N+fg vrr3'=r% p_O?J7_GVdI^K{2ra,zOBī_HuӌP afB&nDLB^޳a[}#D)'rmZCI;}ۮC.=փAG nGg'{nc1[U@mo\8xKEH-d`o#H؋jusD((tM)rW$\{UrM-a- d?4oJ#Jhgܖ l (A!Z usBI8>VRTAEB&O RJ7,R \c=d@AOg߮ _|\^{*sҿԋ}1D–n X2q~M_=̦+e۴-(@? ݜ:I! 7J 7c쟾?+߀ߜ]{Je 9["@I&/o=%@ׂ#%vf^`7VyfUjxVL'T>ǫ_v/WJI\<͉+Yr1OH<\(ɗFH9t~j!83q]_ј+ J,yw6ȭ|p Lj" u "0,8JsKP ҏm ( sLqm<4K_܌U {}aT'q6(ALw.;H2+Rs|S2WZ&e&$Ȝ-ޑE2Ь$p;״\O_D@L:n1pwx2 `!F En羋g-njb4fA[, b&I=!I h Q0ᛘ2 VTz16|(t79J5R!&Nj gm8t: utRNqWꙤMibR1zIZ\ix(wfj3>:->ИKcTtFc3fW:%i \1C1Ձ(cD20N87: SpGA;L ]gtc(Q+0A,%XIҢD=-1RR![w)  (zqbu!d)Q!Li]9ŪFgj{ȑ_1nE f3n;{/;ֱdߏd-ul5Պ5Ď~X"rLoP5+t]kHINx4B"J *%{Tz0R;, /Pd <Ԡ9,"c.ut xz'IrB!R<3%LVrz]>K;1١vCc92۱JX+7^6U:n{7vB1M{bM# -jlfM`? JlOQ}K]tTN f=[ǔ}R*Ja㾊-Uzq!(+7%BIc\i ] a[2Ts96J$q!jƵ53-$'u/{Zʹk{M|t^5Lۺ?xwt%7 gLZ Zh0^h>=K)H=g^#^4CVHZP|KF|kEo=ߢ?z75%Q [~"ҔA}z|A% 1bZԳrl$J5^ hf+/甑[%hcs :ina4AOYA͕u)Hgis8a!aǀi@e2q#bhE৘np~i("Z}b&1NA(wt hIXcoU Ǧj"8R#u9 *̶m#ml/`|2p+c@b48?V2 0F#ZyI+'+Owi5фU@̩f6)l$ՐhÆ9!zV]7r ؜*<[†.78nJ) ە2M͙(uT"I?=c( )p38?$FE1Fp5b<-dekX_/%{w1鼽_~?Yry)9o<ߥ HI1_,SfG_*`xˋf1eD|1WJt|[b4i8 >c ]5u08Q8L' R"n6E: aCZgd F׋KbChZ&rjr|UawWnM<`i}x_~Z2c E|UC(QJDɑUXxF{;-VqvZPbU@1Q$H\ec @e<@tQF=_fjY;=09E;">zJ 4ճ,^ CZa÷|j5OV/gkpH[$q~&<e}X)cXLTGa>h .1_n9n{v.9 y f3Ms bb0Q |]nTC(yzQ"YJ T"p W %`r`;T ڞd]SqS.)Iʱ@ 1lHHYKe=FR N-.x ^H[D˺=Ftf9׋M!m/!vn9kz!.\ea6& 273&xfFy5ʣ {ߛe|9o%.;9S;E/k~gb#+{3/r#ώh"Dbo*/V'* IzV s4a".O$h5N@Z3덝!$xeκ|V.Ĝ\EsIde24żLV@i B\Z57m巯TZې1x(ehjזq|:#SR*= -,~1UuV^+pn*Y+NGa2PE-a hڜWbd^|_lDUTkS!աho,Tf =82C>J.QH=X9Fw3:]wS! Cz!L> } -%d6V=,[1ikܦ!Ěk4(t]y3}qe|8OpwW+~p}18f re`uQtwo>eEҁ- Z,0Xl,_f褭11m=Nzbtm==(ڢ9FF ;rtuO dL~m,ue͔)xF.FGW`ȍhX0CѷspGG䊯ZOIb ltt 60 r?~2X y)@b"8 ݨk(~i?2po9qEr t<)"}XKh<>𶶁)nkk1hSf^JSS΂0#$ʻc+ W\jjlw )9v'5|->XjFEwR#Ȗs@7fvKZq'0{׻.B$$-f:G *dm9iV>Uạ# UK/5bR]$&^UGS^A]Z]]=}c(Xͤ<PS>kQN-$9r:3ӳ 6x;!)32 \(kB^Ŀo@n$xFj\PGQHYg?FZZIւnk}z{H-(/ %2~ ƠuG_ȹqr*ErJ V:`6}qf|xH/e,-`S2U9* oh,G<Х[YF\3Mbz W jju|=5 uCpNs i"4_{0=Sq948`Dci8CB:d1`:!F(lq2Ak'?l:mI_⅙$-ebZʐh9CsRj6Qk 8Ziiitb]kυaįb0(ǟW!?m|/oE57n]c QPQ9͌U@͢D#5w.=\lQ|zL'!SDnӿ~fdB.h5 f_ސ=i{)6u=.-6?7gmJ| ˚*GWn۔TBrTnmK6)Ǐ64' ~[^OkT`9ezfQt ܬ$˼zrĖ\,{{}Hz@4HP` @h3ttZDaf*X(=kþf v̜G4VRw.Q+{X?Jޒ^Obg}bS *Ie.o޶T;pZe)A˹ghEo.V_߽YMhl6?-"ɬZpu9X< :q >(x|1Wĩz?I*IᕲV;!*!WFy@VVi )pRޏXC$`=0h0H@XIKńEm'`<׎%([aϷH';^QTiqkRՍayI@H*EѦށM߯[OÇ$P:)߿7 [|IUږx bF謞2.o߿1b1__ǜ" N`sA՟gbæ-3zsO6ɖLAr "3M߬ocM)!ňp9*Ys =un>NAL?{֤ȍ /=кK943s{bGJRM3g}S@PE](L )ʔҰII1$kaphѿpe `ڟ%g6h$)Ny_%L@BhJYnvc"}{54k"tޟasX ǟ.I0NqgGX?^hf<7zF]#ʾ&WlҙaPY=LW!6ШUZ.hh:dm^P 4x5K&+Ed +j'bu]5|:P?ם(nЫepqmx_uhְ[6+ `DRҡR(H: 8:,[a V)A&%IkP{֖(c}j)!9T8"VMl8֊j*Od%OQĂVm{pH3 uQbD\ה`@Ԃ0.w(zKCPRʉMU4C@:AJpSJ4KvjCMUjRq+pnon4@IeR@Ӧh>*Or*xu664|r"@S(~̸D ODG[B vH4ؤ LEZP>.F*f=/:֠ DŽL&*%%M4Z%Aإk:XadX h *lCiʡQ|̕$aOt6A+։V .c d+: ;wi؃jེPp&<;H|OZ瀦ah FydRyKpk t; 2k6hj 1|s"!~ {nS )f2եV-CP]Qgti]Ww%XvJˁ1—A% Gst(%@kNV:B٘- :6[*{(MY%D8/H3*ϵV?zٻ˖J.G]޺T1 D2%F#%8FCcntq[E@?\e\=MT4iv; AYZeB8.)V;ڥS?9/s9lW 0هQ4e}5/:^=}tr37>+Mvc*KKtipԐ!hl;!q~IMj/">?=dϋ½Bѹĵ zv=}=}-uiGOd%n@-skIs3pK Ȕjn['Hz~u6=w|Iˁ̡XiաeYӭ30C\R/GZsշdpW$pN10ֹKEیTc 4umOƽ/̄iL>|+""EeDj7 էKrRy\C4 'է@+pqu.ZeuA/9ٯ>ھNKC-(B1 H#޽$@kc Rxь*zaO'S*0g)Ēkvc/a sS9{:@fG'K-HZEY5k)W@BTT$?EvOw>5vqW6"j e/dE ~3}E軏?~ (@g"2 "s.V ƒ$ΤJ'COw)6vYTWb~f7W+(|[`3B .oUr.MC_4%_M VU')ᕴ)x_2,wVO˫6^=)X5_"Q5y@gϡkWbb#'(;=[ Cm] vVʌhL[m6% *0Y+qL৓ ^)->7yFi#3Ӌ=Gd2A gCij B v[M 7 lÀ!^˜,TV*пJjPX=RߝxuQWuRLXW{bDjj&jS@eRϐ;UQL-IݚR:Sih7:`;^n9Pw`q{(^W2X2J{sNoGvm *jv(8[%=c 0R-848fNcxrYtف:c М_aڂ\s)e*s*@:K6\Q'Q{IRCj(K 1q*D:&(MdEDfVT 5eDAw+dL$BjU_ZSjJtEv(Rԙ䖹eKjxUiINڤ0uJЙ`FU|s|y(u&^9 ЮC̈pz3,H< /MUB2j#omtg +eeaYPKiQ/|Ii<K0gGpۛ飋LѸ۸{>&O#Os盬Iu8rGYxjK9_B}6{(֔.r\k^L?MyЂc X.^ΞT3V.4C^2TxDWmƅQUŠԩGvF yIkҭ@J:f+T3W˩"h:l͊e}>OXT|p o \ZuÌ%L229A~Nni5RBpoMvba>_E95ˉ qI m0)|Mͯ%g~Oe&BIޙh)u2/ײ3W{hlA6D6iOƻFk"~ey}0 gu(W 0tw% Q23r=Ӡ\M"J#+ex_z߇tf xti[=g\>C;Zԇ>'0}&?;L#ȸYo -MЗƭL$K!֓DKto$\JRA>qTS};0 KHL@R!I쨲*UV'*1\`Ar- P-ť-݋`UuVJ JY~=z)V(Mobk{zY]y; cDZaZ3-jzut*ͺ z&G8\}/RFp8s6V4.CZ!w7OgT4E9y߮6dO1> ̣m1UJgJ|/Th2IXYmoǼFD>U M1^N=Ͼ<e䂴riW4ѵ$xyTp( Tl85Hee d8 hN{jq}[eAx"na9-sW&DvYs29tl e%y>zT!Ye"tZ#x)Z.2|AӬ3OՊ `k.V2ګ\A!U^َ,^^*pJAS аU!*0 ҁ)HHʙ,/D;-1ZXo4e)J$WiIӞ9 9U 8cu@ZN`L)5/18}p;9@ 0 S)K›RP ܿv-E&MU֢48qR}v~SB欤/NBL //SJv7; 3MoeZdys{ P!BRT(hkXP%()F ?~^[;#b^*ˆ/WWo}w;4Tļnfw;߱_so]~Ə~h:v:|A?SpNvX Cv 4=v VT0R ޑ a Af "?]2I dSL\+4{W*j{q8x6Q[  Fً ۩ 3xvC~uv&^Ti,:#0Rq0@nPXOyHC#юԎ[At9: PNzt)6W.0i}`tsKqEFK.z۝r.dz,ۼ9&.CS-qqkyif1q߿PH&\~nSpdmNGXKw0 N1| NYi*)y3[`#J9Z/SW5*/J}э vR0Vq}nv+hy-Bxsi`==LYL&~ڭ 8_g,T(ǧtƽ5^ce`Ue\$;4pǔ% 8is+9F AS)gVWQ'ZmJ|q^P[29A؜(a$F(W-E*JfI@-D NMf\ΘҤOӹuh r9rZGχ߶7P2*g[cĖuVDZ5{ҏiƺCzDY"yR?F.AÇ >= YGL NMh0-]q@q~.@h1pK0<˝!ZLesh qdS-pEg*V T` %{?g_-mT{](442zf7#U1(}Y}vRA (߽ )*TBth5um$k%op?FnN~}fOoU5~dZ͕_/xJyg#41I,'L-*VqHӿ6jig+Y_>e+WZo\DdJ/]ڍ<-խVxqK;2Cğ؟7JȮnUw-O|dOwh9BV$-">y*<}%'D ̨YgcQ8x#GXxtⅡ΋z2|Zq,tRY~{&/F:yhԕkEeP⎢G2LDfs`u8>U0B$d: rX Ӆ6$(Ns5)5YgMI;)[b },V:潸FB|K:ҹTYZtcכ)@r^OX_OXuSʋ;"ث5j 8HQủ%TS)G߽ͯo+-W0*) S]25ZBt% K`*]D Ě _A'a^Q3Pg\E1boo>sugƉ1}}@?ROih0.j4P̩TEC%^H#}-Ȇ@G &Ǔw}]2iWcz 4#_-2Pz,.J>f_6u1T\QμhW-PXIZQ0-4oWa ` ږW>[r 8V3PRZ4fN(aF 8XPtamZHsAz+~J]V!xWc̐ Dx|^!{Di 8B$oCfq2ju}!o`Ra'uA:ub(+S}s`q,ŕ?*X̯裡ؙ|T:U"c)H@y˭Ŗd07Ǜ[D6#}ћ:ߋ^~#x7Z)93vDs*%\fe@h3Kd)r|"Ӝ'ƈ $Y $U ˪]Դ9JZV 9B=ޏsR9a)=6d'MQaJbک3i պm p@6qi/u EuX +(eXhhڇwKLinj$֡VLǛ4Z\qeS4S͜agågW{,[4-P0b9i2MaS*q. т-14׊.FRI%yza'`[Jo}e< TD8t-"q7At zY M'WHA:tʓ&#rPmKו+/]We*;(qL8B;DIʭ΅KK]LduR<1E9b_Xth2 g(Ռ%)?%rM9&Q\T3EbHv挰JvZ8%'9?!3'܍ S#6EHU*bȚX\ջXko?:_uݱ{>Lg /?nKjoo_nXx7ϊKE%oor  g_]/kzAiw[΍],_w ~O1v@+:{?(8X ~#ՔjA M(*f j~mۡ?߷e^^Y\p?u7\y4_|Ef*$8Gq%[y|X$0D`FyEa8$~L&P?v63J(~'١Rs~C^HE9 yBU\9N{&*Iec!svaAZF;CzW"&o]1cN-m9qwEAWcԐKjh![CcS5z TGLjg5ʔ:)]"\T$juF\K dYע,ݽq2bQ51=@uR c])5Ÿ1d\j!Ous\P3 Wc39jTHiMLLDAᲳ#f:DhL=DZ]3pSbqNxxY}u>EkWq:%Pm'HbzOcʒ Z Q"͑5 Np͵/|c;D-@ 5,#=e?ώ;`yG ZTв"@+&eXS6GA>5\YfMjY( ډY3rK H e.ϊCzW(;GAJ$0R^#E89L8Pu8fSbPm%0#9 y8cxcPBSc5Ĩ$\nzns9;խۭ%Yu]KMM0z螟t7l%7A$  P%)C#.iOyԺ\Ou)m9}jA.P9jrA`SAu[gT숌DGt%1lԵ2I#naN Q4a#hTݡTcMvSW|{q:V/MKL@j/tIKї2Fu\qu \ !} ʔRAoYr s LP1EJK]̨$sID b rV?@d90osBPQΪ ܤ ~єw%@1#9@B DPbH<8Lf.w:"BZ<``T3 ZP,' F 1Kh"OvuD+cH;3 LeX1t㑇 Xuӝ VTӺۋS}UۋDžSBDtV*Tj(2k 5%2M\'">gfFhOt 'P)e7>_h0ʂjz j^CBQ5\qkwq[,V&|u/>rG~دFn|(UyQګmR4oIښ=a#ؾ>%oe&;-0MZZ/%s>DI }k\]jl5qc:_N򹽴iƁ}AS#co;u;=3ARfM-?؇#O_\!ibg|E,'jh;/4{$4]KWZv'뜹h)V#yc?qR@<܊uHfYLSj)d&De&tZ)4K!?nEqwS+H c+oFP&KcCH ̗10bC:m6Էi{yUY>gLh*EY"rZY Z.]BUnqNVCgv9B/`RyڽbDžgKѤpI(TUf?$YqgDH >ׄ6C ct#5Lc㼠6QI!{{!&G u{aDe!0FO_?%ձ5>C[aaGc=^s"G`<|AkOڔ0mC@r!U>HcهpjgKfPʵEv7Hl=d~dM?Vn;d 3MQk/"1xcŻ!wU[je7[w]/}7vT\zթnM] ii3^ }] urc lZLrK AKj:i0|kC$jI~i>~=HO0Z~0i-S)+YD<["<~נ ЦD܈/93ZVV#oeTRLD^֖a/jQpr s?{>YN-}ٜ.ޭ AfR|ƣC`el {6|j@ݻ ln?0W=9]ց1|EQggK2)S,f\TZ"7yøK6b~wT-[$f/&O=$O:\oC-xjpњ5L7JhhKHTNdž5L-ވԭ :P:%5fh߀6~%IWdkh^[pJ9@&$/kˣ (!FSpF_3&h 4wJ6{*7*@a!px^76 B0480"8cQջ:6(+!5L }]#:wUo)v_^bG/ c'63f;pm"E][ n(pAFriѼ}L ,D+/dex$Ƶv^}zʡJ ؊4O@ѰoYwsY+2myKmv~@N4MK hڲVBp sw$]Z 75.Zt^%S"uDfR1=} ot 5n&ݠ+ķ~,TTxSXYˉy} wdl 7klH!dnlN I:q&\\rMx4J` lu/g&>ÆIGw\b}dh$3:}D"j8K3%s*:v Z,:o#2LvX"ǑupJch{跪`J!бz&-Qr&ϥsbGЕ:ȕ鷷Ø`h?L|G[m؅n@@'Σ܉#S.d2+Uԁ`&Y!UTNl3n ֞kۅ5ND`!lӊ~ۭotF`y__:\L yL)٘5b$<㖈-UxFhnω8y(}SD##i@vupF,-&\1d8u<(4cwa]# mF8#$^{e=`Np.AK6B۪$2_c$!%$Ϫ1-mgܪBPAE^ e*}1KrURUq2\'bުb8EPA~fd:ҽtBw?gI=x9CQIn ߪYjL=(l4>p79"(,Xeڬe[bcWg4V&!%d:!Ph(sDq&mL 1\fi{+\M{yEe_jb߬XQL/x֤iz6_߼7p&z^o?vݣ} vpE@!;ٶ8.Nrۗ8גSpZ}≭7څ=0g32.tGonRlZ/.F/>M_:jzvV#캚t"ԌK? )G,a쯜M?Ok- Q/ {k'ߧm[}>A9jua)Pt?7]>_3imNݭ ?jZtF\(IM/Ah/z5(H>-&.?ŻtDYG yq L <1Ptt̊@Y:;(0CUpٖ*ꄁ,0.$(*su8A,9{B" -TE#Dsgog)WO4ͼWX"?8Wl-&TXdh, uܨkY(|iS9U;Š "09 RhgH^߂YI0tKo%3ft}#@[\ޓp%q"$[ԙ7׆;9v FQ| na ;>t@V'2vqHmKLmym-a&ˈF(hjC1olBM৷=DQJ99\xZXNpmiK|fQ] =ڞg֎X WYz4h#5nN$4/"FN$s_Cj=zdvJI ]`Q B.J݂\1ŖW`'ZuYx3KEb%ύqj!67ú&0ikV:XMF15[lTf䊶Ѥ [Ӂ(S'PctY@0]$Q fcoǼ:8]vX)k!xMH y[2I~sdj IHh)^n'ѻ ,\z+my7A7k Te7M矫RJUX1tP^l.\{ Ye 9#H|R7`w)C'޻j;uPͽ,+~s =$&yL2!Gή?y}T6RCd"ёa>ZW#'^ϣڷ_[L>\d0ΗB h;VjRt1?~2@<^WMOl= 8TC.@:ЛA)H .pQ,k[2[ kFc Bqdw^Lp)h+*%-( 3[?;B"xo8S+a&݃BM"SAw#w$ V̫\1d!#1; z|Gh^%|Gmg݈ YqᆺBCM?̿.u(.}_$rTQZ{|'Z.&q4ɳI51iw'糓Z7pr6,Kܯ]^^v_$9 B.;/xʒGW:s2[|/hEOk7* :Mb,B˭Rt!b8a0W{r,)BlӘK)x`Qd(2sM'޽CLE-[ĕ(`:.HP ,.TS|a@o98%NQWҿImDHvKTUSiܢ!"6?D??~9~8W>}'9#DpKՒ?G}PNHz>Ш}L/.]3~ؓ+?WBoO^Fo=fAp:^ms;vSXޭιaSxʞ:;>tӻҩ/Za¬A7*ˎ sw9Xu[>9lao8t5EQE2pX|JDArF (2yE \ANg2]N%9V"{؁J=\,3xmmD:y }>-w9!xej6%bdP. GI"{!8tm%m3⹛Ѝo] ݋ȥopXԃWT %@'cN9 Fl)|P#.<<|w;١ygl}5lU1e'*攄! l)%UͰ&a!.֝{,;:ڈ!۩Hl[?S?!5kJIQSthU\ݕX}([Ug~mgS"dVh#3aPdL43+,BaP e/%_2>K m&s0_tM̧r2dn+=| [#+4i6 jՌY,)<@|.qi"݌(bE S&˰&=,`0r}.{nW|߳owm2'E !PCҼ`٫)j2y>zPM s BVb6Ke@$Ju;/8l2Rm-sS'p/̱%(32P4EYPϽ1ru4eE"AM1MC[p=LR2DɴBTh Pdmp/䃢j6`HĚ_)(ZvKT7wKZ3obo'r:}dB8DɩM\$&spL]X p UHV1ou(l^ILP6J&/H%PtE"`z-lwٜ^]&91댻<#:..{2F.*D7 5g`IK%p 61} ]s]Z@瘉0)z:8Α.QPOʃњ0l^5$ħ_,ۚQEi g01wb%8XJbʡI7YJR @XMU:;Ց¾EY~BDSHsE`mbJ,RER%D@QŖ &a"QbCNOdTrc>)\U 1U39"Hs;ı6hwe[YM뫭U[*:| zf /6F'g!uz|uZ7ȋ~izkar\üsMmϧkP k?!qJ u}=#%\ʉgo=+U%rUS+ԯ '.e:eU{(Os㹝˟>6ZZx(21^x7+Pyh?_䣍wAqX_?;eD4썟s3ʜm'+f~(Mȍ>YW>lwy|=z:Ko0U_6u~wpofL!BF?Ϫ97ۏK><}H>vt7v#cɇ${+n(o ,VhË?{(\6˳y+ýp~᧗f7?||w_џh_9S1nrj/o=yW?xo3% lbe=uetx#˶H<[8Ì|`Mm4Mu%wzzzulOlW5gjY%kv ߀|ZɣF^hF&jKxgKxgKxgKxƎ5z9&Mu2-2-2[f3V,pV.M)wЉkBN/i%)zU[ &f SoK %0ڂع+Ij" 9YR=5&D 'k9]n)*4ByI |,APӠe$IE:\_$d0gP?{)v@i%e1RC]* N,@s`K)JHڠ5ZrGZ>DH#KRzhwPfsDn؏oh{5A`]\'PJ$",лQd`Xc$b#n?yPX Vq4bWb&233ػH 3hZqWowARm pt2K7 [AIwA& Pwf7iޙ-߀wf Gi6wz7l$GޝBѽ蜍CfQ]3Bԭ'eVW:3:M< )3bg}ZLտ 4+KL:Rux+ؿ]2Kޕx Sง< "ݲYJ,VדD?$KBrQ͞P*a!霎91:^Xgj3`5CS,ȉ&bbFl$ @jWc49߹;H(G@y)0EH"'+g@T% .m0#I)H- *U!0=,XJc䷬%ҹyМs֏1#ht Qłf&D5DR O8ѠĘ4&Ź0Պxss8d* o˯f'2#p.`iM2:)z|E |ӂd\Ь`Z>›tkaz6(xN@ ao{pcD궬! (>+xgh;y3BM?+K? ~_.'ӻ:q??Ȍ۠_owʆ% a r7b 7/ܬaKe@B7 [@D*B(8Vs۶qo! :S"m {E"hv QKs~`o`=yKZL0{ 3~E`y G@ q__'\O~Bz7_Hk @z>%eO*KQ'U ^~&9XX*UG^5OOVY `Oj7RsIJTߑwoWY 32 |wa^7`povL:W|y71)ZnH]$\\LlJ} >dۼ$A&4ϸ[B-x :.Srx+J1}1sA>9( -=;R9϶,[㍑tDXxl^0}kf=+i =AQd^V w+%elǫ6Wm' Ӷ2cp&K*N~.NJK{-_N& o^JټÆz8F##(*%eOV[ zj"#kԚBs:Ih^MdlR8ps?Dgk$@r\$w0M3M;3xmՄMŧ,^ݮt2r~F2/0rHKu>ީuXqH !#5w3ꡃqk EH`i %i(0̃a9f&Qb:y` \}5`oGkMbrf{U;:1B2Pa_SIVwu : FOAY()MkajeV]C߻kN6~$R{zFB`#4l߂TQk鍒?O?NWݢ&;ƅi)HPє DX_ ט*)j>E8_eNM:wuS4jdL> K5CUm7Y ]ǦgL77E22Ԙ/P7p{Q;+=*|(FU9^˝`"Dp#Kg!=5JAL OhSƺ{ ֓ͲNZOb)Cd»'ZdOJٺlgVXfڗYn_B-Ľbcq(c~W;s`9 fɞ;mAa!9L+bCd*6s8ks8/?, b. {$`z{3rGY84;N 6J(Uه06!} gmk3 Ԁ^n8!0]h !CV@P&`&8B`ALK~錷9j",r^:nEldFaAL5QQYg6: v+ iЗd|lL>\ג5AL}QeW7_UFoN;0UcQ &m,N13Nag\ugq.1[3(`6$2꯺G&ѳTaη[˳c3(l<;ߍiX(t@\Qo2L!\Fsp!)Ɠ;e"iM(&H(g$N$1Ti Vتo4~kEJ/;zsg'{g;?mg/o.v\@VUYއd'XoBV+TlLD…VOANRetʴke4MU鼆A.: fJ0dqy?w$`I/9BT=5G+YIuS\},OVC'Q٤rN҅+^*]ʾ_!?ٟ.&x=Ͱ('Š~z8(Ӄ <6Gk.FkWOaVūɂEݢ*írsW8>/#K vڴEJt]h_jbҽ\.u,VOlo@W{Oa_)[v>oڕa{zލAqxSvaB5 ëϷ:8N0}^>8KTS۳i7/im9mQLяG=ލ-KL z}zWR}x0,N¬ +Zܜcɉt$ 4|*'bo4IYNNu2oyhNuwE93oBxN.KUNe\M Ð:ZQbaT-7U};w}'obЛ;Ew"aջqws {3:I›Q8.LN׬!0~s娔w]|3Ar7v?hG^tuؽ?8urb?JW t~`U#z7*6xbY F?};*_2kW%_. غHp ?i/BUFxoh8z}>%G`gzvfa,JE!*y|fivۛAq?]/0Hݯ!{>@*ȽrS w(BE9G2VtGS=~ɐ5^lA a&i%[ld÷El-ζYn9u4^>Rc 4_&x~z(y*y֕n72L007 Ph0Pb.ApZ-~d߽_qlijh&cn09sZcϽGhaA8D<*Km4DiY R0ǑR* ڭQ-[3!cD!*<Z"K|F1 EUh 9G@t)&IV+t>|_Q>RqW^}ߟ>>k={j/F+mT3|RLdMtt^ :^j_ $UR6xIN9# )[HnAHbːEG|Y 9!*f8P(05H{#Y3'9Jt*mj #sQ@EKAwsu*HP ix 猃UȻ[pp+%h0FKO"1/Fq)V9B LhHq*Jɠ5CwZBZJ0&N@9 IqvVbe3+4bM1͆pg N+uS0ŚDKbc)n5ؘT5UEWVms #ab= qΔŬ(fMwa$,2l8L}4ʉ@ @CEGG5zk-AqkFZjtx4usJY-s[(g)n֠,1r%~I9\I| :&2 r Y&A($*X@ y, 2xl^ɂ"ҌvȀ q;t`t^8\ͱP*GTP (A9큐J a3Wk_YBl|ʌ812dV[ ͫaPUi>C4Y>bel7Ɛ5"xA1bŘa挳T! bCgE-%)k0wenyWr_ XDZj&<+[l]G;.N3hPT#C爃[A%Z#Hn(E͚+L`Be)yF~67wYjHd~uYVqj.P7Ĕe@ӘNc:w.c#+Jx\rf(5ӹCC0߫4l ay{Y5كf4(0]B]4""G$0Jy7٫HR1SHgq LO"[( KOeusErE V7B^S2ײQFz$E$ڞ% UA6Qd2y]gN Q|YX+cjKkbDJ&@ج{%v. U;O}9Qs Z_`_ ɐ%}-yH "js˦{!J;guőđif=y`ւ.cq5ݖ. X/<;rQm^H1"D,bSE[+|X>sS}A803y=Kq!H)㔺cx< ed˫k/VǩWKK:>xib H R?X*0ɔ !䚔~{62]`|/w@^I!TRa$KI)c5}ؤ$HTK:̵EI}RQ$koM*\wsItJ@ S*#nl"))(ZszOezL&Mq6inYߍ=L4) 1?RN-oRw|TIi\ߍ=LJۍ ڋ'#ӳ('Ԅ3D-b8{ݠb/s2AETޅTpަqѴW;;ݦ'6G?;ajCٟʼnsOǘwި9ǣwСǩya70@OcHnklB۾y(Ğ-OC[2c̹X6n /h'j 1)q^$_cq0޸;=麯n}qu}Қu$z̺"cf4Lnsp06>GOQx<00/у᭼fKj=qr/_Hxg0D0hKGGaVz2>ݘtD.O^߼<>d^kƢ>@QW6nu+yw;^2E#F]bxVfc:>褯 < 4kCr&xc̸4.ޝ8ƟNҾ(S%i&, !$n;WIKc2(8OJ֏vf.55x#x{qm+Y:So=e(γ_T xW^l?ʸxk|T~9q}Q-l;E$_cJHVT:JEB'^ ҋxVB2N`W\7+0U-K FjN.q}1tJ1pokGqG? FھQ|*<~9:S1/.OxN{tml^.E{El~ #fϮbQs <<@=L:=sUu}9Vs1wKwq{R|vG]w;"bEIg77GC{hOFSq!-ԏKGw[ $}\`:wżn-o澍T8~l@R XEC{ݧx| j?'.T an-Ў+a4]S%S/ SB9y^Rpx:Ͼ]MȃwB'0% /눕PB.$}1Œqo41>!#$r@42ۏm7s_],;"xeb书Q˴-d!BwmE\ɼHMg+BP4OL4s:FW.$YIi؃{Qy etuXP?޲RK&CkY| 7vl"҄hs1M>D]V(\nsdWEJ5R81٫GklKtZA s;kgۍ {YgHcU+Tۥ/y[B Go( ^bT#s8%T0Z/BW+ jDcfieohJw˓جw+?I!րap3FMl_^݋9m&ɿε`y䩄1~e+ߺq4Vz:3x~( %߷6:lCo hW`) ,+Tln'6k7fnxuHRBD} CiC}tv5B`#;U䚭J~PvrƘ^D33$6CE;l6a_|ѭruEЌ ` .mbhY` 5K;gBWejr31r''Hjij>뇡8iZ#bw(C.Ą#$^LL0}A9(9/﷝HPNZҶAL{#2p ~גm:-v5m1#ޤtDV;1&Иۄ8 \y7)ݖ}L`ȥkbPYlJSDjCJQ T8W[RL`Jh*uωTX!rX;MlZ,*p:/ngY?iɐc^{F,+V ItR*'VzI6aZU`iڿ=gUמvY9PRZo}7grl4)p9@6lW jOqHVlD=O?v@7h+^Hw'9ORYxth FNYb4dZe81E#*gz 8 ҰΟGi2i\ieZW=x`k ,!G -WSǭxq+ a2a󐘎EʫPyjOÈy0Lccz> mf :4 -Wcd1Tl؃ 䑈bRCAŃ\?NM#3X\'fH#NDkoJ "IDa^STll_R5ۮK MrE0|Xv`0r@AݲfPڤ(2)tr!3tf#>2yUjbʜV3x\A,b6e@7qLiT1dc#+hUjb;Um 4MA=aDae"'Ih-WsUu*I=wӣ*Puni )vuP*H҂*NBԤ8e128}A{anOEn荥؇rGvj&NAʼ 30 u;Zhs vk[G-ҽJz.摧T$Fk\4gk,fb1)o]\)U'jA캟 Bs*&jDʸZ.` NisҞ~xmc7ܗО 89E9ıHh ΗL_cqNwe=nI4;[Rއ=x%`6ưG0M3i.%RSs8,Dt֨bߋR`QT[_5T*,՟ 2ɉqbD4V)RУ{CFUK8FuL{eP2hPr:D2Z$÷Sև7(G)p^%S#^O2p}aD yhzQ m!>2tH)sD%z`>GoFͣ<&Oˠ}ˀb,ĈD+]LXˤa>eQHg3' De>MIf$i0:ݻ}:5D߅͸G)Jza!dm3߁oٹƂ5Nu_L[7H#4hp)J һ$~Xo~+$f6_ޤ ?Eo7q,4tr:nv(vjrDm08;TAL<yڨ~9ء t4Q-ۀrXm(X)EՀq ($pOtP*jQMN8N&QV/;R~4yuqbzsUB>EƜöRHmX -vx^wuQ@%&OE;v@MeѨ/umPvhI  7N9uX`\RZ9I7h)70qNcE Br윺 [!5 U['tA}rXygf6G;&;i/$t/(?V!-oFL7ɀf`MSNe!z6IH8D1a %VS"%9Tݮe?j&W x@ӑ Y{*Yri% O9y+ 7|c"W+C3Vb uټL*hv_=hcaxma⫴OJn-w}rAG^VĵYt&Z&lu+ILL+"bp'e@Z- ٷb]V殺6 <`5޻O4|)w+#~%R#ͩG.ekڇh'/ Zܺ;rwo'cMqeCvLKKTKw?yn2^:/]Kyˑi4+qiTc+k\M:K[ . yۛJrبx*(8=2RQ`DW#DwT<]:ڌR<,8CUkJb0ёRT}S$rѵGVegOZz$%XˣUXD^CػZ|D՝}~u H&~e'swQ#X;_wO/z(ĈmYm.Mrm :;W?rͯfv01&+H Ie=$5vpnM {ǫE|:N} d:<׉ ^$AeN/R?^\+KL9%K+ 6YQTGoxg;Q)=:Y]DCkonߖ )!OI<G!-vt*/:el{>TTT5^: ۭl.5f[O٢- h?7j׊-`h4UOuajΆ}sO6 Qh9r˼>FQ}V <ͤM~᧿s(TePwOBTrO#u16&#"+$6*/+D dSڠM(V"w7([hfJoGT8Ž`+0șT3B"|Ciϕ,[H$`@>bv^L,?=,87QEV)malHȈΠZ`A{Hߺ"|`T18r /zLb4qWr!:]m(` `T\:I騍`iQ(_A z=5hiEi#-*U+N*GEIrDQO]ekpNp_o{l`/y:eĩdlK\W .208b6tC:)| k ք(Jc";YHx3-: -&Xs,ƒ':@찖WB'6Bu}JoVd [fk` ,ǐVzIg"n!`HZ r.W^f0!8v խ&jS}@DnK!N FM8%e!a @Yf9]IkİN7[_8Ez[49<@TiQ>?,"M$310b'g:O 9 4\hG:"FG$o!9I <( ;ˬ28n<9-:ǎj_ qv={=x%cVT$zvl$:xsU7uZ|nҟ; }׹K*tV6=@'m{fJ>$mm^mTփmDZ{MMBFSgk/\[xr3d0*}5gqn={gW XstѲ#( !ie9:K)`!;;󠳣>Zi1H.XJ"^@EpL891L(T(-vF ."K[C` rX*d(e93=%*;縂NO9egrP͐hw{^AC{ mrQ%RX&nbNBfk 9TsG _ ؔK/'ƅ [i8a#KAdWQV*cؙp☴@@"*!oL{ɸV ڥ${afF8FD=t ;;|K2)px Td1\ؿ4v5nJzZ.ԗ$ s]C=(6(kH0< r#q^]GDddDdQSswRukFv~ɁGu>hΘgD,`Lh{jq&c2o lRl8Q" ~nZot. { BXYـꠑHAH)n@iNx$HʢXXKkov@ ,EèU`ꉴ.\"4 zu}{Y dOR07a7% F9t?*ƣ8O|kZ {=$8gIX?d'4^fNFz,6fۻ;+nQ ɟrN#"[64R`/A)H{a0qK<1?Nt?)IRfG[::yf m)<13IgDpQ{FbRB e0׻Bd\vxX9Ș?5Yz{+H`nM|}phbڞ@#<w D0~v} Xz @=uڨ\ZH:s-j|z˃sF&QKJ+؈GWU&e}ϡÃWBC0ìC;)3D`{XyWk5>`ŦJMӃ/beLw4۽qM ңGn|1*qo122.0 HDI nź݂\{g HA*,w=ݷLc;'I8aq7qP`9ݥhAe5UrYCțu(ϴۍ}kk3-{o1m؜]m_WtGiB^8]o3.@.ӔEl*/8^J\w~:_DHB"JdT3j922Ry(*⻝9+.=|X:[WEvDJLg VȰb1.1-s~|%wAeBt8|#Q8ÙAKDwE%`z5DFoqPM4´r˟8i 'b"u,nĹ-bWBF,I O"F:69w )!D Wz+yp܊<,VT÷I|7"F(,+h2ŀv")8\34a)<6(f!(_G0@w \PNP~v.}: TܼzCiǖINb9ԥZ>)TN&yʤH2 SYGe0klAjZe'MPz#h͈1ҺNRKrJCNjC;k@l|uSc@QY o@֜Gr}>#.`KH@6o(ƻekDmTj=ޖϴ-SKyeNQoquT#- 5@HGG7u8C^g/E9&IAԀM5z yxpGX e{yʧo_tr5Xq5 s,=8y1:hW|9kJ/lRK(W6Qf8De w˵[͐uaq7޽/>gj-d Y9X;u@踦QlzCjCi3yOGB*\^)(3T'_N&x+P`5%r(1T8.|WfFEl S)A8jb_.&q%QŌѺBqGLb)V{VIjR9As<Dzπ_/b@$aډӃy1|$ՙzueFT "QMx O:-t!72o@QNb 9* QyBNt`s>XX#,8۩zB)v!_VV`Au;k1(wko"Fp'̀O%&e]?0I l2{tRL_>=Z|&?ݔ/8oUKD0A\=Z)JdO0}7@cl^jrF_\h g<K!=~ #pvSgQsqvI|!Ol_{`N{LK*GܿnkiH¸T밍z0՝ \vq8oϣ3pf6/  0K-Z48؉\mzl@L2G_2EmN=e"Jo w439 R~[kG3$ G *-'ND^jϔXP弖ޱ8%baXj\B۶۹/f#y^din.J.m*3l]U[E BUcPqgC|3Y^ 668x~)VmJW1 ] s7+=8m܇f;%g\X^Kl*v@[L(R&%Qy>tb'ە$U6N=O:tx=7}?'vo Ps +ڤ$bLkB1A0G$ ӱZp<9?~ HUrr1._8>}Z Ss34x~zﳳ5`0hrm/U^[9mq,O,xvtgG?ZTUƸǿ<Z/r><9 ^|óςWu Oᅊ}VbJPEr3 [ٙiJ[uUmS1kY>dz,fj\!eU'&ҚUdܛ5g:GBJA U\Tҭ-˯.ho GKǣ #Ùq.K({7ɟAgiίFa?zW`}nEi~?vW\ )y|i &=/MYv2}׻ogόƷ' N#-Y9X0Ҷg׃S幝^ p$&O/ͻ8Xܞ/;O,,?2ك[Oϳ [Bar:|:ߘ=X:X?U6۟ +`&O]/9>KZ;$tq7BGwx8eWMY4LSsļt=7[;z1p/FHt,}>u=:c'U-k-~=L+G0냟ӣ0%18>o>La>x o|\r뮃 EߍG7WW.Ψ}?L`V. ]ڟzxuu8>]1S+󟦟@d'SXT|^EvvS~͏]"gqEޘ33YB L\{a{{~ћ;lg{vg{o,skqB.Tœպy_, |DHJٰ\ɿq^":˷'~^{cWw˜W{O_M}7իW[m[ "1Y!1pG'dz{AN\f^k #ě.[viuvN3]gWZQdӌĤzb"vNaWf]=8k ;s UaWv%r+.sz٢ԑmDzaӘcd9Z)m47II,Xۈ0r*")Snb"Rɘsks8zZUڈ%sʧs3lkUG=y>F a ,E*!0 ۯ@LS ؍RC*lCir K !Դ7hf]vTZi`0,AT|5k, gXcP*$$桂3&D5l\R5W;p~qBhD"LD'pl$ I9yAjFDR3ʨ4(&hOV\\:TEX2LtJ)6PJɄj(Dn$EcP+pID*U&uĭK@͡X\ a XPF f]#"eî5t wߢEJkAb`0b4\'0ACPly%S8M" |/l@TlՈiTJZk,RoHz>*nчvIZƅk@'$ZnZ0L0/0,c9ffbN0x腄E%YAa͚ vc]׽5}\Q5sk^jT[wU Mu.C ZJ yi[iSu7%z͸\oE"5ARDU) ][r;3EL5ijV.eREZBa/W?8=Nl2'A,TGz/՘uYT_dEz݁wQE_rWC+El ,ph8|k*@t?Ed:-ZةSJn]\jr)E#{ReJ@$Ԓ0#@Hc!$F,Sj"bW5Fู͐&c5̊[ڭ(WfqF/Tg]Zo`H;T6N0Y~1S]T狙yn0_̤$uW~òZ? }yS`QBI֜q #4\v4r՗?U_J3is}ӈP:1 +l؛: QM8 3\4M'ڍ^/$ﱭ՗' J8*0jcj`γ|k)u:ؗϪ$uge~ I 08猹Png!Yy޽f*kp=ƂjΞS"y%pNּ9O.yUk%nEBк#N.PvSIkvEpM'5l+Y:T t~PgTzFzK.~}瞍> 9C\]_U9٢*Cy%?]$0SbRDzN˯<GkT'v "s"8;[y ~_ ׌s5THG]&j.B |wFW1hiabĄL&8AMFDi*LHaZcԂ V_&8歽3a `d+.k&?6Q&啤 j9/1H8HО٢ZJ_Fͭ]UFz[QeH("]"ZhꝄ+ZkX(R˫ n͗7K)Ӹ{.E7D[Hs8ثz(Ri|OpM.]CުiSH =4&E=-8arXEu,q9* A%! ҷ}N.3B5T-ڌ+[Y[Iv.;qW2~Gu<>) 춮.iؑN/I?vB{M!AK@l#mdm:.2OD,DZ~䶈rߖN %#_ˁ}ʗMks *z ntژy.<\*թxAZn(1j)Ha5݉Uۆy;桭_ӭ8G]筧Dt{FDƮQ0v(FH)_W1E gMt9sFH8oTC]o}]xF҈;جB?|{0`kޗÛ ?v2A}8~4G: 9E!݌_Չ%YU@JӤj Zp]zSNשlw04)+`5!X;1W!kԭlΆ|fl 9%6쏄iVIy&T OEo:`݉8j-1ވ{Ȑ\u,J"CKPJ=d!³/(uj^g̐D-5\t%{8I"AX1U=.Ҥ,c; 5 cn'71wƸKENJ]T ְ-W̉u6ȯJX]Ėu|3K:[xo{!ykbi.:wM]־WYvzl2} /wFmЊ(Ъ6ך7 tz+a2gȭm8;Ȯr)H(`%[ń|*NT.M}bA|T4V0N/^jbߥ>Dk鯜U=Rwsjb=TZkԺ[m@+jͻ$.U?lrW[{aWc!Q֟3EֳB 94vh>]PT0+֢)Z.B6<4F-^&8m'%ڽSKb. բYT;$ jIP1}2CcLNl4Xs=_@W6ܤ#@(SVQ w}*Lx ii?|*8iHI:AﰐJV1Qm#EA @ldYɾ~qr7yo׳.\RY=17|׫vʻ۟}Է[{GvuR n"!1脦 %F+b"aMD#B48>AKRRYF$aE8#I3JklX&(Q$u F^ƺ'zQZp0QlFʜ|_ l'0;CL D0J P8M" FR >lcf21MCתpA_mpk2,bHz>*n:]mQ,U*4P 9&AHB@QoV4JdHDpbt֙bZF~rUc{^+G{Kbsm!Fkt_ĢqI0H'q+,Ԟr$ȥ)l'"d8$q S2bJ55dY`)(4qQ v~"0/կ-k78 5*x w})#ӔQ)R, rC V,5iLZ p92ٻ޶nW]lO _&[,-z/"5^I(ߡdKǖdM #pfHQvE->k0<^`(bbj>ݙp9|E۷'Ebjak8]JlJI>8Èe_=A Y"`DЁ'*)C̎-o$dz ˶RjpK'׵ݼ/N/ЙsZdJ8:t!u6%TGW> !JU&}C[/H9 t,?X!3Km\—M,rqn$ٔaZ6k՞" 4J\+ ԥUTpj&KVutEEee],tk S _(O`!kNEs `^D 4\|' ]%x46! Akq?=R2Kα 5,byc@o^@F<w3m%VXO*3CF<'Whgs"DֱT]|džwBho`]٧O)Bz@/Hv|h.A*ȘЇT(tĿST~N_Mlk2O|?;ţ+~UlF޽qA:TD-ؠ-1Qz!o$шOLS"D"A D$L!@J`㬞1GrthﱀRTU͠wh *.V9#TdHA TZUN: GR b#(}JJ99&79YvR2o~hLtǶ_ǡl0Կ!{2ToFpGk F}F7k0O#GݪE啋ZT*ƔIəG4;j;0zP^}=x׋~yO7[{=Ywrӝr?_29:<(h] s/(L/߬jٜ6ԧTY|V|{x$Y"٠f{3Fl+Mo@O/&0຾A/p?~z_ꀢd~98=6˳ɫj|Қw~7?]ީ-DNYW>ޝ>Lty>nBM;D˴` OB7FZ" gYA;G:pZVsi0&\ ~ g$]zwiVCr$zwjG7-kďqwo:f*6imPW95 O)6ãJg>50Aiu~tGsquJU@?yvudh>\N5kIb`x(&Jkvn2˖8EKpk8&q6{q6lO}>U,́e6fKl}7Nخ39`W׸}?7'SN.F$qDxM&2@"*/L6H7(i/8w;|No~/\"ӑ}X&D8^ruHiۼgK׫qWM`0=1W.{o?ؔ4y 3h=k{a+$ah Ac DDk% .Q+t`ʂV-$뼩[K%Xe{=ʗb|VRƝ7ղS۞]:1LѺv1NDkTm)}$nLJn!B\ BZ]` Uzf45IR1+<\? pz.eMUviwxI!#ˤM3iӅ3 2ii4Ibö7t'H.J g/̻ST=)D:!BdQ!QwK0gp`TΞAs`PT:^Y˂bb*pp?  º{*3ԭmfÏWϏYm5"qحoiLCev'+HfXO>&U&Xe\f4FJ6DXJ>i 3`M8մV TLևU20XAւ~*#,X)6%)VoXeas1c h:)HuAiHhpVC0^{˴79)m gVQ(Ŝ\׆vJ:0S6]6zPETGg*te_E/LAI NJcDe_rhcA" ,JQ7%Lw@3}EpA1Tࠐ@Nv)zQK N7[C| \ "TG~^%ݺwuȲ {B"zR >2/.fƠ$sMrcNLJh$x- k/5 pQi!B@Zq EAҊvV}4+jf;tiu\u\SBK կOhIx{3ٴn'2TSw6zPA"TB1='s(PxggOۻQ!EyZfVRWaQhW5^mA",+Nuở1]H _.]\{A;ֵU/>M7'Jjl y_+f'mcizsuՒ s=ѷl =Zy!蔢=dbe2.BaF3aF:$5I٠ˬK&Sd^L:Z!X3iXITX rLJ"%Wc eU0xBP#GsRtiZrFNҿh9bGtiJ1^x<h.M@4543W"R䝃MIltQwbݎLx֭x][ UH? búiGntQwbݎJ|tT9xGVCC>s-))N>޴n*yuDVA)}!9y[@ZhN HOnG N_ 3A0-ܦ/ߗدL ʿ _z3S)6A^Npŵ8].2smo޾8;`3|k|ugu-u9y-U|pKDHq{'t?(ZyF$n8CFEqL@8!̦C'2:Fzusp~3h?[κ+u% Q֊>∢a柜(D[lGU[qk+.xh@hbU R[]ˢ⦴PՉ^/>2"꽿\ͮ.~탓\3t5.܅W;}rKo0-ZhG 59iଓ\$K UoZ5* EAMpB4V݈pt{ߡ Cw,໴PeJʃ7Ѫq9:*7D x|]Z1xkCigR&zGy2AwhA-.nBcq&tFt;໴PcBsQy@#'m<ڬ |*ٞK 5/>x%ajb*-i(A,3-bFUwQtjRj|j ^Cuׅ{Pc뼗e׫4Q;(Vߣ]Z0xCmrL)k OYFdPeU\^{7&Rs-d;P>xJśV;Xm+~:=^R1Z W0f0ښp$pW F1Sr30CdeTC *y}S6^.囻 4yYMgT5| rDbÐ]m]-Kwyxg=kQ 0I4;`3n;~:i99!}OL++F >w|ztn5@*20L?و6s"ԗ<%HG殎-Ѽ3TcO}bX5-iKϫ ˀ 8|C rUb듂U rp0[@T"1OB!,w8,F*3)74A6$߻P@xzY8S=oZGuS%InD2 M@SMdFEU^rl/X [o׸ܔZC/V2%LS-a gx^~* <sDcq&2UR$>u,@uZ}7"Tr>eBg〾VFQMw_|/տ_?ʼn,g eV6.obv/ Ω5^Cڨ^mʥ2<ѷ- gښ_aee=)WIIer#p(v\{@RRS$%t`E^lFcc8ćapm\`Cf^PFR# lЕ=>!>BX0 ]m~^l2:Dʺ/"2 |Z1ai=N1UOɡE!x-Q1iI 7`-荻RT3\ V_ ˎMK'UNI0$.BbryTj,hs1HamQtTcU*T%,37,JYq;nc4AǻMDvٱ[Pw+a!(Ŏ5!Je~qU+ep#j6+-E hđ SG@J$4B N{uln_^䮺7W1ߧd㙭aYm)ڸ7h#07Pn.34CzuegM2l1/naZ6WhЬo/Y&N3H5LXmVGy[m6;mLJjSivF?h(8!,͔8HR:ɬ@Lr! # PҗUZ+Vf1GievZ!ZUǛ<{EdLfC!U\Ю~Y \/:G`M<ũSp誓ٱ5{H޲&ј|I.2 1!4aE2W a6[UT”/nS|}w-l m Z-:@0)[_->T z_~^߿Z o: s~R'?X.46 SeDJ?5 s; ƿVNq+LwXJ*0)Za}`"c7E s[-GsC;PѼG0*fu* Fi.Ǿd\@m!Q±IscGOk" DGSN[ݦ,Olb`t-%СKJ+~Ln5)>^{}- 猲ƨڭ|Ihic֐Ew7r>~@cǣjۗn5scL=Y#xj2M_4pPh~ly6Y"QmO|:wNƌ!' z|yn[JqIu[/Km\òՕxJIt}j|BmyZ$BIU6xT4 KO:$(+G{>7:Ut2tP{hT#FG|i!y$sRSR>nj &4`6p4 |V*|{VGHʓT!!aCpaA PCɓ 9!shƾԨhZLv1Ef':}9*߭z8*8/=BZhpTH^xn?'t|O迖K-'=^N>b;J9Wr3R sFQgI^ _l45~Β{A-w׏߭/Y^\ Nfe fz nNMyDCI/AXxhNЗRE/xA`xh? &=(NYz6N @>zw[F>=?idƠ3DKO#AͳUO}YEm9ËG| nW7cYY˶%k/ѓtY(^7u{fUbtP%/*/ δ1g9?بVlg*!9p5hz9玚$aEd˰>3v+3FsijDޛgyPZ jj`ezX|:z(UXDgypumrRղMc2n$֧ G7WVtL}͜t qFhf!؋kꫯx).RJ:Zg[u2Pqe7;isv^AO"I,Anw?'Y6 DlIPIhO%Д1/ sf͈gg2xfy~IP%jf:}I lf:\A,x^&i(d %|91/y/tܼѲM" eQ1qT(se!t  jO7&8ͩ hfB xzwFr1Q!H{V#!IMA;3|3AV%'P ,zw:{%SɿLiiOoS=՜t' ѹdyisɼHydw^vk|Aitpzq4LD^G/Iӳ9Re0*5ERq\{l^v)bص_߭pB3[ LYM)27 SPOg)T}c~O'9A5TR6ѡǠ6#\Ma1$mT *wSlL8%  =MElQQ,x5EMdna`:ߨpP>!kCŲD0РϿzElL(&5*$B]"ZLtyAj4x"6]E֦(*px5ڼQLSy4mեAg`P5Kl.˛=o^_VmaFX/㲥PWFT"E,a”SbZ%^#HJ*5.Wk̤]v ;1o19z};Gҷ_|o.L?/u_f7c.m/wnuDNׁ 5I;ZPj%;Cpʴ:LZ}*Mfh.pyyXvhI}гec^[V9JNI R*8%C+s.=Sõī}x1J6|.=KM"i>} kUfO;=i-626UBxh909&!ӊ,7 22jZ/ l#&jb<(!w1+4(8- d%S05h YIٛ >n \Q䅒Ȅ`4X t*`/6afaUjnRw8I?ݿt(Al^O$OKSW51%!Wʈ`rNqL*IÁCcEh 9ge&VrX#H@(@`,o[ɂG)- C"iN0|JR1.ⴚ?ۋƅ1/*>~7\61;#o?Ue0/ۉ!,#,\a1,I!!ׁFr :z{$VRm# ߶tvsa[.#M w[\*l6/6.pQ׋?_w 0>8]ݭb}u/߉?/ݏovrnAn;cpb,P d8nvېM0!X۬!*Onl`doggJ1}n;ܱ Dsl 2HȻ)yjBAǻMbռ[vGS[ M4˦M Rk8( jSn-VFQʬr>1Z:NX:efJRPԔafD3;RO8b2Ls3L8~TIOghM19{|ݴsnNn}ۄ,S-ޭ|&eSlgpaAdDƐ Xne\0OC8#` \V Tj(~ҝ1u'In*뤖Ԓb gcKIcσ` M4Ǧ?:X088XU( gI׷E-9cF|F]U$ϴ1c&9;ZӸ)kue9af>@0mVŧq7v#y'#7-׷(~T3FZk!QԒ)wći{ä©;1J(* ϳ^]e !:ͳZꖩn%H|KsG;/cPKA5F=X&礭XN2VBL9K# *+"%:{ &pu/'\3V|!:*|( crTT>bKB4g Q~v^DDZrW1.\ėH'__|lX-Id_꼔p牄F #PhpдJx 'w>W2`'A̺&4cJ9Bj(y}ڱLӘڞ|J :/\-4V:RCcʣ @hHXJOÌmsw>ǤsLYAjFm㶁z=A?;/@X},VXv j(c&&Ï r,݃C ˲nw~@zlRG/E*#b4 o"s ^J O͝`Z7CDKV b k3=.7\W͞X :1lX/—1V!p0!؍d\fb="1%k1O-"RĜ2ro1LRdlFsi1 -d7L)Rnn2nu8 {.p"M9.2jI/dz,#dd(qv7Ɯ%,rЊB3±Sfq:q3囝t´G%>H}>HJӊg"'B|Š-d~V([׬ {~6 '!_}W'y3m؟hY$ &'J3l| ©I;1 %ӨRF{dJ5hRTĘ@+ֺ610,\oh,KJ^Ic*7BM; k}%(pQb V!68ezI~Cv-C S2bgNW]m .')CGEaЌ0 ?C-fYO yF󑪸Vy4ֆz)R'hrٹYRwb7>#HIF; jVf3ZfT˭1ًC|Pr=u:>#C'vT`|å&D3# \jEyFk9IJia '_Q&.v$Ԋ|VsOD5Wtwʳ) 5*FSs g1l jЂ>Zɏ#r;yӸ-»>MYGq#~veܻu w٥oց^SW~r fzq*U?3:٫m|X"F|b&,77_p}d65.nWhhj9?2;V 㕶L5D)Q-TuYM{i| ٺfp}h>l@tl #ёFhiL,X5y25PwV1w ws|om*!xs.ʖ\^Eε+y*8hjNfw[ƨdU&.V]c֣q.~j>p#/>⽸D}t='شy㞍7˓_\,7ggS}vE.9y|=?\1-rԸ iwB-<3#%jSĊhD057J6^ Ó)_(cRӛ ;4W jԇA^3b([/:(@hZF U x/?o]4 ʴ}614ʸv^`z1fk.f})|,*{3 9|~P~#굷3Xqֈ}eg04ESU Gm"K_tGO$ a -_WOb 6؁QIfpM٤2fB2K.M.GDr`DCAoJ9j=yMeCڷ}KYk~M· JIi$!H4Fa48YtG)=xqKSS!ҁIBeRk"42-HJɥ$5u@m:WoHiM܌AںLX֥ Z̊ARpr 2nDaug3O'l^=ڡvgٗBFz}/OMNخS3GiV/5cCt [dNx(Qefwa!҉+Z"gO~iTp4=\㪊7Abb@ yg35ToWÁkaVZ]t}6PݧlOԿVc㱶ˡ'+>}6&<<\u~rrJZG ٻ~[Ob],'~!o@g`>.[T'Ґ\EtJo}Z7|֭թ}Fv,r*g0uiWh݆А\EwtJ:VZ"RF "2 IkK5J*\jԩBq'zGz[zHKo:ױz,t⯝/bW͠sqo2WYRQs@䷋fӉri-6Ěonӧٓ( >qyAqSCév66RzeGhuFx~vf69'bHI4ZJ,~J"zeYEciw7q[zaanrG)z3Rwoe/3^jry_6o6D{}3JWv N _.>^Rgv|ҢZ{r`~r@IObI[Pm"i.I}K1#.–l-KqbY|^SEC1_v]pN{rM\u bGN S;,{>r9jr;nOo;w@$æQyt%W BE}52gH6N}"8jD`'1 6xS:~Hk D̉qAjŜ&(V%')@d \&J&jP!EԁT+;&W 4*t @4$%*Q$g!pC?JRsqŠ$~]B'&jJK CА\EtYݬHŠ(Ӧ&*-KOhqzhNi2u3T9XZ JT]!w̺UԺ !?Ta\rmN_l\p=UrhԌqR3:ѫ(Gz_zXmYY q}NzLwW[' UG97k82VT12Jm)r}rӪ6}&afM}Bvezxo:a%>mwS!ښp hOBc Ru7@ǠHô[8).znAXä0f&lapb1me&祘;i+(VLc7щv#]HwG/= [T4sә"N+(;ƴ͎J:2;V(Uǘab9ZNAXb,kdr႐FCp$CZ"(#5DS2Q !;u7)RB2@T \V__oXw%d=S45q"+}FҲirkvBൠjZ^6 p*Z c:ZaValfѴ.XZV&ۄE…T;\{:"fAr$wOE+`SgZ{KMg꣜RsЄX4xʍ:6 6G+c$gf-~ʢZ [qB yWjL΢I%Ee*ѩf=rtHJ$h5:ŹV9Cn^Y<>ZaevUk9q׋ele>uưjhUt;Z"OTLwZޗ/7f5:%=QAQgtqwڴiFQ;h:ek[ sG 2* z$:*Rc^@|R._rY_=y1)҉QTWUL&} ;3Ō&V|7辷5>-je32͊,%,e׼`9o`O?R<)\x% ?;n-9)"zgU ^d:*tsVIEo9g@% â30{R3eb5묯e߼}ꩃ\Ho ~!d#ѶBy(8qp3ݱG7x?;823ݔR8\k3Ǟ%n4JT˗\޻yzTZnCg vfpg @z~1`&LgƩJWӶ{z*M𐧘@՝K=tS^ZH jC6wTkZ =%@7UY*"L&t$NJZn@%Μ.JHxJ'Q bH-a8;L3Ax J~j <$TZlPNJ=s{J\"1Xb$1W#58e\Q+i0pCqwFUHfkuJR r \o#Ja& U19'А%H4Գ 4s8ʇ= x%z$N(Ipd#{\D˜MxmO'qV3(Q,2$ uE9?p(T&a=)L;('τVD14 )^ $BK͈}gO]UHbPEQjc0I8;Ԥ8ѷ64&5N҂m P,XB-OM"5X+w5 }F4( 2DJ)TQcHcm\qU(P2I*["h4vdy*(5zר j$H0NjOjꉕ@PmQQ;NZoSn0gF<1ͅFP < 0#O$ N;q:(J|\(NEO h>G hLۘ"!B:+9DgL1Y#&ȡEu;ݡ):mY@A,RH z_j8,Z6r/VjO 2cK%U%8\pdh+5+e{M -UYKmj ?זQU|١K/ÞY6x"O:EŻm}˫eUJv Tf}J}|F^lN,J{[=.\K]`6Ɠ훿ɯo.˛3B)?rq1?ٻ6ndWX|9 UzZ&kW<o0F"!i[45$G2I@AwF7a?Qon_T/@e,l׬lRdݝGox\lpEv.kS ;bXk~57|UM<6E^S-íf=oVO}Een=x߂m-{r@4f$%U,Z7H(DG{$T׋_p"$ ԑ;Z&s$` &*)DBEڐܐGrc?՗;T ľ/[yD2V`X( iu,8(=s숗8SY)86bnI'9/5bt!y4*(PČrJ{ Ô*b dW"ϭV!uǏcR%*P*8ul@92[va[G5E0{m5C1R=T H(DpqDz'h|: ɴ|wkvgM.ǟkO{MfNlh=@F៹ lDBJcml@BrnCnyėzl6߂ H7p0j'$b}aELK2D)fEj7r7}|6zб,BQ!QN*s#fyaDVjU`J|[w(v~0pՃzOC+[G3PV Y.˂&$?|2vāQ/79NH42DPlL 2qTs̰}3)5jkp, 'ZT[ղȐr26<@E)|,1!dzڱՆ0T<oCΐSB-,nkfLE0225Wh4$'Ƃ%r@ 9ΐ] ;9Nˁh]Xj?B_ <8y%pBAtK176h%$DqB\8 q<:*So)p6/L r< ݩ0 PIBatYfi-3l4vTzYYq+d$'#4d{ǵNlpy7|4 ++pJGq6BQLYapTda͘.tyVNJca".o:2JHE ځ'c߲}1ٓ7sXCF.&mNʂb9 :f ed1|uxbv"b;rXӥ qTɝ/E;cEPy .kxvi@n~8 zdZN>ߥXj#cr PHѩa6RXJ 2NCQlp܂R$398%9U.a0fOύ3sdXkur&vaшv(6^iڍcWk3?_LEBpYhѽd>,Bz8'(hZ.D1Kj 2^d2D+!2p:_I|1^XsvB3PĻ:zIh#mkKӲJsɨ8c9Yq .VZY]US/xƕO60 !;of5U;}sHZ8y1+hL4BTQJh|H׎ ,HL[,$#iP$_I@weƘvD3pQIcx.Q26Zdceʢf=f5` -;y%X1":^ yqI5D6ҏ!>~dyg**y8.ln0f4Tg+)\.Y>;>W\$0#Z*'@l?/Áq(V%A*˫$.IH3+=VSj||N{؋o]ف#  S,v͗_NF'?UhqH(ˏ(~{߽_L/L? 'gv ayƱL\pIs\3*4}Lt#3BxU "3_=+cKuUdklO n?kK[%CU#\?UKz.<}1<U$6f յ/԰wlJ]oi?H,ov)IP-N~<04`j0v唲ڡrZ7?]e|‘]>^_M=}R5o.B{˾TH%Txn 5J2ϱQӹ2k4c%F3jUFdh+!Oי^1g/H*#Tɵx&e.$#QX R:YN\: [.+@FWQx?&] j-m5"{ʐ 8Яc62F҅t [O(05ynVrcp-)l! 3yN#StF_gc4ƃDPj!6gحvc.}z8M`C 0WFi f'B9$4ʹ異)VT_nSi-p mUS˲ÞP|ߛ}N@,~a Zz\ <Ɠ2kV1lkEshF0FT+d58 Ń#X7PG~ ]QrZ((vQ֙Z-1tPBx@,RMy14n.'2 ׋/cz$~?~1u<>f?q ?of_FEJaYE'G]K%)H`Օ5s~<O*YZBg }'TZ]FΞkIfGgN>Ϸ HeSkU;,Ϧ"4F\s"9g2WpH2:\gI/k$ZΉe[j4b1MM:(ꇽ0̝NAgMy6gςQgkr ס-~aqm |[TK%KW1'naGȆIKd8.Jsf 7 ̈pM2~=ŦAymjIK:suC(!-X]'g[sE6@9HLiPHG0OY#5n{nD9j7Hq}7l&^WK~ޓ <@<7bE~nSG/Toǽje_zzW `6i6l @o7ϫWA?~^s[+H@Sgf Ӯ/0Y`ajOg䴥5-Q&Tycn}"Np& 9eks,^ݢK62q)r"5dqx"g Ge` `z詆a+&IJY)PP/OgLy`5L<gޚ`Ffu[/Y0c &Z ?OJ%cbRFh,V15xPI; IԴڃmo(FT l'nDN3of6$_و^$$ʅS;n8jJ)VC5`[\"Ciq,X{^v>91'I\ LI.j,34PE3ֈ5[Bl|jlɜBsBкwN6fLV́:7}X DR׺B1d֫p]Zc[ܚ\gkqRQLk8ʘZMLܣsVm=E08$QG07SS1g CO|sØR-\9fiJ!Ɖ *h+MS좎Uy{")ovQG"g-2VHpH))ovPģZGy IFfP|ҳixn}FG"4)]6@̥4Nx"m%FU&*Gb/`*ZWkC˱_Jr|Զ?hnEXDbuc*T7p@'abI)Ap~kϳ 6ښ,t?2"bMQ2g3GSrN\c@J5^8:Q/Uai#ƹBU`KۿC/(jBy*@cV|8%k:38qRu\ D|Ik-%!%e} }"FS$CULOTLxDoLPmј;y☳k%iA{q(Wc0etjsPc 0s7*1(SlswsYP. sMm 1aaJR< uЊ9}9%C5Vm۟"V;d>jf1<1G|#Bm_J4[jM.Eӭ%ɸԡyj͔bV"ɼ\gYB2[Ss J3YJ{oE]CT,\3a^zf[E;N"ic1 5+jzE\9؆{5{~݀Ei =fcp uQ{OʢlZ1Lhvj{jnUNܜ!U3#5عhF}z;8/ޝcpNnl`"`v:u5SdrJQ`TsΘO)Pz*Kl q}8M#]`Q@fS[O7P/6P01XO~ڜ˥y_NK]~b"NCjg'1"9DX]R< _Ijx*9szq{L}q1Ruf[{X4"YZi C%n~f;[8\rh">$FJMTLsʫ-iBMHgҒџTb1; Ap\*DQ(j.N%9u5؛|y޾f2y!(vN=CD?.@-`273m6P)8Ȍ(|q)?FE9z,˥K%MW(=s!3Ei\iQS5;X}M¨1-)S05_}q)$B<9 ?l-+̮aY$;l(i[a@.Rk#`~iG핹wQO qFivEB|5."&rju/BJBfu["(|:O?;W21~ {C1uO&X| Ժ-N>"E}ě{b”QYk_He]F-[H3r,Ǣ{$u}$avsyUO5Quu\Z^(9(%v^)+Jk/fB'(vLciy9!>]ZOOOGsx K9Rmd |1o*<.plB|a3hh3m)=O i88opɳc'>Āk mXR4E*w(EՋhy8g{C0w9.JԃOPji_cQsra^$~-|1tyڧ&gՆIwF ~K-8uʣoqW4)NN_g;=~>xSŎiX ( ܙhRZ}@H%h&5"Nrj/:q`0ȀNOX8HK:sl5gN̪|δi>DZ1S:Z_x 7f3gbvN7ujO3T5*dIǭ:L0b n|h5Ja:!ZQlUHk6-XĸJkƨRmS\a_>0z8{64F=F^lY Ӥ⺩!I)}d"(21DːRxU5P [ٖ)=?)Չ)º)Vs3A/a.Nl q6!H94=]B~pa;ӯA78r1r/vʂE|YsD/c|۞i<ݐf>J0~zڅ/L*+FWA{^ato|uV=5rV|nά*{GGhqg.Tfo!ay!>JzF&8Q_nꧫM[:dC^_!՜t۟ۻ_{0Ż*3l)EjН$m>Rjp]i^@{uhg @MSgj㸍_a]T+UW:r@fBQ _8]]3K%ÙFwn<68**K "Mث(4 myjtYɻpZbd Ch([eUs^~regRI%qKy]@5mw52gw?~1&*~TNd?gBT~'J3O7gfcn/.r;%(~Iq ě w}MW'Fʜ_cwߧm_Jx3* 6deR~-h_z}/P)-wWU}*̺u񪹸s\7thRymQ +K*d'q偍WP .a_qr-Nm_f7fuqmVpmZQ:YԢ'Kx&,]6RTgq28xi)]Rzs?।`WX_A`V$Tׯ+B|z᥆- &s\_ZׯR_2'tAPR]\Km TVA,٠xPeғL a0>& `nԆh:^ˌޑQԠFwk4`F4ۘvR0BIDR/s5ҘQqMN䠯Cٍi g UJ؃XҴɘʃ3AIitO^M%]bŘ@ŽW"uS8-Cj  0Y H+=?<дE2#<[T kJd|G`%f>x7R_%$'6mG,P(pM4 6!m$iֱG Z=I5OW>?]?J%e8#~(9FLؙa٭[K/U(CWb5bgS7k;k1FARHcD=;N.NAKXqF߹tR!<*bg 4+dXX)=J-B[(8$0W K@cc!ȉ-%`9|gHe(%$L6GSRʿO2pNX ֬nS c\IƤ4@j+ MV%&ϘX^F(8RCP )i5賃]]#ITJy!1wRw?eӅ CXʷP]݄gO* ywRr8̞QR-}|87n-l%z-f-ʼ'wvBU󩥶L-!%媩sYYC3mWGbXD\r*=ѕCѥYN2޹nC}<@ւ;ZOo:ZB㦣ڎctt(0;htthʻ⌞~tPb,v NrJ>'.lssQ}^`h'h9$0W \:ZLN~FhưZJ¢ _L,;Zjް% *9 srym 46ZSG4H7 Mkx`c * 'H*fJ,QFN?śy1xq+_}%+F8h*8x>ԪU|> CxvA^P 9Yj1y~j[?~m?xBʞQs@͋ ݈dPSZ{TW+YyÞcݔ,!o/>}(pF6ΣY;獽)@}r[ex 撐W1q$kTc!^042u.׵mH*y^Ƨv ~s`<$H:Xd>$%tU#P?@wA꾙F182yet`J314IcSpg@:`i,V6R!%u" :`KMB[PӰja W +YzQPmQ%H@)99f pH7d BTLT9'#5VbT%\qN*`^lݍ*_^Z%̘79Fl é+Sy5y]1db`$CԨh+T>T*\ HZ܏*^CVzÿ",tZۜ4.2iTV5$YmG3KbI̳PjÏg #:UK^jElv[ɱr%@B݇fvu5ݓNcЇCoSL.UJ(e;[nG9WG C}ORܽ .\?O=6kUp~}s۪?1Kgڼ\GB(p2 u^[GSL%T39[jUޏ;ĦǣnxT=|_}]%+5]N]ݩ~zr+cNeT Rѩ3I֑M ޣ]iJGTVP/6 ߦvSic$}MSt+W=N};Т>$#R.ўR>-T?Uy#u"-$JNoW %PlkMlCwϮ޽,͟VݫMçۋfza:(;n6JoCЎW뛓٦m|V*n${h87Wre1gFgCcx2&>|.:ȧHJ(m]=;ct.m݁D%ZcGȒ6|.:ȧ9mj\g7Dl걣pketCqSNj0s| 0!V@`\ph~U@.2K M: T1,| CS݌CS24JjT\GWݴWAo뒾ߕu=Q/ZGfgB;h!11bQ(},ܺAi.=(c6JNؔ|`1?M֊t$_vU |GC'\lp`Q:_!fqw8:l!k>d^ 9U ǂ?7!>D-Tէ/ُ1]ŬR.yZsޞkZ8wn/nT(uwَ\)&/5 A(l0^ykM Q IڳR0ktS SxdjSCThp1xE'h=\CR0:y(#Rv +C95vPS6ECtض G#;}dd,SȀ9DFV7"?SKFڃw$#[˥A:2gPy`ւuddwfD=r&roԻyv7E99 4WUrIr`gCV7ʄG%JSW?-1ݛ܉oc>xvkXUNp?|yo o={f.cǑy2S=<>BAScSUZYEc:ŷq0uu ̟HW⦙{ioeUVK!1f:7d%NlG]i{%x%NG7CU8+ju xn爫[6M6UOZtqI5Gɭ_Q8:Y8b!KTg%#nҥhQZ'DN7Ϗ~ &o>CcUhxyب ij80,=ӯOn)+plJ7r+3Ć2 dh|V_jĬe"/,I-{`O-,:?ڴ@32 A ota"1Jy2&J&1<w!>eN~~XEr(~|.HֲH!߸)<f@Dt.m,€n632!߸T% ܲJ92)\]kp5o8@ 2FBҸw~yʎY?|tlROrW˃ ei?HX]ٟϿ?{#/{XuH/2@^.A"%knߏnMw)%8DzTHVU_m6.=KJ>K3ꗗ?~& Z?caib3)߽?xď_r>Mn>ݷKmST_?P6Z'~Qۃ*;e4QqQA. ERY?;'ri-̏Jآ}niX߼oOo WEW讈W R)>YiiozҊ=) ӌuyo7Hq,nyby5 vf4!<6 YL9DmNd!H'̈́Ѭ͙ 7ng.fm2In?|jwi%7J꿖vmA}_˫*7n~ 'I)"j#ԧ$~:w߼*>a۠a4QO`2_$=a&UPX2X86p 4.p6;);Hrs%Ax&وgfmHKC#`F' 7'dHa] &X`ڃ`B_;jʊF8֊4Q4s"l3UA=ŸGL^R0!B]Z+njM]y!?) kGBHBV:ΛL(UgGDD4* wZ>JDvwu4W7\ɕq3iKb&ʸTeY(&/ ud)uŎ3璥D2 &_y?sIYccG7jp3 N)q(-EXP'wR^p\Tգ"C`tD.z8aBv-)HSXnԢJ0ZN'L`„ ^orׄ܄\%N_k &M\64RAݰW6JiAKMmjV`vRm5;-Y)I5\id;si:4ʹE'36uWkPTh̐l ]'@+#?:߼fY?L ?*z ASeljYȩPF7T6 ڸRV-z_ݏNqx!{o޽k=v5܄+Y%¸ -%F"oƸs;o@M1>.}//߾!e?yWr;~9_v/~+_+4M{ַqi^Я ZٟE sߓ`//_މq=h|Gh޿N! %2A^}nf1nL[Bj쪣R ji4lɫ-[;B(DPL]*\eU!H6s#';FѧpІ88GY}/DqF/o3+PGe N(H'cc)*8r0flw,EO]hԙGs'PZB,уkDeUx'ue $HäqEnuE5}Aa vA9Mm}#C \s:@Yw{')GrFUR/\UhgjW#%$1\ҋ4C/kCJ)v={&HUJs!Aو]1O(1_t8ElAPtucuF(bǎ3EmtIURP#(džjS~B;[aʕ#NԞN_Z CnXo29l:o'}цZ֩|oI :7s}BaT;|hQlϓDO0Sy%|\xfg0l=G蕑#-mUZk~ڵp%kXx,[DOꡗ*ġ~`sҎEOQǨcUYj -pT{bgjܞe-Bi&4ga1&ɳ(M:ȶY,mbrur("frYݠ*ǎG,Ktiii*pr{{\LT0̧mQRӺ( OYC4d:WY}=\ =ϓ#RIMG~#b;:`N}> ;(Luh׺S?Z *:A t~#gu[rmu0M0}VװA t~#gu[2hm|qF6Bs0 &%O;z>EK򆃴N6QrOK}ӗڑNj}4R#6+A4n]Qz(Dlj|ջ:.x|ts'Q |gx wRq<&T\HQkF t.%r Gey2 봎"^zAv2xltC17.Kmo:E@ ,#&JE+Ua48NydO@j$Hl8#qT}NL2Бo\1])^-g`ZDkvT۶d~u΃ΝCgXā[c@n`75]8 OW='S+Jk=+@(@hwe޸;y4R]DEr(=}'Ot.Bˣ{X.'%*cThE^0$+竆 ƊGl%*Xa*Sȵ:``yEEWKaЇkCP/+. ":+ĿxŬ\kK9Qmw%oOI}ӗU\uJQYj==*8 k]bR&ꍈ(gǎ`DX'uA2%*ظVJkK(BZJF\{\yk-ikq+8ve<;-M_ZGzR2A) 3Ȍ卛yL1alZ(Cq,̲r/(;2W\x4683C2bhD:!ͻ`/~nsAX>b?֛>=js`/x@mo^0f VDCyk\۸G7VXwVad0^hA{,aQO`c( ?^SFXͨ/Hj:..\4=7s?N.w{ kn=FlmL&u0XWAjoMΊ6% 00)ι!T+b^[`%oJc$:xi $@L> `y&z qd2LTVUllAT R\eP7Tr-y`HGܨBivd5gH\zI] 25[AKO W96<_p%&ޗ-ҲS wbLR eQ$"SV:YrJ)Q9QLf)Qwx#]XͮvafO\~cٱ38˸4}(q;4aIy総C ItDƎj5q#QP(@EQ:tHfA6[y "y$%]E.-[$}RF4%1ސ(W{ߨƘGi Fk4`bP'NT_Ck^26vRun]܂k) B/]Qz(MK_5@iZNjhoX(`Rb>ͽ5K팽RP]koZ:1JOI}ӓ]Qz(RP^JAt+5Q0KD)BJq; "]Qz(M }M:PV:RNL]E-rڈ+ҺSV:ef knGeMKHg2˦AQ+(ʗ2%eVR$'52[V/#ż\ iY\8 ʪE  OJQWƓ%{mvS>7cd3L%LW hN/0]m36hw8m6Y8w/?^MdJ cCS&)P up$1HTX9%L으)IFTue$#ï \BE) :[kiBC +iXd@+.JG#LT"n BsC!F&fm7m/z\Wz3\ɥ77iK&қdY4 jқY'KN '&A4>gr%Osތݘu+׷68S *U?kTzs\`IX6۸aқK1/]`(n%T9]z򆃥7Xzû"\8z){їi7sa( X[9&t@QqUvUFrjR(ҵhS[}#m5 [i7`\S/{nA~ EU( Ku[PNe~)oR?^Qz(%Rrۂ@0@)4nv+J/NkkQz ~ioR;W^6J YVRCi(JmL\6J1q-Z7=AG(^Qz(M#.7+}9zdpb7A$0|˒v4~ř<6vwXUdUJy7Tc|VRFH)q转}M4H9]2'w4PbMu? Կft|ˊ6祈-cP'[F`6/;0!xM Ҙe`LlK)ኩǞf[8R">Yp+|Ņ75RJ^aD|6s>צ? 7I;rQ;a@<|I7t#T<ĤkIpS/W C%R$Y?0{L/?!w ?~֚L`rf,MLz6GR9ϽPs ^K_G=tTM-͔#2!7 VLZ` 1Z)U-#uAh-vQf2 .\퇔G^%VR~_/nS]Bej-Yg uSn("=lŏr9$~gU _dzIW* =Į0kӿ|n:3;&ɕi5I>1e,@N 3w%v/I4 f9j/]wK 3,ؔI^4]bLӕ@:=lCng O”KC+n=M7F'E 954]{!fO-^M4]ùtK2z+Өъާ],%t:s^G/ -D7 EBdCLu s#( `tR%Kzi}L4"y.҂{WFRLdREBrްsNg)禚F[FF}1CdEyݖ<ڧ 7feMkW1wx.l R&5{*ϖ½t4qZAl4vn PSߞ -x(~̢u9 DˆCXC㪫WwRR A#{o/%}Cveu`RzռвDZ#-[Ow?e\<9OEaPvfg`#Ď5 ˗u?Z bW.\)5|Vhp|E+CԃSJפͱXb bRAJ ^ƃ^X |mGIEp[j4|8O}/Ԧ\<-lo479F2~.GGPi)&@ ɧLRSZb+H#?y !}<|xF;nNQ$62Hכ7X0>OG96F}C&05O̻tNp4V \0 ]{58dCϭZp{:E O.f "` yCRw74Rj{ 3a5`}}5IAKM6X-l28TsYcNb%G鵖e'ڢ|zPd?:5>sRVh~K4St=H|$DMy J7˳.F[cq)! *:>o{hPPZj57a Tsr Z{j[jp9-[gcwYc˧Wn2 ,hXѿ` XX3 uBILI I6bEqfz?)Qpu`J]Ngj6Ngh$0[ BxRHdkG""(A_AIlWq_$Z)traRT|*U0BLpZe}|VQ`3#(5)l\d\L a`p[B*.^mjpVqIF_[F;t ø0 7yUyPqj#Wr^ܺ rް yc-$>9+w]kL/gpV~[y\BG$8H }_F$O^W W2&Z:z>' {|)Ỹ HؤP_$uf]^c+i~jrֺ;'Dtgerw.~~;9ӏw@Քo8w9*^Mv`C`ΘF bW7o׋BNW58ߴXll6.0½W 0;߼"pw5_M9B5뺠A=a54R$Kg g9 W֊6𢊨/\ *ukj1^VxjC4zC:i|\'R\ʄlɹι&gA)'8v)oA0_F;yy|㍨N]M^/]kUKvZ(tc'g^7*9 s]:Exʡ{N!T7FsZ#zB CFNU7N&̓j-%9Jy'y`w]INcBS$_ПFbH3ݲ+hB7lέr9DJ֕}|e)fu֕+MRʚ ާ[b>$JGXwK5tJrN9T;tm/:D})rqz/=-~vě=%_'ﯗm,_AKՁT?LJLh->4q.Ů2GsS \g/98AR6Ҹ_/'v30\Ə-eea~Ͼ.jtݻαy >/hkKbffpbPLaYpΗwW~Eq_mq#-Bz跭>=͋Tެ >}G_-zN;{D{ZMF/#EU#[a?skqW2B#GRH: d|P!@ ܅S +:ZswTBPTZ vRGLlgwZRrG6疓ԝkJP1jRqpz&zvTS›/yI%/DiQŘD~ζ1z ֥H.NiÝ?+DĿ}JmYiT{?-\,7M[Dw{qYr iKP:Oz"FE=ZZҔRT9\zoD;TJcM6.ӻuAtFw-J:m:z,7.6@}ƹ$mOݺ t~Ļ;yh}z ӻa!߸))Uotʬg)3ް2aN)'<B!'Cȕb W.WP(ڀB)4e#On=}>AXYhH712Ǵ !X$E^;o= h=5!b]~@Vj'FT%HpuIgݼG%vvv6_I6ʭ2u0T Iw98 ͽZ-mazk,IT;:IHNK5FEgqYOP")ciTwnyJD9<KH% u/kNMpFj,|[+6|=5yTS3M' Y; 7J(7KQ~Rc7ՈHF7i8up&o3Ѹ}l,lNtR&c(= *F+õJ1I,# k3$Ϟ`ҚLSA65 }5h")|Os_qp<^Jc: ,?[n׿ջ%r 9j,z K otRΞonE1C4ǃm-Z0iMdk57SG-u~0PڿqkȁqvO S2+hǺ&yܬO4S m39mUBqN& ^$wWH~JT,C"?4-r78(e.j@3@gX:R[%5Fȼs-qB԰hZvdژ.[o|K&oǰYT?n2#S&~*9fBu` ⷤ5Zhu:k@9'ZNø‘"Pجռ^&K΢i(pun0)8JG$(#B՘:.#)F4`d5EI2Jܲi/ eˊDe rB7o`:)V a NSkn E1[ka\RI9 jyTw,T3.qg/G0ˋ\fy5>bqX)%˃Y^,}Å@D @ "$4EfvOKb `M#Iw;ׂj/c2X3$PqƲ.5 E99 E@nElc W,}>yEZu?λ B.ʴ ]g{L+heF*J3|T 7Z}#-0Md8a s"3tqL) |C rLmKygG]DA +>B](E:\n%#r@2~VY~]-EÁ_,˧n2z.(J+<8 Aޱ ͎CwGnD2M96a9A'TM*ut+};NdW_nS&@k= ZKw{{iQݹRۚj ) uZBx?hhiŝS(P8~w-qHnc /}^.f0{nL2VmiT{CV쬬dʒжUƃbFRѕ#1q>{e)1<`HDLBC锄C&"ЮKy:ȱn3]A-[y)`uBCjZjy`annc7WPqH7۸IM6nMxƆ ?}Qn6LT{XM0v?qb"jk]{)F,?DT~i!G' .@5O!41uCeFH#|B7(@+Fu׶Z7yxa5Öj@zgqBStG5{I6U̶F`óOojyׯ^wUm3W!Q.727/@ Ndv[w}9o@uEB$W8,13x<9)H{0zq Ŗ$#Z.I& $BD" ,g۷Md A/m׳O< U 3eCU vlyPݲYh{BjS bdL SF%2mG3Q5F|UmTͷml_AΉXC-k5V^i cWp'Q:V+E7:~~ $j?u1!뻨m{Q7\M18dl7+x6{x{5P"?xVb)ecj7Ӹ%(nln[_շߗ0<GvQ}=16\WF{[hɆPSq'ƥ)}Y3 +G* cd`gSψ .ts# xvأFXP|FY"l%՝pst#~rfgn&&JWTraM Yல뵴6f T7 Y/ib 6DN7(qc<ծ/1eJKͦ.oMKOO.d ]%QN)cO`ktf3G9f^ܚǺ ]8n}>, m GSh)\nSs,P }v[pLE_BnN\DՓfOLwA/^qFҍBJs*]qU5*%]Û& elD;r&Mb&1Pd!4䙫hNa_l =uS/ EuJߑcfL[֭2Ѻ !\Et5pu [)RT9m,B.v+oIInCh3WюN% GpiG@R(({-HVvpj{g~Tw]",6H[އ0U6EB .rBT dG"brl^PS(,a^pnӅrB$+! :2٧ _{&IJJ󠣵KQ0[xc5).* S^$R4!-0I@GJT/Z(]PahNV_PS@4g, IDy/%K2!qˁ|~$J>_ P su88Oȑz ybC6Š0*9!O'WnjԩT|J.}K\+]#x8Nla^ 8z$K5 T $ EKXdz)3YB/RFjͭ^;Xi8<`Hό!N+)F/$׵j琮7Eh5!QI5 *\M`*LQnh^YzD d+-29E+^i~]E\K4WZ1؋ΓC(h-UNjSukhg_pM#H:pܙ`㼨9bbQ!PN8(gb";LLpxсrdK`i+ï$E8+<ZT<(H_N(%8"Vsb9!-K-*tA[Ot(` r}cw!$ꧫ-:u)mۿ 46ᶎUW_6mk\=].!pC?hTA7+{_DlPG1*_ʉL z@2 uḟ͚/6cjpVyW!2?yMhlPxcSlϋNxQ쒱M&&{Q?qؘ?/'8+;tc,$ܙ1)X(^9H0A!ijIa^cm  klݸZ_V>n6&R5K ˵t-i uLj%=sfJm&2W\*ٚ?fc?[P|n/wY. nӻ>4ӸHz8[fK 2y9b͒\W ?m]-E;-l-V,-=GujZ5L/ZzZJyҭ3R5q;"T_w(_7rE/[w{Ґrܗm]RnKRg؎-籩G$$G[m6HX3sFؚy/3SodyHt?/fA5!Sg?/&=7 kzJ `aa@N#( ")jJ0^EUjֽfRiR(KTXZ uڢL]%͎P/Qs)kʼR$~'*DO&N[K9R^GiiEs➶i)TkaK󴴢 =ik)<-EvO(eVTkdԛR*%9 -qj}5vZ3wOlD}1͝[=י jXzHiۚҐxזZ#vN9B^z%1!ʈç4$Z+)s^?<317<ٿt>,+n8\bh7ı B}xǫ 9k:iԛ(2=ޜeJ6(9 >a= N0=jU4ߏxڮ7'M,Ё}H^*E8ij<8qW(O~^%Jy󪮟6n %' D.nL@}p\ $JE:-\ 2ΫRo|p~j _Noj/kQ ۢ2iVs4֋O}痷Vh~oRWq96+'5+W?Gi7ōS3戃2(L(% Ev(QOQ"G -t!vY6*B;s]081B2~(2BΑi,5aY^<4kruS31;5k"5+4.RM͚KjV#YY.&Yl%/Sajzh:틦|-ko!.=nڡ>R:Ϯ䎣OdNELF c?7H, j }fu'QJ&{N2!078RG3HuE9vG[ 4/f匰55Kg6KU統[P;IOCIDF0=/iIKJKtƫ!K2'S*RQ-ԇxrHgɌlECZY:9Z J 'Qt!'UX1q~:!E/|*h8&dYhF|cS|ĉb@]YX=!:($ ፈOJÍ*`JdPDҌY(oi*D3;Hd2QLbhn2p8<'x3r%wts[8G. >w*3"'\"@4icԀ`%`0P.U&_՛k(7gm8,cDE%Z.hwL+/߭,@$:\b,P(_BNMb.sq0.RzMh.i_/aXSׯni5䷑(;^%Our,L8#0%coPї{؝o\3ްh{*d  S0.㸲-INQJ(HI vѐ3!G%Gdkw`9{37T$yb a׆' \Z%'!5Bñ4(0UFDu&D:,[8K,;cN*e$-ʨRq <!(-،ju52<4W.mOz5 P#hc évΕPBɓ`E66rR]<]<HXMZ_ڔ\<=-c_U|o_.z U%X~>uUSrC5_CߧSJ绸1K՘?o"B"_=^_Oo>nd1e9~LL\$WaP0hx&qL"z_?ıĶ$)Rm2-EB0ż.ػuP X$13 +w"[bT}PͧP\>K4hX9$`IϳxF*E~fHabO쯠>C@rLfv 3 [M;E c^ \_>FF5鮽Y 8k_ o/߽{XzFjl и7 Oua|0dJ|#d-Zg=%-eD' YjW4JK7O1imMSLO.dbH*`[s;eR\'`Ds .I*PS&Աm0xOR'  983p]X[6aw yH-t}'y G622JCQ6W0vN= `M.侹~hO5mɇ M DnJ[[ȏeܼ]m Zb~/^yvje|י]\__GO_t}se߆W1$zH2U(a2ZwxuN-=yu=xti (h=i\E`{TS[WNw4nfL̺ueZ>4tG֍B֭+rT;X"L̺ugZ>4:z.5AgKrFZZ(N*Gr"ZSz:JyEӛYHv}~^j۟ڻt2jߦ?#hqv8{qE}wi6|/2tOݍV&P>>$_[.LZ3{eyt7էX2̈́Q{[G^uxxͭ\2+nChUT70~2;,5JAe`6/)aVCtu*UCObX<, ]kX6E+eI$)v{i&nkPjX:: 8|DjF")k"6Z#œGh(rʍal"ݬ(9S;j˛7k׵iyEWlhQg*͞K؜eN6+Ɏ2ᙓt3ټ@ksR> si}tߜTo!}K:lT)zu֗1yӏcfқd^UϫWI?Oo TYX [p!4X:4J,-3NH*y4ߦ|SI_UtgEť 1Z;OSX?p/o4XBMslVNa\\F@&abACG-`"#U{@<2,KFKm}aJIw6= #|" pH5!A[ar +M;(2RT6ӄTR^_2jc<1рhAf(AE H/lPadsnTWLÉVFKGb\v$q4e4-z͘dbjGroGț?4D.. pUޖ5ˉ15\>t{&fCϟ^dE&mùs2zɇn&899bi9و9q߄}҈)*g˽γ496ܩq9d~'H Bk6X˽]LFF@ϸk&AK#2q4 DXa+."y 7g\\S\\C&z3.o\dj4ƿ Fp7E+hc x>TL>N1jjI0nzLղv55O(EQ+>>N CPcPHO0z;t=uN׈JP%n>DeF=fhZ76A1܈܊.ޱkǴNB2r]2aRIs?8۳7Δ*-;፶䤐L3]0Ѧ>#It"X+-:/C mTZV"JDWH,wCFy^f1ά 1p)FhI;&rGr'Q`_S`OͬC jfQNQ k9C0oo]h"\?ƿ=[`E^\ DL;}ތCڳZWjd^_g\U߸f̴1Oq aJ *,;c- ⯪0.}[cN;+u̲ TPSkϊT"w15v30+?iNG΃.2{X;+JGak#GEfcŀ_p Ay@`ɱbHLƁ{s>}."}evc*V} yba5+4GfJm-(ñlAM07IOPD'Ht?ԤFCL* ^zW&Ȑy.1t`lxFDZB2ݯ(KAsHoр.fSKyc0=RlȻ>1zû[憢Ny8N>rt@t4)P`]ipGrc8-"WL)&Od昘Gd31(4)M3U,Qhb,meTƙ ([mj 1O2N<=1|FFi~s|FpZ}3M BJ;gK>B8YO~͠4<vF"1T)>K3vm!蔑fЖ{ga!DKl ([4[ RL9msWЩyD6r-)A@,Ĺ Xw+A~#ǻEāyD6rTb! 8`!NrZNr${)VB35^W|ryG+on4*WAW JOľZ԰C8:4<[)-0RYWn(̥SûsU҆?P%y ܧxfn$U-|񱖎'yk8pKlz%WZ|_n2!5+ԲR+5 i#x<*Vcw"$ׄg\EEcP')@sVV] 8‏_6)a) ȯx Hԭ8PoD5&WbAz:Zk;H?AOz9g-ȁ6UPhpm}WQڼחeU6rG }0͋h4/i^Ms{{y]Q*m)\0s[M+%t 8[+o2F 5"7[=cZ7v~޹ً'׽l?ۚARNKX_{5HB Q mm, 51I6w!Wq/z$=U_,2g)=xҵ'F(FҚ W#ДH_ʀ$p,< fc$ V'cR<s-N, /`5KkF*Q*hSF$^3I6v$* QrZkICD#lӀRP"FT'{!>#J!4y_e?o?#QFc+ 'd>JB|4H y pΡ{+.?$<6=R$8'ntC"jH^ldR@YW kxͭcoJiA)}84WRC)C Ԁhss#ym}t;ςPq: %E7pZ^@HU+W[*-Pq1i*S^$6hԩO%)7^4uIj:ZC|$FΖYЀ1ڛӌj% A{*[zB`d)hUmшI HVT0ɼ)r+h&./]^DvyѶ6` hjuxԍiW )50SsbhQ'~렉LW)b.-?Q4 %Jp:Cǩ43sx ekxkpǻD 8uz-OݓhK^ BM)y"[*]{RD)ŠlYH8E||u ujAe#'4,Qlpz!r A%jaM3T iX#s @WÑՔL鐷O#]<'t~ c8SP$Sӛk8x@itR}#MKooX:x7"pER8"cJ $ ƫIoxQ.9t@bOh%ZgS(zI}* JD+8 QzE\.Pa =%þPIO-p:%}Z ;[88#y߽zLpB?g kՁE(/,z./% -!E8lt[9 X%T:K0t2A<6ueY! i "WJjV 1fe!-vHz8feCP5+UX%sIK6& WL%\i ϒ'DL@)>>ng__{(宄9MtV 9M/Ug_n޽Hqz, +#^Ck T@USuµj|SקDRpJ%3 VZut I*4yAX!<:@m$&8eHr`.kLϊ_1ݭOA~ c <֠_־T͵ξ}&#ow|H_/6J<";zO7߽ R\?^ImPpb4\{{s:6.& .U}@ZNQ!Af!AӸ}8Z";eJ(V8Sq2 s[b}A3S Ll;5@~̩f#<a\}g#|P6"o7^4h!0Ekha51 TjvՖ.PJ ҂@-kmxƧz_5BrphdZL3!Xh#}L ZRMFp)8gLQpێ mN%tx<.>dK Ls0DiTD;?UX,%EJ_`иKWl!QZP=Mp;o:{2¨]Dds(r$Y+|vp.t穇CLGBw>mOdj jUSmԶyrhkӓ04ei ƘI;*Y-WbBHZ1sYH 86+w9l} ;b|ae?χtͿޚO/Wr`w~HaCqBO?>:kO&@\%L 9गwRfI|*eVH[?wDPdTYt-jj'1y#Y&;[7F#3!T'.Y 'NcjIZo9lX , ͜ *y{NTuD""ƖZ!H;7gn^D-$ EbnvˑWJB= ~Ěwʚ/!VJ$-D?{ȭŞl+$xO$9{ưdo%{Za[72bX"h`=RaGuJX8Dp%;E9cу;m1C>LHRjF-YGفּEJն]HQ4FftilFTu3@C^l~^T"הif)G,yYe>BLIR,#n\ XN:XM,A¡/ܠ.U|:eCwlDhn "ȕ~iL=\ϾnyD1y|&+ ck GҏԣШl.3<&^"6ΩK՜]s%L)M.ji#$iQ##9{))dHR;u魂uY;僖\vPD^ؓpMɚ+fn}Ss\*#Nm[ Tk\WՊ1 qJUBK)U֯]N> ^ʑfŠ3z|Orl9x9hvu`ǿnul/DuZe2T֎hAoҊ)lWA0:P: /2Zu-ڠ84C]Jm2ǝfJ+00 7L9Ekv-wrK6# H E0%׆1ρU-* " kR_JmeVin}udeR xVjJJe/W}ZsɗwU_JAXR_<ڞ#Rn묔2Æ}ZkR_Jml|cRuV bn DROR\];aXR4xңRTm=+J[V9YYDx9?ƚ?}T2"j&;Ԍ{~ .Cj f=\wv\/xr{gh?fOaIZ?A(wF=,b<'_aY`Uc-)96jFknּ<6;Śmh0;9h1mOcؠ#q O-W^Fd A,OyRt9Šbt >s6X!2ʹevSEa1Po+bR !P#z{E"+Kj\Gjy覴 U;J%Ŗ+ʮ)QQo֭U.Q;hs%VSoԛ:׸T5zkD9jZFtRBAgȥv7{7sCʮi.h]=-PoD3gEެXW \j/:4\jsuԛ'H1︨1ɞHMY``SG[I}*5LQ-uVvNah무ڂ=Jm6niWC޼.3=Һ-K2AX)WuVJmY'xcR`uV -%܀:+Vgf@JJrGl+ĚX"N9.8n+J8 _*mR.dGgաvSUr'`_Btli/VJ17=`/q@**%Ro!#Rovzd:Z˃v(sbZr!)Uko +_X WVzgc=,+-,c̋T lhd2H˰hּsԁGmv5ۀ:03;ޜh1R;7nzsF24?8P;ņW ̧$;IejMPohvze傜{Jmp Z$qn`-zSGrNNRU\RbxS2QP )ERc:Xar& ѣQV-sDߗ;WyZXr| lEI|5K?k=>O2(;Px>⫻~Mٖ6L_`ϟ:䫻bHߐVk%_Onοcq*@Y:\7 z+S!HU~<9UIsm==ti#,Rky%`PL(U S aUMe]tJ!'[3d Azu!M}W(gafw =Du!qzs T=ŭ36Wb5nzs ]UZ\1tFSf˫v:Yz0)G[Jgq}l ꜕ob95Zq;ca#?Airv.6nR {Ad,rlˢAybMrW^g<%uW|~G>⻟bUj:nl1LܤT"3 $Oef)ː5dɁ&Q Q=+U JM zs *k-RYكRuVJm8Yq[i%m2LҪK\WvS#V$h6WFMɬ<¼hQ i9kN MN,﯈JSJR8h>?\΁E%x!2ƄcpۛϘe86Yjl4D6ɚ5,qal63vA4FQ]8il6E)ᕝSF/6=Fʢ:$WzU|Z_&?|H#dEwP)rO?)m4__|dz]NjkNMRw"QY ZHzw]^'VD zYR+*(L7wI,"$q(OY H ^YHD0 SJFWȏԔ'չjGXD:jzJ{2Q"yۼsZv-d5J@U.(Cnrݛ\Հy n\bbb54AtRQ gi)wwTM5,!֕iX;2ܱ5Ox(*dTz"M(1d0\[ S#$ɜ(Cѣ2z~\!I$5ԕlۑԌO^a]]y]v"g kgzr{./?{n&D]f'o7GͿ-٣SX={ aH׳9FlC<X9؎EC~2jdƓkX{StTJ-͟x+" iF@ |Sᔕ-됶E)SG߼ZM;/VH_⇁)px%{\-f߀ hrSN4ܘc ,KsF=^hSYMN F];Fc$v"hl CݡR+;@@^s+ĭ+' ޼%}&;k ;>6>RћϏyYeˮ}.>4IhcT<Cjр`ȣ)n3JG\z=vޱĪq)1 }ѻɏ0{+۷}wNfK[U I&3XA/j|VLgXQ)0ө>934WY(yDF왌֕Pe/ٓ,}&ޑ/kƪ /B`q;}sVvxesz]OuDa&^/ &ZίƭJWʼROFw{W2Hͳ*M9p9?LxVBK,eS(<كcgJ$<i G!G_G"+$ظ"5Tqc!H@wG^.?5zv{oWSsY93Z0ɞ1kz}͙e0&iJx0=t@XZUֿySbĺIcq%!w j],3,N7xuJ戶zkӲ,,is0T1;۵' YۈT*ݏ<4羍A??Oh;C)UabzKqr^#ЁmC}TF԰9}PGaOFx_#L+v@M ׋ ;.oMbt0h t9 x'Лls1VgޞZMhiK6咍;! Ѣfste'>}9s(|)oW.ZO<4PV-?~ԶPs/׏ߟA+>P sWwꧫ=?L~?ꌜl1 ?0~SdOܜ\٩FVXIϾY8Wػ8n$W:211-} >f^!CqL/"B]$(Df"3p9B`N!9tWl/}tޑNzyŏߋΡ Gow77ON{E ԙ=JFߑ6Uŕ Fy 2AX :yP@TMDF5RcGxm,4 =; lydyy&BuaVg-~4dt9G0fY+I(We`9prq9 _~=56 pmfmqu+~=bzsrlF{%XMof])ʺk>s) +[5?cˈEzb`E;K_Kw*XC7e֊usn5QnlG4H!/EKx;nD6t+ u ߑ%&3fM5eITp-F\- [))S6)"llDp) y*n2"((>Dbl%6PĨSWǾi /VhE!KdED_h59I/2zme8}(o&yiFX `YZe_)2nݛбhd%Z2KZj$"No ]G+RDB邬WJ?'τ\جt%H1L)rn͟^T FA"ri#qF^bKSAYϠ Ka%@} y4\2CJ\CU[|bCN^1P@IaszGx0k΍Si 2 ?Y1mݦ{r#ki :/-CL>W*19\t zY,2(x~{5,S;XI$N ?,m^ A4Ϋ}UٟDw[&FE|ٴlvg/SNIfz{@n4&FEOdф壦mUעT@|*@F8j+tNY@{3Qs[i4Sz%bBڵQ8_GY4nӸN:mnf9Ԛ'{kaM%%1kg(㠴WܛThϹ,/*|;+&1 B 53빮EV %juH\pgI;f`Q0Z% *W ST$Xer: ƣ4.S4@ -b4KHjeX àXu jU` :̴YPF(CQnH)=Fh^aÃ0<jĺiLfQ28VQN6"\2ʽwuŭeAl:_#'z#Gϋv%G{2dI(|uЉpL%rah][Il,꘨b,>g^$+60RSuwx']Ncߕ])6}zao7_q V$!tw?(ԤʭfcaK– CO?~u$J8n)y8^1$Rq׋oLPa`W7lTxA%K:[b[H'-^\Ž) kEّ4c .}-Gq\M XXmlW9p3n{'r0Tsh2|TnɠĄUW89aU傥ˀ]s%*$LP=p%JMj r*] 0yDPP :Yd@D%X8Hĵx )<ؠ뀠SOv6ﺸEߜl9ŦoԜ\CKG+mJj߭'aj3Q<zUʃfUeJjfz*<(ZR `$%4ƠFہ=nI(q*A[gPA‚D9bTnI˕:Ptň€R4RJ5‘TlUeUU0tĚx4ɮÝ |jCx臋g:6wn5¯_oÙmk0\tvł^gv;2! #2a%S#$^A7J8S0ka<LodX'HAƴ Z6xQBm+@$ u~V[nu-D2U-$vdWQ7t4rQs_`,ExY3gl;Csu|Uc|POz6Z'g%dLg! -YcZJ:κ8G9yu#~~~(YG6_]~Xl- GMJպś_.oY7׵ڄnzƼ{s1כ :n/of#"M5[;!ئY}?$5NӬנaJ\Go{4H9_M%3/I (KK]Ab]>\fg_1"n]=9WYԡQ]e} ;ɫVWWWl$^m}Ɖ[kcb:H8H 'y^6_6bd>侶ĺJ-HK*ﵖ@w̜jeHE*Qx^ym1 4յ3h'L 2'3PG)xKJ 3ъ?)G-0=f-%aX\1a`(R:#,W|l"0NFȪ&v4K1ll"ٽ}qv=v}.\]i G ;W"89$RX@"w mnI5Q9l+=54Ox!Vȹ8 ǡ/k }p g"b$M*>VA 딾#GMhHFn[YXB al@(V<m8clY`+w>ɇY8Ggra?.E-~wٻ7n%Wy$-EHY{ l8yY6ٴXsYIpjiHe%CIxF-vX,U_ôF5VJG^~*busʳ^6X[W;bևg䘄s%zVLFs{;Z_ÿ'Y~㾾h#d]їn%-i[EO5MO("0/Ό&uGZ=ja">tA ''ku,+'mT2FJ j ֔+smޯl 8Ząon|u~cƀ2 |afHsTE.]ᙐ+ita5LiK%&?w'OlnjyE1Ys@>_N&1k/zò:$}.%f60@.uYMeH ehre4xy{~:ۼzWv㽷v+֎^NC~i#Q.gDxqX9m 3آ{aK, dDjmT` fܷvcV䰏aORcppn\kNuQ4[RR`-bX M=Gnl@A\0\9҄] +0J5BkFFL2 RzG(z ,Ƣ\Z[*ՌQ8~~lT+IN~4+-@|j-$9YQ[iV*DHAX>7V\蓕$iŏύ v`aNo^=vzy晹dv6{ȧY|q2/&_~MjVN]hSr?My0ssŗy%7Vk#Wk2܍^b>2E`LlQ v&p'+V%RmL)P%PR&M!TҜ-*l RzgbZ>tXu߿Rp^U NT萋|'*jt0M%]'ȃYOxuz+9C9.gJDG% Xh~j E+ߕ`\̒  c{n4NxbH b |Y1K &$B+vYP5zk"fۻYp_.~3%MkMϞU~DFH{am=Bg&s7c1}6\t~ i/Ŧ77RVh7`'3"ȣmZbzƓ$|oNe~^I&3px]X5?&JMs5}R~PUmEO8騙8e6n;!qQFVI?oBtӑnA3L1wsA\te~ZZ6/$O=MfrUWxs7ZDfYc2"dwYr]OTM:[U_V NP!cEAQ1'+L+G]H̅-z,@!1l"G TF)Ryn r Y!BE`?[0@tPCYarZ,>*(\݆G`F Ig/Fr:/5nQ; C"Ջ9|,3&@_OGQIP>O!gagQ u+2m( CTph)ݭc`P7h|lD0wICTSrb+%8fS㎝O"o+%Q0x!~Z^?o.rP9zr0BL*^̤%Lé`R%Ih%y 'TL`Z$ ')a|QKRbV*:w"'?fCFxҚ6^NY+}ÏR<8ƨA$Tp, a#NÍxK% iKysRd)EaK(d?R[S*y B~P b] X.քӓ&jWYi!8#8eƒ5˃yrB!z*U[ǵAAݠ%.,,;Mൟ1-+ ') \T'^E(!։9K5ah4+4Z)e tTkFO0J*R4+]PULdheW`*\~"[Ng6ξO]8zڂ) @87*ÌVHw?>~ŔAú+Vd&!,^*O^;[^VXf3㔕$E)tEBlcU#(ek)N,Ms6\U/uzmw$49IǴ-Bԛ儳,/3 P Axc+wVQexpy&7ˇmpIFyOTGG%MutJN%͞j*"Ŏt v6N6 6unsx5ܡ= u.7ҍM_ <&Cv' '41過 *I}L#|vQ,obZ]G] `V~q|kOXܮm|屑lrinޟOdպpׅ.\uᢺ.,˛LCsý9WZfr/r"FRΥ72TP./ޝ=-lq恑HPUڍXJ_f1n2! ?r}nXJk 4ZHYǒԤE*VptT,%\S4֢Sz 0UA՗#vUO(l#ѰmE/FΔZ,+SHk`>ah GgL & #zHt"HTP\(Bla,!P/=%YqkDmDLj!wRzE@<5(^ o%^ZDKzxsQBg ˧{v2(5xةzW|SŃi&ZCP'M֗ @;u~JP^85,|j;SQY[)WszI1u[s׆99a2Dw[EWvj0<nLD&k*Pmn!d'XeZ-2=hbL.0$WQi DR4ZƭŹEŘqA:u˼/(d-Y qSFx\yV+Y1ktlSEA \s!&O/pօNé'p%]k) Ez,5T./[n;O%D I7?kt8TK\Pg}A "X&p ]7h32} `sS0Oa2Ž4V|<*tA&yGTHl %C 'V{.Sc4ڡ|NF-/ܶZKIR(I`9} rmpaZX~GWanzHu%^y '#ĹyS|jUm{cRӬT腾>+GՌ9MҠ+E+x+~7֪m;Y1Z)4+BLa ӬtA5j8TJ9Y)_S!J9YiIudhӬTĶ7S<+f 5©A1[)$-4Zh }r~ &V ɻl0Q(tJB&R;PJ')TTJS]? Zӿ>@ltm^&тiI]bHzKM AA0HͅK^T0 eFS3"J,cTg-YəLy].) z=騑"^iPOiGH)Ivņ&LBkto+3fSl2d*:>XVPpsrF,i% g\.= T,Hύ(?Q;$LK4:" UbΉ [Q ߳mHz bN^Ĩ~OM ]n[cD̹\>=rwko,`~jt Qs8d DCGWȪg˳KLqwR%X|Ջ#p˳_@)J<OQP4? !|)wa0(|0 T\g3^AE fa鯞l)YL%mݎź"@Eӓm,vxX#-a(3҈DVĘ { zN GE\S_^s!&WO/\j/ԪNjitju@\w14+A a5!> )?]a5Pp$WZ+BV嬡zʀYgUܳBxssïd7H)aD,of ז+AÚ8/wJc[B_u!́9"w\Bs~^ކSU? h<W<1iI-iSíY*5rs,ieur 0'28l Al:O Uht} ᒃν~ļˤy,*C[6GҬͦIrCkWtLGH)%OtM#pqB>F1RVU J 5z[ihY +I)SG?xąLn2:ʅ 2i"t,dw%xIي;Djٻ6r$ri ,>*@ܭ8v~"`1,ަ"E,ޤnû7ɤlngO4ȟS~GJ}P6.U}};@ [ x?4-!`Q@Wlm ׋%&ib մu![)E~.Z~g/4Bp6=<b|;{MZ X =6'mӊɷOXYK.~EAXzcϋOƛAٍ}n厽LuE꟏pwwUݤww]V5u޺̐DX )m]k4`;.YzX }xLc! 6%?4UօZ!K`aRG w9ω]Hk h̷#.8*.a6Bյ TuN(Eb&6$ۍ7UUP>|)%6T!"[=vTlLDv'>l>պ1#g kԇEB=&p 箄R9w%3ލX-擁z SD#OcX^Û?ص^}B0iyǝŘ+{0*TuXeT6fՎw~]im)y]`w)q*"ςJ}B}QN?:2Ιݚ?wP[2jt=eu3nc!7ՎwoO޷'opwW7A=z Γ kY \G*h>g~ٲ+bٯcl;!Ô,rHhޖL_;ga0uAL Ѡ%[4P)@S"G|\ ֚Gp9`};`Ovc@* A !4:i$ ` 2IuSANa7 Ƥ(ÒTdgfp,5.5[gF9wq/ EwB;F X"װN'$geJ|&E^(E 7mXfbkupRHAiKV8~kM}&Jyk D9bi(q87H Q CQXlZR9cJL 9jO pi`TD%D7]쨓郅iEX(B':1M;nrYQLGdS. n_9pkNEXܩhSQY]u**zSђQ:uҨf);MXt*$BR ].$e)THZå$)dӧA9Bژ?F?$Dv5!̩YȊO5so6˗Q0>wI>&R,_cp%`zcr.:9=U6w5#}%;f.mR(>>3<$4@YOшwۗ_LYjlfcȪ/{v݁!j6`]Hpn6c7\P221 zeZ7lm6 *c_Dك_7޷톷 ;оre9ɂbWoWL@y2kjcg-\9Q.Q)A3Xe(!"'A{G(f ?˞b{U}CO<ǃw}KO~E[8.>_>=1q~ˠ<vrɵI+ReZ}c& Aݮ] ~'5iW s)<3)u܅(2d/±Ex'i cs p&zv HM505HQ7(u&SQ_TuAx%iZ gOۋ[,2JʚfýK?kwYͻ27nK/+2MrR6t8t*`Z!,~ެW 󴽱gi=ǰhryw6MtMϽ86T%3]s\ yUm j+8֌wK_w*.퐻}{um MUAI\e/?Ka.(sJQUm[Δ-SAᒖ"'[,bI!YC gI9q,?O"<Ӳ$=M"4lL3N4cnŬP±@+jI1>X^%W,nHTkYAE%[ߊ#:kkaTo2_ xAZhUOE%G㞾(E8LV>x!1bDЉV4LE0 `YW5 :D,g  K5]YBB nˡ2V ]J**\Wl*(y]]jb'MqQ/I5?d?]ϺģC C /$ϙPq>gKlORmgFȌXyvrP7k84ldyQiFS0?7g2Nd/ =Cm0 N|함VyI6!r%茮w6x*24Mi=XvB9jO 8`7ha,vEj8#Dsy[kY>PbR!q G4oH鴥vjސddm k R6`54D|giJ밍ܴ#,"Ε#(YњcB(a/ӫ2׸-z>vI4j 7{79bP‘s MD3nB$k}`0q8-gEK>OӇ/v@Vcw Br5r} JP]XYLXwB9^-m)SȼwT;~gTI ޠwReVPMCg+=I+-|\A(;$Ԛ|a!H= zi5L+6&*74z7j2>%TT ֎[uk JAb"GaHeVPyV ?Ys-ԱB&B'nES:6GaE0VzVʩJyQX)2+m"/RM$vV՗T촭T2+RsyKT_nRMp'mpd$UιP{ZQ+:"t8I=BhzU@3^!Bh%BQ!ԕp7;/gFz[Õỏ+R3絬 JW!s"CeB W+VlK^קa J(MT!tQ/j￟ n$F' L^}$xZb%7+/,]FUt1ȑAGwpT!Xyو9)F))F|$k2 `ړ#ɽ˨5D먌HLp8R &**i:z87 =mXF >YnY,+,nR۸,0H#'\-kQvxV)MPxFqreu@}BB' 0C@h#d01XA&+e$`hRtu0[=7 87y44hiӺFn^`˩;ɍMdKmJGuT.9- pLk%0ѷ:*x$wIg]/'QאַVyӪsbkVQy=vHU3T:D hcl3`®% 8n"xR Ũfy 3lDhc)2@${*ᡊ䡚F=rd㕂 nZ#_KìPeQX1sFyUHE3(Wd([=F[SedBEiZ~Ym.Xx7V͓8| F๒\!QHhJSIofLL*lȮz#w?z@x{U-Fx&=;(A5('H`3$RzB>2I2{zRd4ljn^}¡..A?)D'YhAeLX4 ! JaNR("gb'›CF< )TNCն tС"?&l^/ꃸR1x0 淿?7?b\0$+-lC~#@u_3wG7_,Wq|>;_|p z!}{r/  K!L/Wo'MA!hrO]abĈ=>ĿC1})( ×/#^1xz ]a~}1귳,bO8gW<,zDa 3 wJ@L)1Ռ)Ǭ5mjbcq(Vĥ$}Ѿ(A5ڞ0B G}%}',)BeΒ*B.)^6!9KeA E$T8O$T@c<ρ%Hoo  eRzJS⼡+/eV)0RHXQ6u&0D:j7x1߲xfyw!̘`DXA;p2k-3kIIl7qD 6r*Cf$HMP52U󩨖B6~uRPہǸ"Hjy.pTtUyL;U2J0>3E (a 3őf`96R`yY˴ZqJcKܱil]Y7lRG!r2=I"u4BHe4)֥[ Ӷ !j;pAc?6#aS$σ%' PB;s\cP,֞B,a[ mʚ$p8Z"R6#CYX3I<è;lhcDM 5 qJe/mGlXE)`aƢ,6Rov _t@uu<__.U@Ovy}Z֯3ʯ]ٓpaji>9:fPa,/&x0_?"BQ gQ,90cw@A+؀e lUvfyg*j+hDa"hxʻ-gA "hxP2P틜Ei:f;e[WvE= v⇙5q> ǽY̧odP-up(#-783"B^ZI%DxT .;;=;͵xl`YY-G0B"Q CuVq;|M~o/ -!kitROCJeLb66!ilTNb=lBxi w&WKU4Kkd{@˃#ź {nviN"[U4G qX7wк FurHn"L{n5E"[U`G<%IvwHg"ʸlJ%Pۚܫm?#z#s{ Rvj5i]u{wfu0WW՗Ck$tW^;^9o6C ,7<% .)0o*9n:Ϣ"H麉]U{MK! ql6?oJjN,no/A&+esOI'\I,Lc;_V6@jY}sڒ1s}+|??o0~7q\|@/4qU1^s?9YۻH~_Fw3JϾ|Ci6٥Ӫ3B ؞OU?=<خO_\]'4\)Px5{wMEnka|'h͠0 6&S+ Hs^%۬2B&B Riv %d"爮B,ҢDL,.׌^`)3,EKK5D^'*ʹUWtt)oBDBsaP-8 NڛiX;1-bLGv;1/ukipk5zgL>^4UWӂۅ{gM`!.ɜYA*\5?9Y2I| K);,vB%nXM8A[cD[Z ^P[GsLrcR%eᵪl^ܕ^35Q]^X=߅h*RDzWoRo*,٬(F1~lPW}N3sZ8"=i 5YS}`Gsto zpoXm"ȋr:*KW3~'.TERIʘRgW-M5C׺(7Ŷmsy\$=x4$y{\&߸ظ5JL64?;mddH2: ÉjOJ<"Z<' 7l8FF؞ąVmA_\2~PH,43mT[ {mC k"Jpz) rۃ"\bMޛāXH.iVvs듫DCH,$B.Hz>"y7Uy]!QJK<_S0N$,Dq `C gS4ȝQh@SJځ#*1l&wՠqQ(9MyUqRT!|Zii;#4ԇsӠD+1Ĺg۳)%_ٮWbq=p#ۮ%0[ &\/塣I255+2Ωj7^f뢻s?m09$d$ڙn: H+C+:lEMQo;)HkD>IF?7:kr/_crw-SqeB2acXuүd [Bm\Қ"t3\Ec-8ݘت37u}昏d=F[LLUnuuȧlO&JP(V'C-$y@ ;[F}/盲t`S,oheAC[|̫]w:J' ݣF#|A77XgX{!xvuw}B;:O>]^07ښw$wyv6G,#t=ܦvC毮Ro0]eCKҲ?y|l|8_3òDӪ&5Fi+GL% &# ]^+ΙW)_$V9B4[ Ƭdj wK }n㏪oezno36zxɰ^7v09V[}{Kf Pp Z =})>_UtڿFgvd<jjͻ@Ǭ֥A;GY8Ӻy2,99 \-C;,T㴣$$#5+"ƛvggt2Zeo;7)pNO  ݙO)Rz 7*ډ(#c!˸Z/?fcˏmETCC[]^sfDtFמA#GjccJJCy7׍7m9[SlctG G:T5*,M*[!g"8m¸Սы#z^zDaژ§{'\, \{1HF )xCү$5Gk\Pi + J>;@ YCr؛ZT|AS2ob/2ҁ*J5\]tʓyJވ\N_]`zAq.KVj7>w;`95LM^ XXA (ӓ0li̗LȴR9 MdnԮfɗ7v49n؊0x!t˅~BVic `QIGC3ΥdeKЖ&p¸Sh(|,M|8_5k6Dh׉I q Jw`TDLvu51*751 m:2pF\ B_KBlGl|͗y4۱;D`<\nʲcn}yP:}źE`.֭?8к !sҩB,rg[*̰eq/l9 jԺ1ٶl* [AMVjvO U;2t;s*H]@=HZN t;Qeh ^aLa3^/E![ Dcjh $Cb>kuBCn3rP/7 пvyC5l ـ#;콕c̮ے.!U #ז}?za7+ /{'_l pIB+GbQXrd@ѫR%b1XjAa/(d&S=vyCDT\Fü&x|z{G!PR0e"/n]nzS άcknim*nثJժ$1@A2zp_*ӷh$<>H9[5G$xhBN:9- ^BiA(Agp6]JZAһu} ‚…㟬^vhdjE/7WWw;Mhk+bʿo *@% e2ײ~ͰƋ=Wl^|ؽM[OZ6=0IZM.ƩrzAAcʣCX0qOi)lΈo^dYEBAj'ME)1?%܂K.%JN6/LaS":.Z2~>Xׅ aa64URbaCLւŲZ|h_RZ1Xr*WZ <8"V{d]v@7AȬio:0pd4I6Ɇm+5a | odd HPS-\W|h$4XN!"Kz1#/d[apcGFF Fj\xkJCeQ 5HSq_,_*m>fB&%7bG46>[i -ap얝ݏ,JOp&0kFi\0J0e &QЈ aLY:01GBN=NLTҎ M]/=T"o\ wύF_aKrI 4xnʷ+Wr޲CRWNCJ_`8hWْ84"``3aT*4) gY$R e!a N+^= T.8daR ]\}VtDem^YŻhRn,.e!¹|mnq^˅+=Z4~~x5xxVQ>k=zO+iӏw,,?->#`gξ]N#rGÞWkO *K@h*vZ/-̢ 3FP,![J Ql<%j -R0SL5ut)HuA*D0dB(Ie4iy%6'Q͈u>c\[ RۨhR>iqk)z^#wZ T BKQ}ѤZS-OZzZJω7sZ*A@KS}ѤZ-=r-⧥@@Z 8]2ꧥyxRaRF:pңRmT_4VvWycRC\ AhiALk9m8cD2O2j'#lFC1 1R)Z䆿.?3=O~n')~<0ex3{_y}ԓL:0__1>Kǧ@@aJ搪Kg)u#+AZi)Y ȗ;'a,`M?eZVRs !3 -Is3LI)L cnҒڄ5 eGJhv6{2)SD!d4n0SaD[@-?Chr{L.0zQ 4. x=[ `NF'$)E}^KH X"| ;NK5kLvS10@H_Ir !Xכ@@< \hY @RDDk#$L81b3vRsŎ%oIi<10?- R?fRKiQ'ÑkF8VoŠJ$ZRQa}L+̨&LcxQrQܒ!7~MM*KB'Jx&UrdIrNSFR}Ȯk{V#Ap8l嗇R/mxǗ".#*Or>n:ѿI"8{BT jUCj;P {^!@AB_ڱNf>$ED 鞨=c T!Cg1eHkjո6 "g9"$ϑ B!HZr^bQD%h4\3i. 1heQO5xn%Pu<\=)JAVBpԹBj*K̈}QKU,)U6zk8SԒ/pE#*rD&TPHIxF0NTPx7 PlF-z⛍Z (e_Q\'Ř*Q~n Ԛ)`gys/])my^pT3( )R@96Z)WNåeMR3mT@ӣqeVŗۧXϟ_ϰo"&/׷.} I#~˻3kT&B"vQm>=}G{bhgt:zaca֌DH- 2;=ٱ֍$iŵݡVj}|?肽E4t=x\CC{b鬦e>MW\J*g O@j~>֞ZZ:A P@;Urawq#?Zt]}0C~8źNoGw?jRinoyFȺT< T!/}jŽV.t6ʔGK[~31ljAR ryN~ɫ +' jԮ̬&y # y*S5PwU Gubb}q@f xZ.4䕫hNnC8W0->vS\fkyZ.4䕫hCNI{'J;]=oD3†N,zL5aT 6T9JrύQﻢjy>u&{ޔΑīϢvm ]v/TkcNrԞZEYggRRB& 7N4wG,%cmd+ i-<:e3w)'(uyW8^wi s=n4C4o% H~u#o~M92QMHIs=<\)(ݖ f`[s#=njxKGWjM5=5jiwOYKvtyO/T3j|=:A=\|.wH@Rz3Z#GcuwyܕEak<"NHJMlj}vc6~_ޗz_:}Y|;[IwngA*Fa-ntk4zn?W$v/i5c"dOS]bjȵXb\;OdZ;4{"׺АWQ:X7:偏ľ#ĺi27_Ӻu!\EtJң;Ȋ!dD dHH4|:АW b: ]g*e6NJyN^HlpmT_4V N}pj#+N tMVꤥG9i)rO7Z3ANZzZW=v|5[&.=G3%=7 64]rIgI>$5`\CY.sS{Ѥ.Zd6KPNʲJw33c~q-Zf58uհ;[s;?]w~jlk`I JֈD.EҪI+V"Pzh}p4c"좽J[ӚIkuľe'Z`$UϨ?|xn|t`U$kӪj&drcl$EZjm -͋6˧q{,;-0vntyqc9"(B,e9B,3"ɌB-J" %A$/ KW( 8Thv6{2P4kjWO@U:[kfYθ ^= ͟^03&^*I 2P XBJ1GP]<܎ӹD<(ROxAdM_\i;b+Ns*!dՁN/Dr @+]-;&{5_Pmwr1yZ4qӜE4M\mDiR&4+ȶ'6Q{wύ=mԊ3rL?'gbD(ZXO FY;un#UuyM!KxG<~EpXX8մ`D1:noM+hC8kiG\+ţig"T1"h{7}witiP)K|E]7*s6}x(,de{TRT97G~t\ˮ-ȵ-f= ]2%~IeA-awP2&TN[Dweq0/DBopxw#~|^ם̬x*JFs ET8zc؅@@&]3vBJ%ǚ})djQ1ȵ-t>%}n΢+Na G#〭ӎ M!߃j;ච4V̺h-!Tp|BZՌX'k 3> G<&8@bCm(FuH>yymf0L449,.dLފگ>bɖos$E9s8 DH Md[9G\SoQO!,CYp~IZP% | Ɂ3H\s2Hsް-3AsRZ"%KW41ti&(Јjpne<#(M#;2H ]3if/7~'CMzȽ U %2pCD^P00ƕz peQH'~S4SBIw*إ18ͨݨr.SO(@S,O,TJK{k`kT?ܝ 3/HdgiϹlYQ[/pqNLLPղQ=Tnǽ<%Gk/֡ s% ۴Tq2jKʀ`,'* )D U*ޫ8'dSNL.SFPh ңiՌk2`  Jp dƶZe! "90IgGj#Yaz"e"pEc:<E F5ϬJbJwQajT(fI{EfLpMÏTm]z»Tf&wBZq<Up1ICFVsR ~[i5Y7Hd(B&0[daR>^fݼJOsFbh|Hz87AI8*c'zXJ s|܂H\s)nzZY9Tf:S0/W։l?vΒBZAs|.Ϗ?\_nsnϖ/[dNGƆC Ӥ_e{H.Cr!yzO?O<7^IS>v=yjG&򝻿mn냖zYB*}fox|7_ئKNw{Lۮ q>pMW 6 ъF`At]yƮ >çI}흿Y~2ߛ6FmelZ y2F!٠ABc)fw,d7ix1V aRqZ; ]JA-"r14LkoŻU%t(*Q'^B&LA\^dbW׸޴>FP>^47vm޹$[k0}t:ּɶZλZ5Pd߈A|y#Ywdqf7NQ1/ރtTzWOGcա_>_Z0n1xfwQKLȷnD" :"պ2bnڤ ׀OjR7N))d m^ߺQ!9//B;LrP|sͿcĖoDMW/3sRl,Y v|sl[y"A=L/|('՞ .kYy)| Nؑ=]Ik'l嚓=5Y憒j$:l5jcnu&])VY>BBwO!sdž;F7Hih_ês s޽@c+@O8nŋI19,lƉ3 <32/ ڄ;xI 9֐Yr)۷sV};g%wP).kM7@1mgoOp1=IhXO*BӑIuaƱS>w폶{rxzwzqx^y:t>J%5* z]@}?~6FrW~o\}CzyHPYFvӥZQ{Zzh|KpY~oMz. ܅$aשۋTksK]>2c ||Nq(Τ6Np;a@ER^Wp Z-is(B㙡JlV_Y򍞰o #2s仈xSe~~'7 [Cl=+8>=W'$lUf12_~t QXҎ,"Ve垗XzN9X ʫOPOnѯ猬2,τ+Ciz׶ωː$pר.,(R Bb4ppv#ލY:4ߨ;fD~(nyuib)OɫH~xoy#1/~ÿ}pث뫇\>\,߮ڒ_/^bQkY4] sR'>\}ܿ=wk{'3LJF: 4QQq&YҀޡvB[2 .Ի_o~8D/-~͐ cl[XK7P=9^Ex5:{8M۞ K?4(ͧZpFR7% r4@[7B>zRQ&Հ Q|}Z[^e{\/zqv>HE}IB[d+BXD,WC,yD A[$ $A|_D);:\l߾ @tŇTUE%*5ğqF﷞UQRQe;w^t7t7'6]pנ" Y M7Τ_XZ2נ&IkDDۘШc\ _t]^!Epˊ1G$F.,NlzI.$K؈-(j38Pl0C6c0)RlKf`Q ŐQU'MS^:>-h8ΙxS|*:8)h:gM/߾g _oyj7 m-oeYB8Zm~z%HQߛ_wH^5VvTXM6HE.n]K]VkTD$+a~E* U9g (Bݠb)l+//C_ת{={otb^^eDffPZAEpU+ Roa!O a@&YVdҾwvI &MU z-r蕥!(6-hb S6qVDau5xlоřQ2t[Y܁ 84m`rVwΨ@\W96! P7GzpB;z (Q!FP&6\RH,OhCaekwVOh I tCm A6( oD5`'&c'.6洨e݉& 3$*M&PSYƊ-&̀1ݧĖoKA=Ge#yԄ! ɨm&'>,iNn`gNeH9yvSŞL8Vˍ|yR]%Nnw.[QOiw*Њ&|~mN+քN^&g5NthYgygH&5 ;ŶhW4;^0qge#YBx06Y V0l1 >,bdB2)_ EZC I%LjM kJKUwڥJLס] b8.+w2QRrLH-iաt͝riYj5=F ZF+ID{#!tkQw G-A`aaiR9BN a-͐JScuHe+[wo#]*dw{MIb2Wq')yV %}oc¸ԅu*7uNWketUrtNWyŴS>+juYtɒb@[,)g! V`N `N E0FK,aLhX3 bd}2n#?G-/]TjGU*PTy)j,wk@BڔÛ*NR*T:ItJȹxaDb<#6XA茦{ͭqj}4^XiD89jG8j\ 덗˵*ɻ}RS tߞln1K G-Ghw&d_˪ɂ˩~B95V_bT!UM J0e09i9ٴXT!Ʀ6 X^PEtAW+,pC"MTi|Hσ?u_جc˺W*7.7$4ۗPl?RFd[-,P4fĖ`sqsruws@E] R%S2PQjL}xmR&޾&{SrmQ$0@ 36'eCTiưd ER+PJu9yOsAT5`NF@+tk9\HiZƹ2PW/^k` ?77Sb4U{I p^ u,x\,0zW4R֭=s0$gs5RשTg6I5W{迠Zb{EǢg- ,҂Q8)So7S"t+HU֣ZPK9T>5}:Fd[ڽibn.mIՀ2UY h[ 3-)B$2яe Q2F NE0HX` K`*bdnUL{KrV#ϥ0_+zGD@(k?a9_G݇A=8ZR^}Œ ur sTs6wL'vF㻪6xGŝTh,t-W9)ƽE틖`ockE).[]wS2w"Z!J 1vNKR[H%7HKf@({t6 :i6 : bȜҘ[2 IsZ+ϸpkNe4 ġBxbJ@QhU;5 B4û٩/ϵV)T"vD$ISEՄsFkbk EVj$22X#0|KGQ6=z|ЏѾܦq2)/\Q<$?lvi_u"FYֺ5lH'9n*HtA01XzaS)($0X"J_e s+{2蝢+$z&m@zk>]fh>?Ek䇓ͼp!z/ oY胰3cp61g`fX8&L#D4'¾C~Oϟu~~_M<B.]+B])\.+EWur\vW=<?Bu$FxSs)FUcO ڃ^0vF`ℶ“˟z_~27}򄕗F) fAʴsA+xKn;\L- L?nO? A/_Nzg{/_.ӏ` ׯ?Ie'ߑ6|Elt ?rs=zhxj>dEB/w;?YG9,]EktOv5Iс YL0TŌ }.&?q{d\f Q2L~7Y)Y#(Lr}r)π{z}YxFq&ix~6A*=<M$Ywm`54F'A3Vr^M]} /z1 hqΜ>|zx6E-~Z[ӧ{pcpH<{\+/SJ4:q- L=eksًLN$Z׼q2kڍ,?`Ƃi4saWj([>P3f3$3Q8*2л_\ˎW,Pw%Ywߖ/5-鴢-^?tT~0m]lWsoBQ//9" ziɺ;|I,ضzsEöe(-J3ʶ1%0^&X Z(V⨸B} 7$[L@9yOGh\ ^TNwh5t"8͟Ya#oYgޏAՂji2ћI D)D)x">P]>uU#3-ʺݿY)>j y@!V= >+jd Aڼi3J(_ //DV3a- ;$Cs jmk<*)V9Q.P"JRero)Bm"( RSJ!p07)E 9+;@!MY^ǦW .)otƔR-i lha8:JFX R &ƱL2ǛD*WKa.S,G0JQKOeQX* ;Ne2,Jݧkᘫ'@QDzP1bkT 8Õ2Z=a80=%qR\%)ĕ.;KG`9>]6՟"Ggb&z>DPHs.++%{J_I"PE̔JNXɉR*cr ls < H( P\ :e4!kOYiI wccem#ߐ8W/4u9$8/Ҟy},[9RV̗qf ab !XAFLK uJg$ .S0aboygpq cBG#ݽS" AUAn{S%Gannk9=ohjʂ1 -tiFZ'<:P9F8dF"I҂}aW$$jYPd %#nJ4[L.iEnoq5AL_ۑZk|+ }[ _uu wuMK6fZC ᅮ/ X.Ӂ'n\<g ߷zlw~?:] NrggGʆГ|&eSƥ.;cgr11gngt@b-Wև|&aS @:RL=%w6zO!tn|xSF)=#'T^&w/qaMoP~ٙin6laba"`mV)D3qDlHiʟ;ã'{FA^Ca@af,~RK^xP򡆾|)'*i9yoXZ&,R aGByӍZ*0^= -#SΩ-w96|UM[ޅZ|"r~וScv ;LĿl{g`1mEA-|ɷ?#fE=쪘}tЪ4; >i4< 1}/ӍcDòXL'B;QX% /8x&gpt+r`-@|#'w΅#qͺHŇʎHK4BGFz]zteLI5;0I|pΰ 0Zlڃ4j'/ݢor2=H6#fS")MuHUv ɢHQC~yVaAx)QB 4<הtͫ3L}Ҽb#bKhAvѓ_B(Inhk50ᣨJ SBdʐUWTU/d@[:VhZ41p'k [81OJRfPV|T~[HF3 c982W=B#!FdUYJ#$Q+4ƗϓM1gRy$rih:.C" Gӡ8:"CF<ɩRC NC|yH|=F[{Q'Rd%>g9q%3d:[rk4VQk+˴B)#-u9UzHy Dڵϴw΋ rȋtoV4hs!x<*I&LE[셨ǦQQNTѕ_( 0$)2|ƻHOgXu!)Xї&;z]Qrx!o" Py,4@+KW #FTQ!eI+fǐӄش*ΌìÈτ)H{l0Vg=,vSEulhg%5M5pj*kJ;QTkrW"06n.2ExORu!`n{MJ19PZ5p֌d{wM! y%5$2Iu>vVqw3Zxwk>*>hq_a#ԍT.ub66 ]^HV&K}YŐ\t+`U?@q CGz=R^KCuXyW6uW*=C*q5ୃwع%rT^T)#RØf~ٳjZ+(#qƚاj(RЁg|eQN}_5 @y]iҵ!&O:qτW㐗(_gT-ȭx]4})~ހF T \NB<Ʀ: qOGH!u̖RYD\SK@|qFPZܕT1 i kl;#ݱy6_2 NRъ9 x[ 3̅*})`1ǁ13Y(JE6(¢iDɿ}P1|K _u2wwn,YGS0O xй D&3ۀP*JFtӠLP!K̔/k쑴]Js~K25,qT`"HƪQ-b1ƻb9(2ZAt\%| Xy$ygZ846\bYǹ z X:D(aqYz0F4܏OCYH*BPٞ Ѥ:jxT[w HSZA1h{NCBN -d p|WPJ ${-q1vQ2TR5>`CĚ'G,r9fdJ<~s99v9ng3yKcf';:tD[,r^.f*i5ߒA/CR8nY-/8b8Lge?]tIGOK^DtJ2pB/շ B+di4+*hdE2 37Ci{1;6'NϥFi/_jDZY~YOw}К~U;ΐ`˞'2 k4#/V-龡=SaĚ{uqPU_gqh _ ]"IQ*ysxa$mIfpPo׶Datfcs;=<.\=<_hs6 +pnn_z u}y ӊ{iQ7CgAA3Gߨ &rp:Yz>D=Mb}&!/aTD[e%QcT M~QԄ\.!EB6d3ą^".Q%nMd:]jE)եN$uVU<!z?ִɦ_8 |:p&X}/EW{;|}]He ߗ;ԗl ~ys]3jbA36/Z.fq9X"^ zXܵ"e&N]{Oc=)%_Tٰv{DluTR&h8bc:H{I,mv?X9a!_fT+TsgvcmNg\ bL')i@г[~٭ MdKy"ψsGL,jX_'?bN߅_Iɸ  Ow0=}[X\/VR ޛ ޛi'xwA5/*[Ipl"_!-c"a4e"%6 eВRqg'?}>{mۇrt=Y/ Τ6 >[5~f,,ҺMS"ecSpBi%\9 C`ڔVZ JGd䴢^_KA1.ӂyQRg#%Az<;ngI ;%Kt2C[$sAGǧjw~4e‰wބ~2__I$7۷o>^}&|oNyۭ_0p XRzk*3iKaVqނmI?3QVi dMdzhOmHYIm6wLJ(D)jC|/B>S\Mkb=G~' O629\M!XI-pRAF5J{ 6+] $!Z}Kh7]&Ǖ_Jg z`[05Yp,~vj…}fs[W>v }<?,dXwjAju}u"ku ɪl_Q{_}z,aܰzxsmȸ}{oꟆG0Bj6˙;o .ʂQ*f7An*{~pj&@3oPMG,j cir+G{mtBu]VUZE &<em[+#/ߏC۠izwAQT֍IӃk ٬iip8m C|KY·voC|ר1d'[5UZ5Vf|G^Sa)-y%qWrYe׸4FFz(A A/3+a+ 1$a<FY}Q)21,|''1S#h4r:J(| GУvZv5ob '0IPLcD KmRG!4!v]Ft%r_)?]9)gj]`x +q]@7tHz"bT+O@$sAW"Q>xcu3rU~.,)9 -GeK0/%EkZ ՃjO/hb{Go ^M`"5 (a2chpε?ǧ6'W%{ƧU[3I"_!v.W\roA(ȵZɁ!.rɺ,sTvԞ_QbYPNJk A`łZ l'vR̽%C++}M(9 ~ jgzU^kW-(?KԬ^c&v#n18X3h cn\t# , B8f1Zry}vG(jϮ(Я;Vf0JGJc7#\G+e-kdK"g^S(qRЪTcLQ4 r}!d\˥V䷾Zh9 9Y. p,ax͖+OeMYf:2`2"R"@Rk_sNe՝kj-洚;MO>Nϐ ǜ^/wgxIۙDM]Ocҧ^c`wK^fذP>CKP>@8Ť(??hjl}tyc;O3`HBI}LҲtVo22;kxOO/",ۻ9p2`&}Iw|*ဂ./B|3xvJi$vκf$}]y;vd*ܥh́~Nֹyx,#Yݤ8SO3na^vtCbuSm@;q>wRzta?! *eX)}]%+w[L %h-i.( ̫(U SkZNdꢖK$jQH€4tM @AʝߴXc ٜW rMV VNh_Bp_YDh:S#i/(_^*k=vi{j#jѭ38qD)ەRcfR#1n5eT;eEei6G @3]Z9em)7ɢ>2ɢ->_[MŜa:,],úXIH>ƭZek s=x~߽o x<f7-j4xi`(?#,iVMlϣ ZW;E&FEF,amV\ES8’u K]>h%vv8^_7Fb=z۵{TweM cG&jaÝ!Ml67Fh'-ճ<0%GeC6۞ ՞I ᠹitf ֞FӣGمGwd*9&Xy:1zDL6x"bKKaC3Rbv?.mI^ Yg]Hf;,Bd2ӟB/bQdh_eo܃|CRͣT=LVt';:2b{EJ9 5heBI#lifšچ`.W^oM7`=ZYӯ.࿘MW]g(&eZ&"Qi=`O Q `*x.y~vtscIaS,QŎ?nТ~7MrRYŕP S.΁ (7*W΀d@2t/v: @ *q)v:_˱D3` . ,ǎԢamqTZ2Gە~" isϱ^@aHRGS) @DJR#P IY~Pe m+|ta|@'<8VHH\}@?-G'=5:Lѩ~+c"΄A$T(@[=Cg,̏_u}q"O0 z(@WKe(`(_{ca_!'DI/N}1Keiz${4W97ok5y\rHۢ.8/"FZemVw򳺒\+ D *"cR̈́2t,$Vq,P ,V_AP:RC8_TŸ QKO=|I;ZKY# ~%=byyxo_XF0 w`ļwe/$З}x1lA_ /;LK){!/ X]tmOػgRLap^^My y$d/BՆQAE5iIpL0W@ĉQd UL%R1KX X}Ce-  I %ij8.HeoB/=! | H dz Чؗ}fC؏¿ė#DdyH)Oؚ⏄8_ Ӗ4SFζ KSoQɚ4+yn׹L3uP6O⇅B<1'?p:Qn:uwb"DR `&0v{m}QU%щEq^roF.6ְ߿^{ѯz'O^ú*iW $<p'!+Έ@=!#L(6ݮeC(vb~쌕A2^dYcfB"!:DJL <5f1#!|pySE @(pJ<ň")j AM(2MJZnlNB5:$C,9PfXKp瓲F)@X#*a)F lj10UXYИ b(W@# 27ndl(wywt; uVH/}=Zܐ!LD?vZ ؂,Oš9 L?/1:v{#(BGI̘J*:G\9ǜs  K o*ƚ3C80Jf= #ʰ5aޗ!xI/hfOU ֠'$ %9 nL)ĉ 4 SEFlj4~`#Oшrr>Lu&X[ ŭ.OKEgP#d_@B*1LIbaJY@šBR CBGI h`Wfu LExCU}9ž @ ւ B92 ʊ-G 30[o %䅱|{mgg_m ZT~|XFv=Uo N8Eh.ގ9Ӿ'ɃYP ӮGe6L҉.c]2n$k pHÎӸR DLvɶMZWm8}$Y'gWgros*t;žX>Z>=8Z XHz*iz_',o:@TIi[ @̵IT"noC1.4 c<2hU1DmHܥ6CxJ_6WIn꾸J`[,i%V684f0JZ~ŗXL5όp:׆ցVgȺi䥗; HPRo;ѽ%iLd26E8uaTM7lZfJZsf5 8ڠ[6ҺFEkڃC1&uZȍ H:'4a4Mulpsy:Kysүa=E `)m.5PǨA?0AuTEMK9Ԓ1"Ƚsxbӻ_ڲwg)3;[ؔҘ@-85']J6X2qS0~O>9: taU2 OJP~@Wtj\N\HK\|ۊ)~x?HF5aϚ&䶇R?-ISrQ Zv$3<43ۘ_rք&9q!t$GdFPDUS҈28[7]AɳZdzbAVu1LGdJ=i%٤Xѻo0%)ܹ\>kơِ~}u2(T\[漺BzijBsQhN8tAMt )E[Q"^~E]~ںvڮxr?"$m%F VvuɄ5zFJo-=֮SP}BEY_ߵ)pXxO6#gisd\Rd>JƓ~+DB \ͷ69- σ"Q{ґ`|JݬѢӪ݋_4upU(j<\Y:?PDĠ*JpX-KצAj;֩k%<|k#'2KBc%X+-~uaGʷ)o6%yx⩖C4=MEO:SL/).BDn,v:&,Y$WgmjwL CKFʔ#UfMRgLܣ\s28]vǤuL*}yG>2Y^#`nz9:%벾Z]/2y攁ыk8SYG=(ѰCE$U`LpPasu l] nesBнAz+ZMK+KGYdû|DBR]Ҩs\{YdL3`[iP,n!ևՠD L^>MJԗHWɮ<_|VsLAGNCb8 9'+N;͍VscSϨ#*V25m+RV΁pkj=wm$WqƳvjطN;[<WbǜGO+,2p Ihèxʼ&l_3WZ/ E' U_h=Ȉʤ==zv7\.R/P4}Nz>!wu&zNV $ 55F Mp AZ)j2ZE=KIDj"; ~:5u9{.PN>N47i/̰KMdkKAK^spu=@$QW`TUm:S`c xAIb4RiE =y`2.wJ A!ud4˞Js RjňPj}I zJmt14sTmFWJRGGiCMfHpm OE\ڠ7$zZTB(ddVSq#v5%qiv5C_;ה \SS4]i99KwMT Mc fkʆDiD-ϓ"@5Z<n}1 l^ST%"uRNEO eb &d&ƀBL RV6B|pW^mrC`>.n*wyl4fdδVfs9vH xf^*jIQ=0\q.@aohq /轑|HZ-%wٸaŪYj֟xWW?ՠ }pm=rY=91ȅ~5Kx'/&bwF=ȌI>1zAb"Lna Lj $Nf2)%l)d_qv#-6Y\;*m>lkk#^`rB۪$`ӓZ+dPKy偛qk{K@|k$4"加nUѾ\L8CRg 2W0}œeO{WKhu%ʼns`}W>w 8,Z{-y}XmNrI%0H5Af*g|YsT c9A(n,$s^E?nw~bhV숔@͈~^Yz%1tK-0 sA{ gY@Μ'ly{y69>_S! XE*r 2"`ּ?RꥀHQاwbU<KqİBLcV{!e'!HlaK9[yELuZq֬bdNsyR0Gl[귾PE5Ca9ls?$ʹsSAOu@/,7WWr`5ƌdvCy j߿"B/l_c?€+C 0+@L"GIP36FS?xPJ_~dˤ)l~5"xHqw=(iMͿ7w`Qԕle 8%Ifγh2u[Yof^K]y෿ۤ&mݖx䠴W64p2QY6Gz_[xp$jOf(+t=ћ58QR*Ph1 J=x @ʯP䫹ɠ4ߌ*60؜0Z4?*F"ub J;?H#E,铳OΒ>9k듍NA)QIK@?$Ƞ5%Bp/*)Z")RW֙'73sV5h43G*8.*aj.c /(&HM6Qn+)9q;xEQmDZZqe$V0y,Bi4*'AD6@8@[nWA+(&qlB<|N!-\襇P˩#yHߥ5J~>^$j $4{.m6V,UpM6^Tu=T`8x?޻Y=9'1Fw_%dgN6n}mQy&"Og](O UޒqR;hr Z,{#)F&¶DY>BRM}Xj8'Kqr턟+BΈ7kDb&!Xᘷi2ޑ(!hh !f%\ Ddy"؋^iӼ] 2(LQ#T=ꁄfP!}FD29@9vO +O'^piNU&8R99cHt'SyEPE?g85QONFz+8kqnlcF>G8r&,=|h))oZ~bʐ/ؖ9ǧ24RT_Gv\Y=X!`r4:¿Ho*t߃خxr?" 6lKx:q1',-u_plL&/⒚pدW7_M$E.[Ts/Ygj_78[&Kr}NtnyhcLY'Wre=.I"`ٳ9roө'.I_)IG[8ۛ1J>;Ű) &D;"ׂ(DGC*D[Q"^lh_؃⨰_flTVjc`'.?o) pƜ!3KŮ|V۶acc I$|ϢnTD306oY<{AM88?4`: 3zRÍ? ,Z/`DOPbv:̅Iu~^EhkR_ 4+M9t @){zdz=Vl%PnHb=QG# {Sw '|vD 6,`6GT6YP_L4ȱܹxI&55^nS Q~%!k,1h_/:rY+Gpqjh9Ǹ٫L)s"pQ.:l'ET,*)AYJ*J"Fk"nG+Gb9=]ITxw{WF/4,^f Z_~cEiB =7Zդ.2KERP/%_GʻٓbȨȑK{> fhv]( \{F6(uYX'LKdp\A  KRZFڂbh xEoV6U=3cњ})SDK+(p/UPLRPFHR:L]K hz3st]ͺq-lZ+ EZW𬟞O+_WTLx _(J" ~g'ÅU oB_/m7gq9g|^wO,Zo?U ~geTIj}ߞ]$.R͖Qzp6ix|A( ̋An׌˧8Vئxl/r0ݘl/qLTUb:+^JDy hqJH8{v?5$,:,2_k$@@f[Pg2nkK0kGo<けM?Cƅ s-.Gsr+E(|;^\s\y]ۅPO2&##n~N$V}F |La%5#9:2BXUQB RJS+P+ʲX i%++5 2-IX(=StЦ~heZ`C(iifz=D`s9) P3- ]M@*7 [ٗaؕ!zwÉi */ [@+"RD=y^v-(5lO ƴTMQ?*nT2Øxf'E›Ŧ=ݖ,!Ǵd/7cDQl&*FC{s{Y.Y |(ݙ»zz›g(YD I(.6(KxTvxƲT$mk,Պ~ή)a4^=ߋA !Fn#XeLP0WQyaWҖj7g(qR:s:aDHIH&Ն-:$djI`tbd|CTͨc n j[IQte,"Ssy3?]VZ\Zv.R7:lf3.|;u5˷$,3om9G|hyfp7Xry}pNTO|lV\}޺p[6e=-]aVk#-%$J@ѥ`R)m9 cȣl{cBM9ЁYu7}7 f4R`o%aٛBc>`iү'PT ).,?ݥJ^]Yč=aYl h~~it-4u˃6m:+(6ui݆А߹)H$qn!XBQ|Rt҄%ͧJ)fbH\tTsybۻu b}5ω'Px=rGf1 ZoI۽$\r>a甖Ī:dim i W ӵ{qJ^S<n^ 9 ;Gqc|;^mWf #u]U6A )9R:AZ{9ɦl8I@R&lv׬V͉#IV]R9U^%ti,PF-3??绤~KUzBBUB!+枡B|Yrgq+ %oi>?VR8g*}ly yIcYŗ;+4VW?(s,Ʀ/ 66Tj 4,-!]GZ+ /ƣidyVd:V ÙhvҴw7 _iln"SD$S'k-6Һ.-#*cn@ ,`IZD]/Ig5Fy X(16sяEw>猰YfvHUς@n/Tkxʂ<,Hy٪0g&UTky(?v-HZr;,E ZeR!nL8J5eySA jX[y5Ek Ժp1-̌2Xp .@yuhZpx*_MErVJQ epxKJc+~KEueNA 1'<9p2]R_ z1!xuRSDsQ!9W"hkqHk+j7lڳJEI㉨C,lh#wbQJLzq,my t_6!dպCޓ^kj 'J&q9]:C2r1'[p\ ʐ/]BP4פ%Ѵ&l^cWSy aUH (` #%AstT2'Kɝ!-|v,@aRC}T#Cn683EEj)O;3REDSBryпnPTQH?]V6Y,ogf5w#8u[;{joy᝹W*:ď>[gA⣕"NF asܵaY>!Y#^koveڧ'Ey;պh6mEOU(8 e[IK+^hg>]D``Q7vk]3}oɄ̀X8 s&C|GTqǬmqV, %u\q-yWáXޢEu8"-f"`я{!e7}G؝j-ՒoF5Ud:S ޸^4VV`\g0@k +\nhh/|ş+JhOϽyy2Sk < o}QaF ௃=UnӪ 949TѥNa3}XDo7rHw|)E!&L6e7EcgzsCh$R1wnnj"P[Hօs=X8eYk-/c奊+\LypWTܕ)UEX93?umG*RId'Gv}!d':.y! j2#]ANL83 &~C{x]7e\6g sY"5KhJʂ\VG9)T)e'>>`5y$8Bhz{J٭)rB~ o FD}m{;p:|%l7F%;n`Kd p6]R”`#n $Bm/v؎D7.F/.}̷EcN Iy'8-:mȚ'7}`19>*[|X\Rώw<դo(@yzS%۳D8zťΈ$d2(yQJqo+vOogaܓYSaD@1w.[网T^ZkSkNS@]M?S>|?_?<ߕ;!d u0\p4l*=csn΀XϩpMD,%m^(Kr/&<:l^fEͨ=48& g'xq'Vpp8_a̮ [6_fh:n@JPޓR Ε`(R3j%(sS:Qi4/F{62mᗧ4 nRE8jr%ܼc3eBRK1 Arm!P|uǴ*HQ fI!؋:E^qn(jqv)rd`wshRyaȑJ@ !L"QQmUr)S\3CBAjhaVV8S-L1UH@ Xd"{JB%oCwsZPf4Ή B\$sFJ+ơcegfd \ѵ=ݍ"A&H视t{ O֢HPtгF`SB+ƴdqAn겓Oxnj"2ձW@]d{{1!G8.oE,BAZ6srځ͑CZiOAbT&\pF:g;^mf?[qa]dSa@M!:l밁YNhcNsc#綗 +sNr"fіJ"et [ꋒJR_&,RpT2A)Q2W8[RøKC8jMNxQ: [(@aq JE" F-uwmي`J֍;dz-HG#QRT ,т0|G`ps8apl&Gct7:Y@d9TIɡ++Dd T);)"*V•髇x߷Uы|yb&5>iUۇƬj`erh[=2lJ5s==~!<+:B?f{K@.yf*{H=j3#A3$2 3T0<ûWz?x݊[0p}Rzp.n'Ô65rkk'/?78 ~G#cðAva-,Aܠ^A[ӸQ<9֥sW~bStϖjvf yq6Z*KéHNvp8՞|چ\E)arrHn4SA5,X߇uz(fbvYX ^O©v9:ӋPօ_^w?/Petrn]Q'zCWS2?]\g?kR񒢾 /y# ѩ w;ao$W+\W'W/իZzsr`e;HB޹0z;$1hT bD'm|/SE`-VvBB޹&{MF#jT bD'm1}n;#[EtM" >%t}:Ϙ'l-05bsS}dM@f|[gv>{g22?f Nj^yAkF B[k_4Ɔ-5ILQL9| [Xp~:`I{{ %2#î`!ae!8Aqu[GWZ ;#vLȞgE+/[Eh..'[hG\Ws#~,V><:+DMV܇N}"}>r7h2άY0ftbxw2r!ݔULeS0FTFJ1f Uَw<݃f;u['T@r|lg#-yso"SBcG76#E$x-F&ƈNKՆ3k(:siN5ctIVRh̭!f"qdSdJ7bΔ*(4V.)c~iXQ_BØ.KJTZ:p &9W†˖ ]U-!@"4Eiv2u9Haʘbrux@(W]=ٍM#F/b˫Lqqy)ñj2:֊D/4 Y\]i "{ sϥ墏IɫKX%GlNERĔTuW鿇^_o6B ZwW *`5\R.;NZ)2 )癓9*R&^NyŽ-*RBG#tE=n35>1Wd^ζhAr;$ Ycz?}7$pr[*i +P@_ݷ3"cIY&QPA-g@w\nmCJlȥBue{bm+)ĊSmc>6RRaMs;}X$H alw3)R܆uGDUC#K0ToNBRE.M4?Ylq۾Cye:.r[dIu<3BcS l+7Hp1[ux0#ǀS_eh q46UfRE+x2;! c9M!(f^{D OFy#TǸQ|a9^Ł=*hSۯeY^[ܕ&8Ro>dLRm]Y4Yʼ0N3(/Yc~[w>|Cp{;dǿt gKQ[g;N!rT/ ;#i״K>,:សz2¯5{9È7Bm.TTQ#&dTSF5' 9%&f>n25h#wŽDT JJjTJza-959$C}='=6t$ f8d9tEGCɨE5e.XKM3'sYp<>>o1a:O3w0Ņ}#سsϽ!I Kҽ碁WDjT:8dg3.n:S^q/;f G\tf}s>0cXnLg9fL*#?rZ](̺D;㋈Do㹎+BEnJ㊂#&k5k|g*aBsX_!ecƳ8צv+N8Q<Bn3:/kLZ5ۇи4i&CblG{xjS<R-W-w7gVuI$HgvB~l2 %)+9C\..?殢溜5%ޤm K8.TF?EDP++H15̅,d)l;)JRD^9+l:|f/ޘ2HcZ9xřry <{,w@};_7[C]cs`wuq e7ZXdX"@ ֹ7OO[ʲW[IXӣHI#vS_dot$uwۯi)@#uj8rl};l͠/^wpcfmj~]qv'=-D¢x]t+(9]ܷZl~-z/;1cCH~td`'=M~r}fX5}%$h;K ֋V=t/E.t"Y1 1ISvKX@H)IDe@h/[/R/S@ zT+"([kx-u?x u+zk9[sUVۄ|1˅52[aqeqeqe"G'rT%kQ:-"'a?{9O" J|VQe<9އɫsM99K5rO؆&}t8y=ˋ buv6&y4&fgNfoC?pwz?՜ SWHuokq[8} Nj} uPJ2q6H:ML#9ìDlt;a^[ԥ;3Z?9fux8ߧM[uL6O|'^?71!kHAs:Re bhrQoumwK-4|vV 唫ʫ_V:uv|<7(9U7yj-|pwN%7oX>>MozIOR>FWR_/ 9u{hC^-=^XR8Vw=)Zn0(B coOH({]8aqg|ɹ[Yh}؂VOŐRsYTj::9p}o=m{p_fЈ!eɅƣ x`]zy6㵁Af kIv=M>/he ۖ ZBeQ|;M@j"4Yd\/ccURBAE͊gP▾6÷Q짐MF[w֮j3mqձ5ntu[=V&tTzUh[$ nICN#(J{[31[a~Qz*}UDu3&cuZl|SM#m>he  6:Js&RY[shm++~ (!R*@&(Up)öPGx?l+1%<JaL6th=ORqq&5F=x(ęoΊ/<+M` m# M\ RXs6=l;8:ig}A;uQ}#7p~}`H<^ R4LgW*ʌaɿ*@H"qc94x+Vo([߯E;LϠ;8y`,}C͐P3&qf4v4ԓ.vBa4^qMjcBښ K.s3h9!}AXDjaƂR]f磸z.?x 2 /M^Kͅr;\⒨2ֹo'B-B;;}*I@eqwgn< Ď/+}b2W'U=dE ctKn͌1w*y`}Tڻ4X$@+1d;\gZS?]4CB8^嬊T1>4dlD :SSzcv\9,d }8{ء=P09w1hM`dQx1+SpFM e Jjij;s72]_fgǻ?јgjw@{yè# Nټ什5B~mIHzȈ/sB#_iĶԈ=`1 <͍5zqin Pom/$AvZ-1~xo'c>d>Gi$h쐬4M JM 1ӿMCOs-=篯7G~7"[]2 R[9ˣؤ}QF EKsEzG#f$BwA:4]ƮhQpswX-/ `uΟ7WR;rzej:g!~=yG*I̟mB–3[5Fq\cb!  DE) H K:K! &* @xJL~zpͻopA6N#qi6d6/<$~x{~|5?, h.쮋l!q}" N 1X, bq틵2fRr 1yG`-8@%Z kIithodH}C*8b6b-*6A3f1,BL$9J^0{ !ˢYw0ȖɃz`mĜL8{n.Mj bL(f<^6&iLSd©JZ1>$"uȨ![bs^$%Iz@-K (KVed/i`^G%z!挂8AZx A8v *BoNv-XS~m6K!P$ERk!&CDKB1@)4xJBYfb uĜB[W!ٺ0(*2E"RR+jK?Ĵ>Ą1YD.ra,yMh _s5E=S/~ S?g9Cq]d$g)?=IWw-b`~@W73b|+xut)%Fˑb><'R>h8ks??r!q~1O\,ճ=y⯆;pj_̾;9;x{x)a!axN@L'U_lxUwFy<fzU~kx}<߀l77ox`Ea̯&Jq S8b!pT'YrFK&H~D-QGju K=v~}Ɍ[I4UP5T,k 9 E*+yf6A]y۞g3f!4'o3ˮ]:f~Ǔ0iWiLJ!3̿^kZi6x1<{oNj ڬH͊{@Kq.B}3\ yreNg]\#f>uP$w[vB:rZs9Q8*rRM0"j(|ĤdusK$4*rB9)[R0`,AV.+D~sSPZP!2 1ՔQIѳLYHTy2C -k2"zp9kȢKV<^XUŊ]rQoB.71n. }g;E+Ekċw(*cF]0.k+/rutEߧ_hv D". ҋ@7dSoJk{ M06{lLP*,s}X-nl1MКUupn@ͬxƶ+ SKZ|Uc7hCUq:{eZOh UP2 fgL҉l zfNf/ߧBvsmNۜoy:ڲzޤCXM +/b? +yͳ՚[1b"Z9Ԝ> c. GMYۨ>ْ5ZSBIVkkn6/s EU~HٛTǮl8a`D)$wMCJD 13Zٕ44n4*\VQ-)PGz?MmgP]4jLd=|fN'!N)no3`wZeEpĒXȜ=i@hA-.B7W3{ |&1c 0"XKQ aHROmXCpRE2L 09M?? L𔸂 &/bG}Ł1`lFHG0ZS\Sϟğ b3-%1 bLgAVoo?'Ohi=M{/i-IA-k|Z=~M_Υg`a*pbc3mk30ǞqG.UtVuV<߸.<8"ѱ~y0 YaZWFϔ7 Nm;K!tL^.tm7VN1?-I>[dO:|%qca;<⒗"JD !UB}xQv˜ΑYփ#k&$3.+o/7wʯ۬Đ;h}9ۺ:%۾O[Ƥ@-MٔP} N(,MW@eŘY"`4PQb.؀Lj&5D(9v_.ZTqtr%L]hg>f4]w?a,d+8@\G=S1DU\’R;cA ,ラ2* &VXV tBnbu`X r).ֈL WhN+bI + rHc %XP(ht՞d@;u +I ՊKGTcP{O 0pMN՜wk45٨ JZ&a|`!C0 Q;U9]6ﬞ^H۟&N";걪KXե?\ե x=miLjX"?6]"LR?o, ɻ+6[ ;VY _rS삡DwF)V#pU V᥻'s:oIN:jV?翦p쌇"NPbk"ecg?7nEĝPYB,2 t2if8e " (BW^N%V,ʒC} `*"xzk=Z7zt:V[`ketԑX݊1P-9M+RTߣDMDMDMDMU/GK!ʗp [J%ab'%1(vBJcdyrɿ܁J~؂Dvk=fbp.I)avn0ch1( / -4 I0r|ɸ܂zCZ:O(FN2lch). `2qKXØҢҖZS*(g9&H&9ƈu&Z@j&N#'bu;*}X"ud'IҸ~c 9; 1^&"R7IΘĄ1ETPKمB69F%T+ } sP[$F 2,+)(mL5BX7AO' LD@y sox\V aExaᵠK9D2ņzX ZzN$`!ſ-I&ѲDnR/ke[^R.J)nFa)fNq880_,1QrK@Y/.R¡~OHtJ1[$nEL[FWl!ޔyôpLYY[(PbZ%1 VI5s6,\ -%M͠!%(1ɴUQ`-=,AJ/\) *J@! b%>͠Hz̕,()uihr:UO8-bBQ<4R-,0NnYW@TIKpCf;m{i+8.cY̒qsx\&9EMz `HKUcSQ}p+s-.B7W3{#|&-:/2': LG`2p(>j0{)drKLhVxҖij JŚZu*1q\/Cet1!g Lwj~M4[3_ГD`Н=C|uФ ݡMX`_oAroӽ3s?-Vn`1owYS(z^RN.ԺQB/uol5[`"[ތ,Ŗ=%c tӛ0U2A҂-e?!c:^"=$m?WMΤ{0pAzwf'-!T?44Aк8rWrw#cn=l^mM':>+n?w MbEnMb(U;=0 M[W^WΉg:wګrیlBvDr0INEg5,))I5}:"^tZ)bf4鶚3O}>!h:%zT3N Sp3h&'Lj1ʠ{K[` j\viLbեhaֈ d˥ip'Ēc\&s\@ĕr3UG@^/mU!@Cਓ1W 5% kqTo7%jPDiL:gJq--"}"[|w迃nۧZ7ו~-2E"b,8p ѫy")y&-XKA(YƜ[NQiްX՟p( ߁΃Hh6 Jȋύ% "N8+H` :M ƗH+lg "ޕrcٿ"۠Q2y]uљI̧zVǖI~oadU Uuy+(k[7ЈD2n!/l R!(K%*#E|rj\n12f#Z,%?;cDkE5 s|N", LxQ`3"=}  |7% dC8Ch(#GD# %6k2c!` .cHݼx*A}y|.tF伵+]s|x׿m_ !"wv1 b[|yUz}9l_j*zyeF#m=εt=E;'r X\#iBK{/i/-5jZ?+r|4 7BPHlӎ[݃ކ `WʶU܏MJWQ>mo%tDV\,?Y-Y2mk77PHQriy1ۗW3:dF#W!ڦY?Qve T Gq$,SEp 4Br 6%&oO}M>2TvCvT;mAv8e"m_<5 E5LC/J~m|:}o21sN 3MΦ?1&4i%[!pע9_sHpD+k:LD%ל>߱*x:xP>R pӋa@ʮr_}pV4`:Px@. >h Q 쐗rQR}6\O :c/h.Q4q,BPd_7^}yQli;L|q4bUsPF.߫!_9e/ϡ pnr3m@I"TgIb}$2pF(BRu]+)b(:!(h_R{EmbNj'tWMpDݦVk.rB`&`Ɣ E2 T&Z]:CAufc~1Iz~mVm߮](aC`cAAdQO0_1Џb L8JGX y`GI6!WԔ.Wq]Ї\5.WQFN(j6DԻeʫrA*u)\pMc JJȕw#CV>ԤR$L*"Sb E]j2T9K 0|zDN)'("琀acD1gV͘T(@1Brg:) ܯTXD)4Whl M R!"GcI\‘MkfW8y(Ox(K "!Io]Å+*h_J&^}c*zn}Toi/UFS@,fƺ۲b|l\92w?xdVP“܍_һOcw4>8|nhigX;/oSܔzYYewNs[kZV],^QR$o^c ̈BDd$ ۄS0cR'-^=kfn|.*›^햅!*dc&bs<{oYǤ+o_'P~LNFn[X(s3@qɕ̓bMCr Ʃ,i,'oU$yAM!T uBz:+tJVX{Gy(d_mnVU= pJV6fڌ8) u#ڀA&]ڀ1 R }Bi9D@-C[RQ7ObOm 5."V&ElFi +j|USir5փPTE| lcxM7MyeSLy(}@pdV 9n"e.V{ʧmh>>.'TV&kvT'}!sb9h=/Zfna1[NoߧkcPN|j];ޣgnQOo`)ؿ|˼YY _f[US*F~)Nj+h6E+UCwG7!F4gЪѭvECG&Zo)P~CVQɁwԇ ^^ mnM juhJ;QF)ֵk :*9[?F͒R20A^9]/'Ԝ M:ٴN6 y'ֻ'kHB"(-1۱5o^ӍV!9ֶZbU+:Xv(dp9%+3?SAWco9X / \ $yabe50t`žcXg/=Irlx)ki><8 [t|-bTv -6`Xz%pc萘SfbH!|&nUs)\ j<3cuw5:-.aCۚ RV VbXFKNΖmӞ)[aNñD5v ZŤ ZEԍV9o:Y{X{]$" .S>q(x%-[9Xk׮N堒mgv3 g/t,{Fe<'l:u*fBId'{Kd%z̩z@-+1L]*TE/*&ɲk!v&T/ MUGEeAzCE59Y ƒ*x˔$]I oMy[>4AnmG)g|w>Nl Qͅdi 9hH {ǺMA3-oY(Tto.IyN ΙDHIRlB= V8ZWX<}GQݏWojhCND5$ # E r;_A nRѬ;z>[+/ 0tADNu!.su6Lw|SWǛ#d4.ɨ2^WaA>cYA[Hr5DS鏯 \? ~bKlָsOU[?({uà+l׵~r0WedH)еne_`s8k|̾n@/-Łk}ުf51QzٸUzTy5YUҮDUZ־V,S.Z5,H5ˌ92Riv`@@ M/8r֩,V+SQ9$t2i5M3|)I--NpL[7E/&K&׈GXup83k~|N}=@@!ӣH|,5DЛW? ݇*BAdx,p6~AtILW=nd{rDD= aickD%ήPUN0g..6TCX vuƯiSg Q5u| Oujkny_ǥ(N@ztX>&{n/cOl ;+v{|V](R4-e܉h-HY2SlƩ2Ȏ9sL4AֶwvTju0T219:g< pG3gcRH{y&0Gf)Z!Bw`0 s5gYZ4ÌTSJT!U&DRMTHv`0ʱ?|s7&qf_Lk EAJG+yΓ9ȪQJ쇲zENYTWԽ]U|wm#IBk9 v'7 fkVz$ ݯ$xw%]]U]]]+#81udtj#JU,Vܸ}7ľu$QHbTdȐ1;il?T*%Vn-K &ZoNp(1BZ%#*pl JSOس (9*v{ބ]_g@?Ӛ!{~}oc\#r}}b)<]} C$_Z=#D"3([R`sTU1_+;J?Aၑ?&TCsTnsȿxcwi?C/"w"~H_~h$ʽ4;>g8 uqN' Y1s"mԂ TeĨF׷ vOa E] #JXwj,j:MHǓJ(G2bQ6-y_ϣ-١%.|ō[~>YMXp(S1ɓ(EYZZgbܷ7-_OGy9Xd߲m|3KYoBE>ĝ.9HC_nG1f\;!dGޭ>8_v1Y)k=?yt6_6f`:_^^=Pmǔz-:d_~7sm}(MC1Uc@ݺ۬1 ̕<̗<o2I'.MaJzkZab0c i W6_xe͡?#\f1|R?tMgY~zC |ro_^M&X}Ǣ궨B]\VkS٘v$WCod7T/l 5D1 ?ciق 2߾Dq%-.|m-~]u>)&M> m(]\-= rsuW/QGU2:l޿]IqS;7IE~ M]Aq(?xAw5޾2,: Gs%hYYnQ /dFFb]dWցi>fMEY[uD&N5AX딐<fDku;lS͐Bt@b=IOݘ0IAv:!08CQ>dYj%޽o ϶q޻;&32Jيٙ}֍`98덼ac5w'Ke@+v>fW3ruxJ*qFR'3_=jX_>;Uދ9<{WH gwdϤƓqmOgRi:H'ԚxWkss)f1V`uc ؿa*v9Ya~puT˞ ,P|Csn Nufjޔt8ϫcrl. }| {OձCGo3x1ߓf'A+B8 [ wX~"3d|ɯu7W o},ɛpxTWgvd0]aLݷ1b4Onɸ`oҭVRK4\J8eI~tm!2QIlU%X}nW%`x?c*JێU7Z`K괿=>ӐJ>T1\ֺ kCVJ0,}r^HmV=$QAc9j~Mlym6$J\9fD16(Aqӄ!fjF$,/2pZBkdxZ#O5ÞY!0G8gz^ZùMDJ}*Մ6Eٞ~ح3TQQUJw^>[7p0 jŃtk_XU/xӓ}vmy2s#˳qL.ކΏQ27Uf7\Adܝӟk$N#]ڇEU|nY' UN*ՇٷnкAu;_eZ=ۺRn]hwtJ8-кAu;eAY/[hE@rCT9釩3^{Vw:ٟeaÚ;~#L'e1Z]cF37Me|MOnӹ4,ya[GDsJB7jj_W%${*726{1jSᜁU,-@6-7fۨUqhÛv,7;=OF$\$ZV""9~ 5"lV6+PJP'X'L"]@l,Ydɦ}>U=w PZL[/./v^8R~MZmH58Oh0fl筈{D|llO,cT:NvU&bvB4^VN)98^.EH2&uPt`uC,ɇ WۜoDl Pg'#…=t>(0N˨yClp x,>ądPKUU8, $V?ǚ'"#1ijF")Jx"5I|c9r+++*JLKpsةYd6YI*Sc5&N2ôf(S:;5 G$K,&&2dKөd "a(kGF'=5S<36, LV&bJ1A,o"&0‚ucI5 I611C jD _19m2C{c,56qw^QrPbLvz#Rs80ѰH0)a `XJ@fc 6BXv`FR%i2 Y0 L45{k_ oh BwRcdw[,T?΍s@]8;|RB&S*N݆#Y֌0gb0@tQf2标C1r2h&Q0WX67c|TQnU|j4!XR%TwU0,9x)]4d P5MMneuT足k59a lUayg@ > VXqɨwԾ [<$s1hB ֖M$_^L7(4buK'j1%Ť-I) Ivlvә&J{,%NJIA DFQJʌqKlsle*$։o-֛"|c^>OSi۩!֚JLDm{ʱ^\e6x pp P/z;'[eXYʘ D?NȄH؄)87Vrg|IܦNp],ު`9V>=ػƊJcTB0Wˌ:VNkRX$\h56JB=U LC[o ʹehC0 &Z3038u'iq B;` _Ѡ?7Ȫڮ{ml¿(%$]&7ɷmbJxejt様R}7@>Ǔa_v&+$rI!]یqOŚ)J߿"A N<߁n_oI(kI+k{5DV> 1)`#꜏5/% SQui&ѱeֻpz7bas򑁊+s_мRRNAm` =ВQ,<p C{܇5aCXDHJqг[51RN`:4cp&U0I&N0LIdPћ I1cH۸bj]&1 <,Mg5t87a/۰:|BiEghr(5Ž *CpD%!9`g9zpa֓ѕ@Ziuʈh+En<~t9B 9Km*(ZϮ퓋وL)wJn?e{@8 O%Eofpkԧ^mxY >оC]a`2&6 WLXƯg3pw8#3Rc:JPy%݂h)wؠ\-M:ѴW:ݴ]X EI-?nۮHdZ5%x(QW.]%NqĐM"ec)):uQB+= `IHk(N*cY`KԽQD#t:ԇO|#P"C(|HxJ9#s4 ځ,֗ޗIXQrᶐܗFS4UKu6] Pis I%u^_" ; n nj ʉJ ny su%r*wSQI=gk$̭a AAf@Ȍbm*+Mpuv0a7ZWSaBR|X݃.G 95R-OOE뻮MO G|A!7tCӪpw|q͎B]=B3(ԬsPWdZ=GY*!j DO}rϡH5ɇkK #}a߮?<+AcE٘:tcǑV诿X8/tig;ͯ*Cx_Y:%u ~^A͘C9#΋'_R;"p/ҢZ"-u0V=rW5@ceM' *p\9bOsJC5=$Ћν7C6Ԧ7~ҖOo˕b}/xK@f7ָY2ʩ6ҭ-LwV|Ti}D2mK\o|l,Nj)8N$S^|Ē <(Uj .f&\ ̱2?߀v:Ha|%/7k 7 t_57?{YExbRdܜA[X-FJ#eg=h!  †T/|eNw_֋iKe ^B!n@. ҲT]˹;*w&s=p=KI!df1)˭P]:?o.Or8p1`J[䩨8yWA$$\dњ XNBα{Q?>:R?AL77gїyފޡ}G\tth+4|֚ I1*ňxqd}|-}x)4IqtP+Rx0`ii^#܊& =A&`MCr(^A)D}ycHe6-W)z )}C_6JUtTFDdqjʏr6d,sס2z4'ߒ)fL$NW]Nx4jpʘ>[kWA@̯٘êuns[@CD$5؈#/(;ʐISj-Nm3QQmxskJ*T[?ƨY=?16@)aW$T.gV+ZH@VVH]$gRG&&I! c^= _?xThmNzGL; ȠH(CtqTĊ b02SbZkhtAD!jKlRfL3KLh ÆOH\ĸTHVP.@oEoBg!z h@L4 )|襖ịwf>XvZhsChC$t -ۦsUy`e3r4fJ HfUG4+[/ȵ0 fcih] A8Rj1g|rX&Q鑿_@U,)(>CZݾ&9#FbuBCk$Jm2cCQSPg:B*4P!-G\zEOsaٰAuH&dF^3DCpYF_!-*tO8C&7DK3t =`q{D݌.FF5VrP4U6ZY*k 8a ]ZG>ŖDۘ7bP`f1ygANK?_crߥsh9H($9v=H.wD&-alc0g(m5EPD$EB֕x R{=&IN17(KB(u(>G`nע1L<6 HG(ClQ<=.ljoRLj5n}y3qhV |{W՞ P@q@;g Nka/ m/3-P-GPͼHkb&wv8@L#:((uo#v#Ya܇ˎ|ήX-Fz g++;ɮ靝̠л/ԣBf"zS7ॹOb/WB5 =VT$ h9+lk3As}` 7-c}ޡ`xjr.ɳ_InyU LRN:.p>>ȘCmƗL;qV0%H9Qʔ@:"Q%]2)UJl:`2A$|Z `WTTN >xcN2hMZ&/%cN2(@AG ̑ABh#A|û40 M QQ+H4z|=ADjR e O%у#4BwߠRNs4SDn4H"7//&_%|lUWU\-/R"Ua$z N*JvQKE]^Dʋ4b \)#;JMCeP0T)6Z=53FJf eyXIe:i fRKY֠.53Ri+,F΂5F{t|4^2JATf1Fm(<$~}x 둤MJKHI; ڕ y/U JQhLA/MjjS؃֢Yr3kTI-ќ~{$:t!uyZ@o^5/sB_O;m-2E%$pyt?]઺zц$'˕~ԓaL7!==s*%גP`BxAH EMX_WVlU C׷{uy鄱[v ZHEt,TCGᶺn%LPKq/Vnj)y6Fv Մj GcN ֆSdfL{)s:hy_Ggh@% yzUo8-Õǿ5 w ן{,k5E]r-JirvgYn;%wj펉YtLZHd C@qS>OwUaV,گ׈g:+n|.W-wv hOZ:CBc8Nz&)8Z#^tp%N5tYQ˜P;@ [zs䠣"=y"KRׂWBSQX}ݧ@u-MvnEAPL<󁸠J`TX'U Haҕ yK5NTXUTk躷`8DFaL (l N"Ηr}sln8}^;#?սv';/ϲ zg)'sc*z=b23{iwϐMdH΅aO-!{ڷ;BxU < {@e~0a2,APіȫpXK7~~P[ǥ{|Z-r*# xA[WoUnp)p+3/ܓ71ÒuRT+~?~S+=nQ8;vFmmR87#l\~ߤM*nD$L1ӑ"rZ̮c1QQ[! tPFX}ShVsN= ⸷љGHj|q7g<7} }66l(Q|^%Jy*DM4gV -5I0oѠc^Vf'j6Qfɷf 4A!CׇpDw8X*2 g̮ (gjB$vCۇc}8RԸ FJ'MV[=X`ݸaXLhe;[5Z~=oQ T8;C+M(ˉ %É#8T̄NŔ$h$gR?{ȍE0[ny؀e3l 2} `xvIv'=$ۥҍ%MfU*~yxH $lW_c8=M l}eYG,Dva3+:݆ pvDD#k1:xSOxTʏ7蔥D4b4XG{,MBy"8 #`V7@t=Ymqc#Z#:-r/$gk]V*kP.0TێgW'~q :وCӍ6 q~tǻyo1n!l qWWJbwOA_|bn.VQRWˇ$豰_Bca?Y\Њ q<Cp Vŗwq@`!خRCO~AexKt H쁸Bf82zʔޛz=]H㛻fHf4fz51ʨm'mA;5&~i^#QMl1Fl`}`۬LS7 i]lw,>\Ӟw,ZwńÝ&❧"흷-2cqD*@ӽs]N\0 aw?IY7LʍG2<)qk#jBTVb{FY؉SƎq (LK0e9d Vi iik9P.vpʧzXFR3Dm)YS;݆0I簝bBI6&vu[w|lSp8jStx  aMPrDi1-Kwr qkiRǪ0z CMjnz:]lxHfD_y__ tz}9}09 3[8X73gsxg>;1vu s6 to`GX"Wo [|4V]2f'bWLOhާ= qaZ9+T.cZ!o|f)G-X BAb2trh\g%JK-?'Q1B)8A.r>C}Rް+^ ,y9(,^(NVƝΧaGq6dVCmpGSPNnfA${9`^oR”:?u1 ك6ݫd-Q>褻t} _%] |dĴnċl?⺷Ž/{u Q:4uXa^`=qUhMAz5on*5\ˣz%LcJa`Tz ̰ @2TXTkR#Ʈ X\[L|ƅ- B_XC]<6L`7Q&&)JQC52(ԥmܛ, p%q !5l;u%Wj 0Bk=xmx+ڲ&j*v؎q4&z= `_?jfɋ5n *g_ Bxrxp%.=ǧK~ՎR]XT0CR(j-wRlx:sJ=h70&=Y/ffz0kJbf!%1̕XWf͍&'H#. OG X0p ކY&ƾFCBftaoD4\w5ܟiՁU( "!d?#< Dy^0Y+4FHNkW58¬'l>Uh>*xIUnTyx<ϠL%THӻ!P-Fkts.DwwaVO'LTtMG7qytSGktJKX$Í VaK;/#b՘z: l6>}wT2,+%gp!4){隋D`y鱤%ז{DKy(% **@J Vvh-RصЮ'QHBPl=6J  Nf%(u 00SZa'QL [r$Ksd;EwRڌD@Ƹ.KMjϹT ay5PD!$roۙqAq}0%wh f50' fLoD7u&A&JU!{M/$vb9CHgIR;K;.i~:br.ҮveA~E8G(PvAI6!+ogʍW(vDIJD!l/0NHِB£VmV&=OC]8 WV>m% z 2Sn ڡM(k5 JCإU8tE^ZJVqٽPu*x)CvlE{0,9TgW|9rrVu9% E7u$ ^X[D`"hTDB# IV"_??ɤG:\2W0]RwS#>!r t3*Vi̖37}rbVH'N96ԏOF}/;8uFMTdWdJLo)bC5{i!HM/]O9bQ/xRM$H5t~/LTZ41\Ke2' N *fPB,XВ/0B<3^-9z܋UPd` wUl;&XӉYTn/z>#,})`1 Ff"TC>Y}}׍/@)hd RE>QD]RvHMG#!Ez?Aߤ,F؍`=i$}٦l_Q${I8xt؅A Ap ہ1M{1=4}YTy[سQsvFKF| [Ov^6 4K$x~t`!f \I9%؃$"UIr̓! #bR2 0> J,'l kbs<'JQ]I.8G"oX+>R${O @(zc (5`MXn}̷5v.7/c`z,2[dY -~~$jN]O ~:fwwחS\,\ ~;=m5]Rs$<&[8(o_| ]4N<ߧ-/e@&Y[h?aXs@wLA^L?\ hO !Ïm{mʳ@D;m$-:vx@7p B:?[X0?egn&)cyMTAN8EJyb幐` ]wGkB3f"+ I-{ۜB(I۝*4pcu[B"妿 ).[ KR+Pt,Nj8Ql|jgt;(! NYRTCNaE ,uB0#NJ Zd`%ck*rwr`=Y)KRNH{Q.Y¨`=Η2R 3V\0@cV"lK8fNR'T{6(8̞X{3GB"zobkHk~@Ok 0| !k3ks k@5[8ny)#Օ$%庨cMï+&kb|}hG (%߾Pv{k^/Ls7]g mWH+ %isuɤ..1.~{x;y}f *y|ox)fUs'ϟsxzkV͍=>G.}+x4Iq3ԞR6g.}dNr[B1Mƻr"鯗Y/7]>aVrUZHjҐWJ:EpW֍nU1pQ}nY#tn_պա!':daV(ث<$dvYuheg:Fc]lrX׽g5Yחiu\^}y삚w zlPllB|_'F'>gA5?R*)oc#ǩdEbY$ɳK_2JXN7izŌ"~?Jp).V7xi8ȔvtCH94?_(~]d x!^[k{/W FB5O#|Q5#3H b}my&{Lin 36&|0TlXlmF]CYMyQFYL8.OO@`1˜XBK8CBe[j\35H,U,"U+ş;{7SzdޘYl01'k :*QX'^Y9\-sc`Lwֳcf/oU'S/+Ak٧^6>|y3QɣE^Wd[ߩ,?zPǞ$M%OrHP]4䮝ND*\?_*F"!FHDTؼ:6%* Pa9&f! A!A) 2UDX'ue6ؒ`F%sH\` !m =|MBpkc[zGz =N$gOʲ"Go~ru`;]ɷ4a)|SA@GrvGB|LE/Zl &_$˥+g8#l#cH'\mR߭J- g]7/H[Iru܅+>tm=եB J@! -rQmTVaUt8a.C(j = wy9]D&7xd'7=Wpϟ.-fNVbT_QbM3C6v 8F..ڑR[#C F:Fa#ڤ # OdJͶܱr%p P*QLlyfN` $dWV[NX~u+*@RH;K ;, `sL-'<ǹ=T+f[=N$,ԂcH@#.51AiZ@p,j(Vu3phڌ56WC #vo}L 'Q߄H;?~VB.VjeJUg JUg8֡!'tl V֭*.S2ފᔷW_}!hАW:E/0Ӛ b6 ]ZPO #6\W[t{˱Nڴ!ܟ@ 03O6 V%$,rX^L OAƘRhOrkMﮀ7(k~ZV+n}`JI: H@^B Td _<8۸pA8O5T%k+QQ؂;Q*>*1"F5l,^#c&îI9۬zTAp{[}!>t8[!a&>{76QS{7k3u({Be/6w4/pV6'Mf_: `n$xTUV@D]J_oĬi,Dyg$㯄Y*JarSD6bڳ~|wMzB2J6e&den:Fv\VƜOzy{^O$I4i[j6WLUElc8fʶ7t=|i-|ԎPtT;B4 ,T@4/\H#*PHW&E^5ٸ,?~$^z^JF{nVnPn -&SOw͞O~)ꢙ]՗z[(#ri:+#3.GʪLɚ QYLo\ V4á4d>@NY몕#Y;Ku; }"y5rsf7t 'D þNWp3_Qĉ }7IEyy^Zm9=N.teAAGMb'7آ ͣm4G?fNܔQ EƳ^z 31!m9ډ.!uT7ˊ \|c+c$*ǀ:BKe%5#~^lD=9ASd6 nA%|2z?0Ap)0 :MJݵQ6yۨ_^@t yu>wFVk[@ΚF?:FZ]F$ r*ҳ) (z]qxGѧj>l7&DLa)ܻA8hlQp NQF`H{< 4 ynP7i 0lz )Ttr) [Ag&L~~~u;W~c}L.mVd`؞:Iul1o~I"I6&f-[49a@A@)MC A ""1D-e۾χ +a4vt}HU:M<Ԍ^nz ҬO/}ԅNhG;_! ^lY&F?~x/?^D7oOg7[]CzpfX\Kcn4 1D0P8"a@(Vi {Oo pH8_@,\H~]Agy۾]nd> U+ƓV# z`ٙ8  #l a $qItO0ѣ-g =/uoUd&IzVqirݫTqUBfA{ڋ97%cp~K ׹Tp(ߡۗ{\گ?ib>@eO`}Iw?|o*o*I|hph;\ CrمyO[ ^Y[F_p\(Aga}|yF]O"9{rPPxuNuC#΋ƭ 3\qD8~TZ {4zJ#a 07!b1 -06*EImEd6W2(!ݪԦCM7dc`seYv 3oK<6lu/g eNM aV>Th>XXA|eLTh2co`~Gq'e_:#x!=g hUa|&?( i^i*}%Lj*i)J-dm|i-%X`Qs@+ln1َك=a.[}X׽g5Yח6Ʃr/Ò)8 O" d@MܡINv}q(V)Q.pjsMJ?fN@ "39!!T10йjG* v.#@Z&ca,b5uC&2..hD#P rf2m6`J3(`(8?m+H1U\mSS L<#J;Vk(8OKƯn%QC{U^zbGt en7Yz=myK)Rs CU#ؒ^DΕ@c`01JÈ',4MJa^jxzzSU 0Y@i!Fe>GkiHfA0C&F gNb!+kx^5`$ٌ Ik{5l?GPțnW1U5R;|ͅ{¡1׵!?tsN˹<(yr`AmRB(DZΐ._b\yI㪻KǷ׼7q׼h_~DԛA6BD\ \EG\ybݡBca޶XaDS5G;3}=4kY"onZ݃%{͋_7i#a4xwx};.߁ŗM}he øk.Uvroբw~[-A]sv'[z29*9]spv'7\9v&_ȗ[6N{wY{˳"{#렭lZQˮZWޞv2iHci^b̉ONbCKAo;@nÚ8h[FuSu-n}nwqҐS䓏 YSyo֖@u A[2weΞ='xk4&;3_@p2}*wM翝b1[t2G9SA9gEs@)v|zEj=_|Zc10'qd3X S5WZP]' ܠLOeමoUVVT'oՀ|:^0.<53׷͚}Ϧ!AlN$"0bTy3JS*זQ4! 4AX\-3 eB~-n{ҤP-9XK)+w5B]Y8++A"rT15 >,ʤNr"^aN5BrrT+wXGW7o<B1u&ՒUXI\Z`Y?~ؾ_u^^LlY,,`'duagnf᳝<jLR|>Vexy>q+0fѥ Tw::|^ FA1i+JjW,VB)=~Y!+,+_$!qM)^ 8mK))hT bD'&6 >O%4V!!qݐ))Si[ʝYDWwy|L7ş#h8 =ˋmb`FOX}pjᡖys p}%b*ƨ` ,\ 66{k56^Qo6*i2(beL? Ni&ueK2o=0(DH,Vbn`2eӠdL5#:RQ25Q'N5u{I]]_'/o6DC*1A}4{86[CBІE4EX%T-&~1["QG#vJF "|NС5rk0>2^Yvv[QڗvU-۾iGqn$ O@x̛x8Z4p3*]^,FOհ/Z]H߭[>;o \-~;֒J }|ܙOfZ2 >Kk#TdrB>-P{94;<­g#EGK9B\r,eRw#šf;hP ,G #LD#bTIM$w #2pwrEnA92h,#:ez0C"q$#D4dca aV;$7[ u\kԝJ|ꃉ&qZQݠJ=?ڰҊ(1 ?dއ 670 2͇ۯ ꑫ1ߴ4Esyn~7M( Bei8LJ<_4})}8|5o2}(vkKQ(ޑYOiR r4!vZ"4 GL=J£QgB`3>_gTvdƒyُ\w}?6pq,9TyYx[_kC>\j>  0fERY `z?NK"Q'j'ڢ9wh C;Ԅu6A[ds#lU5tpCZ9ª$J]'ǒ!WLJ 9D,p|nv܍<Ǵzi~4 4:uG48ЄRz%a있!BC'6%q6kM!֟thsnS܆ҎjܻԄJ>|DMeޙYOF[nShG$izNW5CQ);g{']>2ДPpvN(w L(xwg6%Bfi\cv:7<6BAGޠ݀ %{]uypD00˰NQ:~LX@ }0QLhN,꼆uݺIhFٙ?& 75R&kZ[6ad jGA+PI+lLwC7F+톃#tQhlC[nL8m^ `¸$ 2QѰ*E,ծ!$+!y9"稧o|Q̳p _~o]zٻB@7Ӳ/o={ghF Y]dzYg\0d~.l0UH>8q3A؁Y.C+۱;t# mu nLX7$Hc8/ JRH\M⊐w"k.R:};CZI2mub_}GR$\o'ya7E<%pj|S̟4鴵8^=y??ߓ%,-^9 #c~;!_v$L9M eOq[5"9v;Ex;;8j?~Sscs^pfd)0 9X)F!BƐH͠. ~(J,'wc1H!t8>kJBdep8&Myf4K%wrAm>Bh~Dӿɛ40A1iA -iؑr0owh1m)R[T@38ʴ(4*(\BR2E(0*$=Xgg]qy㇋zRūlcF"1:h݆JO/EcB[ =Q}K{. J5WO8otp}ͺ˫KS&K#_+s|ay?\^}K_ﲫ姯|^1ۧ7^*{1|bR9vj8tӍN7T.@fJT(xV  +EZu SiSIb4X]K?ȣi||IhXW" @.Xߴ^~l.Ο$oQξ&K?܅>By"M:I5 afEӧs}Wy[|o 8 -ݒ0F6ȬF, ̔=hьj"Ors$3H\J%BH.V^G,W*E*\Ir ma42>RaJۼ@ X.#Dp.P7J< xA73k# Yl42v)1.7E i܋E5 T—CV_׬fDV—+|ᾼ,fTrOBUÒ;[},K]qI7<]Vj}Nڣ2_luw)6V0EnWH[pחD8ьŃMuOx=\^1hZd<̀AI{|- KIy#('[2"-/S(ޝ}' 5'#åv驝Ӄkrv6(zUSo]j[>٩$"g* 7IKhD9<.}!:9h! fiuޥgGѩe #۾3$)^v¾ 8弓A^䖪 U SGWaMR4hwBUJ [ XY)rCjA)O`bxьxΞTM)h%\iQÓeźẃ;vy0dZd'|8`6 OuVz.QƘ.e3+C^IMޞfM8`|faSFT[F{!ܬTZ^BH:1>‡AdS%qN[U.tђd5Q{IVǭ ph؜=V5#MHtKpd9[,PF)jx" uv sF),}zM.>ŠSzpZ_[Oe-{ֹPXuɜc1_]i8sOd=1ٟ"$`u2i-RB9 @<C ~˜.0zbC_օcrsr !&?\; v f]z2w0d?7_8$=UhcWƮ6VMZ!h! BH/NNݰc0Y4@y>Y^ʼnI~ľO>'=Y<=KY-_ŵ[gٝI4^2 !$JO‬ͩYWD}_ a2la駄{MAbGf~kSk%4shVO06E Di*y w1 Vq9D!52xc9)iC P^0JG @L`ΐ7X;-+Sq‚~HܗuSguu6K Hpޕᔳod@[;#{aScL, #Ot`MV/U YD[=VR211#]4૑3 B"jf@hPjbzU"Hs0cV(bbSp ޙ߯awL>,9W˧'isLB1QR!k9rCjypҭA>ta%OZfs譄jy)^eKGW* IS &m۬I:J믺r;Inܱ#M<ީv>5[x>fɾןH?<Λme A›'s$zdQ"MHeDF@o??HQC f%a]q7Fy]-̋ iW IU0/zUek7RVm[=`͆q5kmtu]v@v+0r8 `2U]/hT"@-c}<4GW hB ˷1%+N4$5$f"?aPlL V!ˑyP8(rTZm<86$ZyȸZUߖGWWtH5\Ío3LH>KWg[!$">WL$yEAtN7qLHF>8ݞH ZIgY|VK!osvIRS "k BXNt+~*]nHktr{XUQ ,Y &@MI>ezY֕+p@\L~CAm܄[֎5'__1|߿d#<6nݼUXѸǠSQIoe A~UL:nmd S 2c}# '0 !mJA%!~˖2QT9Ä qj%}oOLJQI,k\we(ʾp(?((Q(!2j E<(98#:"D (cQ߁/kʕHj4L~Yf~“.2;avm`6ֽs||~Q4y_ŹUWչ_:BTFXF=^aya1C(\Ś[ 󱸂z<,bJb$I GYK./eڭ'1 E M,XU^[Y"e1J3f6+@zδ4{`A(Vx稅Hn84!'y3$GVoт3 (Z \*{- ץzYm }g{,%FzuКs.yx8 * hhuM"&2jӕnU̷3_-Cwec{jOOrGBE",||~.h}(~͗ ĦM;ɿ~~||QǏO!8/ş şwn|CRs_b]x> E< Lߞ|#Hg|=[}qK \7h,(-L@#2!Vh8TCB45@p $a Cƫ)NdTOC@\BVNX4l,2sgѯ2IB䬙&FAH[&4%ޠE ATf»('Z!iGT˓uT qĪ}j-\^Z/Y?n \0$tss˻,VѩL nmpF{B]Z^09Σ5Dvp0鐎[< HMS[U@GAlD*DŰ~,FuABŔA E̩zJ`MV/UK.{ Z zXIڃ,ID(`C*g0A1|+bH+KP{+BDh S,ci 6B?3ZSofXHZG:m1x"7%0Sx  RY\O^jn)uj-5+{cr]jF(AbTG4Fn6N+ΪR![j,䍛hMq2z$E!\ RL'&,DօqmoS!tۗr ܚUy~iMoVl=IM9[,Pm$k3 obu](Af) M]F `_mĒH%ʪfYo,D'Ҳoz,/[X vLP2'=\ .V7S`\-Wf\M`aַ-w՚b2T5Qh^hR--2UcRa∲ g(&zcmq &g_n-[7ό]俟H깳cbW-vbJJ9 pP / ׮( dCâwyzD^Ydw)a)gt~gD3i͛a,:(Lǿ%3vBz:#y~u0x` T9ۯϞMH[ !t_k4?\ݞIJp.c+"[^6\#xjw&cMi3g.ޥM*t?Im]REQ(2*rHngyI:[S 'σ)kUK ({D[L$Ϫ,㹱hMʽ1I.$iʍe5+ m4X}diҚxR|p[S"X#bD?Bm +Z#,3 JpQw[g錯JpErʪhݰ򀸬JF6b$@j;09gG#-2'2rHk@DWXv8L+H}*5*e"<ޗ,1ě|/={*-y4Sփ~X~*]"[۱| 3]Q>?&wEqol(/{BơgBxYz~tؾN4ʉ xfUjoJ6j4AZKWK) mn!#3 iF\.m{XF]A|-oۺ*\ۊup$#!iG!*y`lmQ-R ydwP:*o@3Zl*r5{PoI~5~c;IibyqVUVX?]O?e'`AC& )ewB Ca^}0S۾Kz-mjq ?m_ӌ_ï-WV%qltZ,FCc4J 5 0t9WUEtNZW6|)+uWgSM} p T_'ZUga 5UAr#CWܽHT>a "*u1CjE>kcڂy^ @upsz?\jE-&Ȍu@BZx:WZ8M3tBk%3@zɱ`8^˨5 VZ]=\͊!g{Y5^ZLA=+p&B2p6-ːi3Jw=X!$5Q _] b: }G&XS+$cYCd&,䅛h-Rb{w#bc:uQŻwvDc[MM|Ĺ3nu1kRcxzVwѱ[Mt٦T`F|nmõ|IzdjK5tŵh_áӗ:hczqqSؠQJYJ{P!Xfe4c-Nu& {6Jvo@Z Msop+rf7%U3|ãyύ2^K\4%V`4_[Ece6mC)%l0+87:<&G*LR̤̓y3x0u(E#GITb n~ʢǧ-Hl>8b7L*  +KRRৌD'Sد}Q{[wP'Mhr?xD-1Z4yiُUBINkJq2&`p)2YҐE&ͯ3ЃNY34ꩥ#%؋VkB_DԤ݈sXj9]cgA*O%L}AnIB`"/E^-r1az^3ɝgʤIA<(LsSֺ>Sq}K3_>R<.u_lqJ1rN&X'F)6` â/o+ aḯ؆cF6\a򴖡ċR*gL*rh-g*5>h c:7i2Ff3dR^\BX8`S~?k{Q`W2S*a$:&NʁҼHu82.1WX7Yo1*d(gÂuarSTZ.1U@$Yo5LglKQ[eH#EqmE Wր@Px 7-䯓=T P]eO)1^tel2H\\=3@k~lW/yg=hX&8hinZZ{pX-Ƶ&.'-USN**zx*d3g5h>EYQR5jؽ-ۄx=H,WZ'=<ƈ&?ݹ,FC4 3Z;}6ѽFJCLf90Gs88Fc;FJI-XVx~v5s}Rt=:e9@c|$豶Pp1(ka.qmښZ Tm|SZ2^}Y;#tvw+~_q+Gsh\u XNh]]&U`h2.u^-EiQd0S1% UNY K)jkvFo,dLhNyՑR,_oZGPB@Yjm.=l^O*czg!>μ>`E$)&L5ZNc3{+jd9JeDWފR>Ԡ9Ѩ):O2r<>[3ҦяΊtwxTM#I iEǤu }~ԝ+8%Ahҕ֓ËA̓?y pΨeNv-hP) F7}i6H2'kMs>c[&t8[T!F.LO{ˠˠˠ˲*˴*`K-9nR_xZI+QɌG}Q-ڃ/YZ? ċgЈMgJ-51WCtqi;f3./VBS=aE1F\qYlWؽr&dLmg_8ޤ|~.{"UfQcoogvwwJgrQ6Z/Uj_x +}ry- _H<ڳz|w729n$El_}|$*R*Msо)PO1Kڇf)cyjԉ J!ՅO::LGYUnҍ22I2c_M7DVE4&c2S:LAsC2J%PHY|rL[L+%ENfx4Uxs2Ln= 4T`RmЧI Ḧ.JguBf<[!n~ ?/ {,W~ZͩOsG#=4翽{7t( qǟ 1)@>6e&d70dٿd?);|ŭn#МpdM]mof4 :Ƒa( ) -&i<37bM> OC#7 Zn(V6;!xiBe6G˲<^2Z,E4W5+8E[[phj+!iz:O,@*1W"_]5L W!9TjA H^ɹ@8KP,iUZn\"HdF-"rU*囸M;1nbpE; Vx_\[Ȅ9uဵdkS*S1FH\zep-m;rA&? P@QǒhFe|o:M MGGL me Bkd@x!gwPnK2d!E/K~*+)K~_`) L[FZbBVe}T)L*,ŔmabF @ގӬ$8<\G#qOAZ(,0!¶y{wIw׏-pDpgΟ^$qEbplb:ĴFam6[UI]Ѭ4fHZ\*qãp ]ZpX*ܿ{]x=.8Øy}d%&4^4hS~"VQAh`uWDAD F @{|WW2VKĵgWgrH 5&{&W?nm򰈹e(d{4un9hɮ 6(Ě&R|:w5P}ږ}eA:1:TkȎVɠ{".!0fUt6V,}2hZ5F ^MlK!lNv&'tP%ܸɗ@[ZȋjsVڶvv~Q}n~[__ ˙˽M,8{ؼЉeYgIiUY>&1Nʦ}4D*=d(.;F 1p~0FA["Tٵ)u=~]chnsaб]}4χ+d;UÕ4󪍚j'Rf0^^*`jp۪E)I 0:vwg~fhMQ^lLPI+e#+T؁pA|^ OЦF qУW1MҞ 1[jh@its. mk)RPܽtv=`wu|%p#mVSi1ZVl& tbYnTݧ:tr5A^{TX܂VGp6rSNio0`zpZþrL79bLi4S:𷂋?%71J v"VMMx+$q "aP:N dWRާwXs'Iv֐: **t!PX& NSӨ/ mZ6LV~Lw/0W!+R\r{U9b(^v),-M2z㌳'K`JgRSED8\z %1qVbUjRa v7枒΅虞=̏o G) ?fJe^>Q]yjƿw`qܻDy"K(<̬Jv>N[{Cg옗ְR4[x1@k_4wq,i1jRu'>I沤*};dT(-I9m_cHQC1p8 "34nt70@?-/Qz?^2&oQ_%o~890Fl1\s#UpTd7KiI{dC֧Vi#W4i b@}4s2&o\ Sd,J6h5X*f(niU,e+bA`_F r  G/zU$v1e[tt櫤zCD sxL'FPJc Բ(7O9֭—2QI/ǐdoٗr3>4,Aeه 0Xv:qG,D~3x"||CSy!ܚG`Mu*-qun(0>9 :mdze :0hU!RBKna,8%^O Zy0nWޜ#/&ܬfegz_&43*-J7w_RHwo߼v޷e.(1T\9|Hoe`ÚiBN:*%`#J!H $DX Ũ* T!ud-pMK ov`Rm7^Y'7idM{^Ր95 äkp<[#0! 1QRɕk,jH>)O>HA/H 2A睵Apȃ+ $ Y֓EbspݯW3l?-VPa(ݽaf^qAK.4[^$+7LG/A P 2ޏ?h2-4 Y}OO;7f:nhX?pplѐ%LwO^| 歆 Mf?_% D# ⃭QXk#Zff#mFwc3*)қ8Ԫ?\BdS&EsXJHzoX37)ZBֳqCcM2*H(SvΓ@+1~y{)L7n/q|݂ig@Z]3nned@{*Tځ]96jvjm- ab yyaCQ|{:yr dF縦ͰxjǛ/-=7ŸFܙ]'Rvq$H;6b:ʋ1L_c봯 /rqo|W#j:r]G3A`$fk뺦C=xKtfGfCkw6ݺ7:$hT$jT aL.[uAFe rͱ #(񶵥Q53ZP94Y`wۙ@5aeLr ^slZpt47*.!fPɿ;^(;IM];OƷRNƮCΘg&D(:P3c!'Tm9LqU/ߎ''Go_߷&3!uf- OV!4w\#o*xǍE~9Ȼ?~L/[pǐGV4G*(?*M~rAO({ܕܣv%sxԧ#;+owQ*` ..#LAu_S2 ޝKh'qoOf5Ŋ9+!$qua4ClM49Vr +׮(A0}BsiWM,Rq _1ZTPܻy`REf+M-*<-|s?- .1w_eLnMXșhMIurk_kot$HDOn[d5&Ƅ,]w0TY"`B14PJ'+C&_]%R[E7NƢ.'&IiH'e3)BPҞ.Iŋ\<PȓT%iQ%뻟'Z~MQϖ6~LsŻt_5[?ԵCXvTrFz 0e Rv#vR[@ -Rٞ[ &ZMZ腒CE *hL{>@DtlPǤa5*f^,O3pÒ*t`פ?YSYu["O4 V.$UK1$q M.%xnnczSd()y2zҸe2Zv͒uΕ?N.?u,FShTKV:I`*E{*kޢ8#h2v(2*!-EЦ&CHז&Dj8B?@4ńӖ ". b 1h/RA%3$Y#t5_(ǯ俗Q "V0_޴AM[ӕ׷QTOك'](Î v;m0VruVuI0Ń&8,x $ 4a)ל3ęfR`i`'WJǭx}ۭ׫pY04{U.X3E/ &Cȟ0@dPeeޏ?h2-4˹U<(z-߹1w=߹[x G7HۻVIBk ߐ?p%* k{Z00HϗFRa4$r,NmK n%hƵRsi F a /@X`)<_;'‡ڙ'_A AHp݆ H ]pF*<TB ujGEKw6mFt^PmhqNzP08nYq}qCXjwD!C _nN9!X|0~}я`,LA}Sj~}f>xzkV" `to]p(8" Ssܪ*xȉ#Eї#V0!' s9PQ) C]?IQy>, !vX 넃L_iDi(YHB0ihW&9%kBö*5! _6@ٮ*r 7tLq.Q++_Dupʴ:Ei1YZ2Dޚ7)v1XZYʗ+Kfi?#ԩ2apiP%iri-w& /~u |jaeX$9g" -mtr&b2FZ H9h.FCdDu4f+l`Ii_gv_RҲ +IVEت뭫#IS9=I[յv[Z[k}XJwXy4协QG%qVsz15$! A!-X7j -L g;$&,_ra"Nw!^8\QpiPB *HPDBr1b)4X⹕ I3iܹZV-Ug4RSܳ12 82j6;cu%L|b@` tmi 1.dܬ%&OD 7=%/mzJy¶MO"b(0#[vlX7y60fvk K.KέI@Zբ?M~y(ƬXy7珽O{W~mSȵc i?|tGumB&fz5ߠ!_w!-B6,hlGjxeOUiף05_0| os=0:%b7W,yA(Vb^b=Pj(2"f# sSqv dlۍz/'*;kی{MI.Hx!llj[k.9)Jh)hP@hI[)UnLr0ݑ<R5D "׸1GAa.|,41 -2mah ~,m|Y],stB֞e1{s~+_nC7wܾo&#oIsܹqE/֌GQTYcڳ=c޵q\ٿB̗l4U[U:,l: aL3V|U$% 4#6os֭׹QcQ)-̌]UyqNj8Y"SQlK eL(8% 4ytG@ym;c^Ï*CLHIOӇt]DxƥRjjݓ嵡dU8S XjRZ{ YLe<Su:FPB׀t}ίQ崧']{L9DӾU njrk T˵1+1'_-"˛*C-{k@*(\IJ&Bfa~WmcT&*RJџW{&[-ϻm'EyKFsooNgwSW=I{IuIrIrIrI5w@/rRx%P+% KjI,T;άX AtJk6{z%-E[W>kF@ &"/Oi߽?-=uk|񹻎> o4HɹFz-#RvTr!PEC`Ɣ si䥓̔a-ZeHnfPzol9:\ 7aۆVXjZYl'xx2. aH$J/A`bB+"EsWNbZr`.{ҎڥfvY J3: "uF+ˤ8@øѥ3h CxyDPr"I ,bH sz֐)d9J]>OYc\cp tLF\TyrӀN0g~9vIْ+m ':Cȑ.ץxV•VNRʸxV݂Rt@ :)̚x0s$r@գCYnFGZ^3zW6QH-᳽frb {%A^rȷWHPw/Q{4L՚}. 5]/.4 If5?q \"IVn1e~7jP1~9A(B :n^۵z6'3Ef'*YHknV\Jk]V/RIչu_ݥO&}h'bV3 S_Ɖ6#+pW٤:}E6&֬Qu^/+ekvٻRtC)e {6]8tVsvQuF_#+>6X_?޵2{}kEfR8$+;r$M?}I" jG?bhW拥S15w?\]GTLI +*JD[^bq˭g?E 97=|bt ?DBXY^9bq䃅$axƅbȖBW8:c;^1~+s W5.Ձlԯ5&ɝ>$W3:ϚOtmTn{Ji՗58E^"\ߏ/-koƒžx#!1^>+N7`E!5_'בB^gP:u} zky, "\'`3jʴkx^=xkݹĤXp7_{M%\&7֬rh?7i\6f[ .ᐊ%;S n0JDE9I9I9I9:dU^i) V:4%%R 1MHɌI$mJݾpQih`&rX44:Koh]$uORfyyߓ'Q/-R f1#g7xUTW߲?~~S`6WIK!~$ #LաH3~uy<߾ŔM#BVsgf^F<,>^ f[$*jۣ8!c"䱹EMMW8nLԝ'Xoq9N0%a! b(,b FDSx]IHLR7"Eฆwn'S BckbqJo,j.˴ǀClhߑը7a!Hn:`XlRJG ج mT(3+ &^69aV[ [79a#.e4vDܢcwN6D[Ő!UTnˊg-*ܜ` !O*=v%ҢM;G@5^v ,E.Kd]D*^pEFa蛞#j, PNiŋvPs|]gqn7*]s>=|h8Cf3ȤY͞FF58C c竃V_Y>^1 1? 27z;PZZϨ۵q d:$J:蓨J wh3nfT i(|5>Կm]( G 죦O* /&ݵV"l8&G3#%!8>][WPdMqZ E>0#64Ne*QJ (1b z )w>/ ul9a_HZ و~y})(ϯI}&i_|I-:pŐŦ,61g\6qmΈ?9Gm̝,X98cDB(N+)#4H8r%pcA1ʹ E`PAq/:X-"dD8,C,f"薢pPWa|KvT@/ldf NXԽmvEdG lY|I3AWBpB:z!ՆSPQl$Mv֛j R=G*eg61cѕ͇?+tfB걓Ec FL$Ξ,Tʮk:wzFhF?G1*iH? OY(.J0PXF8؏ 6iTn];7$$1.3#5<>w1bR #ݾ;ZyƱ㒕me:{xJy6N-TFcjF!y^h:̱_~qKKӻގàWr&I"gșTYoxfpv1J|{S4N\I:8#?nz!1% 6t9֎K\~7Z"mAh4"xOg5Z{L]t!4 zV/2EsT˻|Sd{VOfag 9 sg /X6v +0w]s1Q6_m~R]*/\Ke]67+m BdrE~>8k];*Bό>w']gSwGɻ? yq{n-.? 8^cmLVZiSC|:I-_=R_]FŸ=ɀVzyIh~2oṨ]M><Ip."(wamپc'y*X"JWvK[}[K/KR^m mduzVLEj[ȳ[Iv5"\bDRat{{FOy7H4ugiy~coƢc@nkBUKͿ4_S1>[TZkߟ(&jt;\wrذ(ZJ(tR|_hFi]j_\ /KTl2{vXP!܋ uIWjo EQ徟 -R؊qǔ+>cr69PĈFHN{r985^5CQ!pRyƤ"Kg4p:"bNt Sf,JcU>F:ypƗH!lIP0^qJKZS0mO0#A!lJ/e,Rz$\ #vnYZXsYeRk"8ǘaalukT$T$:యH1My04Ȋ-⺤ڗqKJf ֔@#h9m/"b61W( #V؋i`_*cO##uPBjZ0b &TPQr$vnʡQN9Zbs M̡ĠF MO5?ײIH0Ft(53j},Z#$1^ m)snz) &ƍeayCWyܛq`FVpXF (#몟J,ҭk^K@VV;v1Mn[fe/i`NدՈ9iRFTĞ[Hs\:Me5:3*KYyux %;vCSzY޷X sh{Z"6qN "E듬{$NuE#w'r2k\}E!\*;ܵJ]TOTJEqㆁh)kqΙزƚAF1I1 $$؊ՄWYð-m5@xz }qqDio W 9J̷h۳L^-v C ;/pw0;^ԉQqsW6*5'^]\D)ԱW6L.yS2ؘ#{,݉J&v5?Znq۫&/.|HyM%q26tt76NHLp威fDE'ftB]BC%UIpZlIY, tgv49^~R8eoj6a-oERil:۞WfKö9פaS٬ρU ] L8Ɣ{cxJdrh=d,4y'*PK;F ~6̞Z[n'{oCTuRrߥтF;E_=UN1D!ÉrhCrlLڒ*3̴!" kaQC<#j}_>,uuں S[py4{y,] Q" (+@qv [ؐeU2e]-`#Y?\A2 8frgvkPW.o*0N"Mrt(ʾ g35'6fCΏNF{3!ˑS ?/f#hMO7S;n!QbG #Boݍ?_LgVO}o㈬N]}V<62~Ώ/g!JL= +nZ.Y2U6H՚v2At-)FciJ-Tֆp͒)kXn!h\NuGRݲMnmH 2%HTLk(hLxlΤw&C8{o )sa,?F?<:\ ^}#oh. qLOg>ͶSOxA4^>^^D0% ="`V:ƵWq9TR!> 0D_">t[7c8܅ތ&1Cl`?-Mٹ_=/\E_vԍD b.F } 5)^&#R`К b8CNc<4696IA Mi첤DDcoPJc`!!{OCYR,̧I eI7bݼ8#OChx(nܛg'ޝ'-P3$K #=î(Q[/%Ih=-t"F1q:zOel 5+Ԁ-B@I   ޖ$,jz"bG)gn"Dz[5Ͳ7W6EȢKǏK]HJ5^ qQ,/50L"Nr֗ $uI} Ԩ,Rb]iS+SJትA; -1w =' e:lJU*)sZڮZꖥi%T|kf0Gvpgy Q4~t$̉AeI6SqhGB,b&HS8cVP;r+Q+amGVI,!` .IYϸ) q΂4iK8{yañŠ3&4,`NvN)1SblI+H&)C8Š+AdQg)#9  YzB<iC9^Se@QXe)%T)z /lE8>*ְKBZ:^A(ٜѰDVn. dIy8[wa, ־KS_rFH\$^boVi((o%¶$"'i5"<)ڵ G٘:`^N=rwjQ ;'x1|{+WW?]mfdL@h3cОl)?]^^\[\!f_w.tR͒t#`|a$=:IMP"m!tRȐ% .n)%0o FX48:DI4RQir@gtF`A k.6A97JUPkUid)0 Gxx04WƞŢdp(r io'fz? Jkqˆ_hj>=PdM+*O7ߕyGvUTxݧM ;~G埲hRV] }+r.j##$'+ЖCP()SR0Yl3hRrep:|P $A[JAa׵"Lr=l6&u5ٓ]0: i3SX!ݑ|oCHh]!* ]~fvp7o]iK:'W,wG\_яwg׋ahp7wd Nk%_5:^B4\aw>ީрi)$BΥ U7?ݻ|o9!JDXP ӹE21e?2noϏt<+*?U:߫UYSs]9~5mdѓ9֒ Йs ՟CAh 99mJmYf7/~)۲Po \uM=bQXBikPҪ.fýPPJwYZ¤¯{<i1!7kLHrp{4lm+=NzxiGX uAg6W 6{3m٠$G of)ġ+jrMJVt. 9- G3[`ף\$}]rir# Fż-3:GMH=SZ(4/0HTJjx*&gk1oڒq>1v `_*ÇH zDjZ0bv C P5^Ct]6ИQ:> !;oո'MD!D CtY9˴Wb~uoWZdQمL{LZQoQ">_Nwȼe)ڹ?TVЏO,洃w7'=o~\9__tu`я"X<;7VGT l[y\\\cj~RwfĪA) fKgġp-W;]@cwgc"•EP|9#N4q&ATJޝV3)| o OZdP^J8Ÿ^P]dN/ήgPXRp/ ƈDț1hU;-ۧ24SD/Qٻ6ndWX| wU=s*Uu*N6/I0R[HJ?w J)WlILF BT  DkARV@THQQ+V- ZUnhs;xfRV@DY瞝#^)UY XB%P R~ ː\{Pr\WӶGNB 7RFQV X΂ZX}Y̧ #V՛O)Uz9ƵuB+ʫW$Z%BѴb:u ƋuښQ {iZy8. yp%O˽݅}L:W!me}!;UhUTHkㄡNJɵ%udn2*6ZEsQV[Ub!g!L! d;8;6H1 6)"&cJTc!+ :|%x ,e|et8k+LϿ7XmgܒQnjc/ o:6dOuz HJ/5y^>G :w tU"=X_AӚw*@6MT^_z_VBR>aXĀcm|XMMy IuzM @5 XVKdӆXQ FYI-3f˴MB83)AN{{ La3TaX7vw߾oL Lݾ[ |}1Ivaȴ=y}A]jXC2G&n|_Ku6xj\? ~vC~0Fwܳ< 8ǓD, шy:=/{I҃7T5[w}L9i^9(JTBN PWVBѾ%0CV1Se[/r6*/4V!!.UdJ; ڍ#vʃѩ2ڭccԤ*/4V!!.flNZ%cr7R !tLOVFJ0RP(Ix03|W7>8,j䨖J@@f(yjfџ! u3\;!]n b)Bn1i'=1?߅Br r> / K8__e:dOL^2yd('xt(!V;ځVe{}TTY9'Jrʝ!G u6xy`h f4B%fyɥ]\e_0{b4^'1:jEꇩRZK|sw0nHj%f<-|gv c3T!Qq/[ϝ0E"=-p_/E%&b KI .Rr))hibVhk/`5OjmtlYEb˪/o)A0p$+= 4d M&$#6SqJX)u+C@p)r,YVT(REo跙@4o1gdSsWLO9'ޏ?ȻO?bS|3ބ&g?\yoχ S@?}8Ofx|Y;!1 ^~{ps0,DHw&b|8׳랗qoyRk#R }2eYaaǽpV<8W7jwC#hX1x. 3+ÎT8{8U|'(Nwx%IZ`gq* R&F{ ,Sp\C:" ՆOw#h)8B6jj柃oWbz2Hfo[z'8oFz|z\C977h sie<_\.W$(^d<ϿkF'<ڋ&D3 %(*W?W\;_ Dtr -ڳ rn r*N"!$2jFP)&oEo^8 Q!-(`[OTf碶L􆧕"2yKq}@ 1>31zӳ0qq6&ñc|d2Ƨ,\>RfG)@;S]sM ʫ8){lNUR%P_+-R­m'KDG7I4z /8Mk8B$2Ęƙ̹$:t^'CB4PW:7\rO(ê<5*s٘ TGac)=WWU H&N>*e>MԽ[#V3,KLǁݙ Kwz:_fM_ LpEHjد)\4?f Ѵ8Ithf޲F#6x" ʋލDɉIa;hMɬVB!,JB{̣;T؇`DQ[o1;}|l̚w8oShwx(m 8nHpiX{AR"{Q'WJ(bFD <12u{\LY%~i(K j" H"2}%Ҍd9u{7 p2KuT`ĈL<$Iq (+$S*WPlE߼KWA0,ՌP۷~ wV2Tkz{>#7Z{$3@,S;N\PR7 %N8R-:pʐ&5EB%3gx=LFqP>ao/hqѤl1Q9;5 I-0ƒ}͉[sEe3+P͑Z"gTe*Nh;, #+1Nr][^;*9hqVǭ!|NQϧ3 "d 63o(O pBPJ),ݐx GQ[->d*ZP1z)oOFJh&M8I#fJMM" 9p +Jx8s0/˴I^*%/ 1]/| T{Iro̠Œ@r8^%f}WY5Rmܾ h/O_Jּc¸EGSy~AekqwxuXpcB5ؗ c{?&O入L؃jÖrGך_{Mg9Fu= Вx՚e v sk^rārE龩k:X7gɽ0X}aaypBW %DousIm1ٌB18\,KAsz{Tt*=980;C7'6Z1|,*p8VlmÁY%Pi9bmhX}7;:͑@uo cLH?zn}s9d-m מG\"B5ca1QgJ*%y%U3Gx) cU 78e3 W?#qTRŶV)L.C/Fo-]F/bLv2q ))i0/ 繑y_1eơeXx!tjߦ!hwWO'DeI Lk@S݅pVj(kz$HQuPTp2DI kH߿l`,١ty;<6:10L9fړʱ1RZaR9$Z5Q,ň!'qy`TGomSW=ѷ!ѣIoR4ޜmmPW|;X@ 6db@ @)AE1*2uQD8bi?-Tc}?\a~|P Ozb.gY?߅BB r}u^>__e: 'J.BYE(+e%ŲKSV@2eMF%MUFSp$sTIcg10/ৰ9_gw680OYA >La<`buz6ۼV$`ri /t5\e_R wKT:=Loޞ4RkZbg41KYX ^A;yk%5Ýfa:=߂0T[o';E_|q rImHS2iJ A"ʹ9(eqjJ+kmIEb}" {L&/0k-Xrf栗%JlRMā8n~U]Uuw?sQ!ҼsN`51Rq*#SĭA851/Pd,w #H+?5r %Q #DUt[7ŀ#+aYP`/Y*C~, 9Gda1)z7r86Akj'Sl`JVҠ ɂORʬy: Krvj` 1zޟm} L-Pa4!0s̘F7JpY0J2Hai yy-%t(G̲ vDE2E{Я%eC p~8勐XČpKǿ|9.Ft*xIY 0>ssw~BcDe"駷'@g8><\_9?o!)smf'=K~|N%q`pYG7I)$,(}ıLpْ1É+T4VrM>Bzɧ Ӂ \4 +)qRce"#'QC46u D S_[Q8>bu ۵A&e o9qL ]\/ FPwFpkd3(`yH4e:„0A Lz:"WXu^zޒZ)C`R*`) ,vQ+19XqvB7bɡW 7n5Qof.Z%#B2X HN#U%GUnutA@ ӊ{5Q5hLV8>Т`|4հUe 0z5>s!δPr6x IV0ו#>Moswt"T3lp}+;TR=)_hӟw:mC>fy_4.8Ŵ΂.vOG>OO^G\%!IjtYս OsJe<]VW.Ȋ>=(ՄKtengi1̴~8eERqhz }Je _&.|0+,p88nSoU%N1LpC eіUpcA^h'C+uPT TwgMi}=x2BO BxOC\*1`M3Q"{&`sP| m]z.>^Ucʦ8x 'a_AFh " O7BK zThZTQ!k5:ȼHH|$N @I7'0}$s}`2Gi϶4:^ǺW3'sH KӠgsg2ܳ/>{Frv1*U\0oRfXZDnmR7^z$g!"mrvt*WPmkEH\Yu~>d:Ss=}1R: J}eS;Es+U'kro?B;+(hp,H:61،J6R;£Rĭ[ДQyXqhP}Z?uvRJ#M2+8G$0GXvX9eS YjH&ߑt~!Qqsc!25>܏D Ƹ?)lͿ?Sk(iH!W UO+DU816gM`#UKg1}J1v! ѯfsI)nޠQ+g5w߱ ymߌ L ,HOͺ3+pĆ$q1e*Pa`|{=77daЛ2S Cl+gP a=g~3L|v+T{c[Pk{9m0ځ!DPA@1`K##;a't.H)(cu*" DpY+* `Br 1b&IcjGA,Qϼ&`HrVDw'F[OYX׃m1Ȓfer}i kF #hfOshscNr dpչE҃RZ͑\${sv;#0<1Ť"Op `ū|(䜢 `%Q؀-4X;9w,2Oe6KĖ0[錳LB{R5oyŸ7Fd^JT c8Y^H;)e`UZǭ~ۣN1S8x`Lie+x~Z`(2P>I3EPj&iSRJ$"IH`}]I~y(m|/zbϖMU~l//*i])fqz/j6Vp5Q^u2\bCZ⬃re%fjy`r>dxH+ OU#ݏKLu.:}H*I%joX\ݻRVhi7ѵ4__Z*oFÒVK=ʮ"1t>E%fȗ(~ !O9z] 30Qyl+,Z+Aa58d )QV.$aoHicTD)2VX``4 ÀXzOSYj3j0 F,SJ.Ӭ٠V UYjG{k8k^[+TF8NUXzZ%u,k2 y uf4 Zkk@l- (u4@ekWPjj Y]윤R*=RٍJPvA{x09xa@Vdzd+oMf.Sz9PQ2m\_^^=R-6Q֏}x9):Bh"xIY񈷀d'>\#2.Wѧޞt?/򇇂4~\zS6٬=Kx~|N%wj}JƠ,GJs3:؁z˒HP5yfd6PAT*B딃A"wmmz:{L &;f'Ȳǒ2/)uid_ǢůK}U(bJi8S֥gp#Lqlߎ-rxf<̺U֛qC>biݛ{S t ^SEȚ!llR۩xz\F1f#$kX<:m.u- lA]ЛYuaF~&:_eY)m3r{m PG DtlKDe!`uɔO_|m`JhcV0X¡XgDB{ !Vr91zc&~uwgDž;7\Jgx~y|M$l#z\#' ]d&K;#i 1@gwP)YV D򧋶,rgCZ])W9qRnK* )^O b̺1uKUխA ~ xٽ}j~ej;ZWiXik6| +L[Zux_C|U^W S5z[imݼxY4y]Q J BLvT] ":_졊lXec . fCac w;R  : ԇ-l-l,)RNΕ] sNq8/}@eȚ-q'uC`hp#rպs#eF)ʎX4 XSd[,uTh`X*lD[ڥӖ ꅫ*5Wվs-V ҵF(Ibʼ&ac-p8XFD$mq%=6YBj~Wé}c;B1"{W5}>u'֝'~US7JEaFd J QT`(EO9g)દEU5JHC4Y}%rkxmijڬ[ަGLN] -YZJ^[Za&~+#5kőlzbpBy3v=ͥa]i1NH[F2D2%X+0e.iJFauX46ߺPSЁzҗ{M }!k<)skfbuTGZ 8o-E#y5؆nGqgn:7OS5fuPo$6w%#y,CMjLv8c1!b:Nyw {M|W),+7(bkZѻbb:}nt"y@Br)I-w2_o@ ~ߗst.sB7A NyiQʘ)e 47{ ^rPȣSI!8N qTq2k!3PUq]oo)3Q_t`\wA( ȭ^܅9$ 'uNdAPGDns{ Z._T9:D9&wڼ981$С0FѬmKp"Ք`A7ߛKİV6;7cuSmۗq1Psވ1|ەmΖE̝TǏ˟#GbT4ˣ_VOדzwuQ$|cyB ]{)O{VZLgjKfDMJM~\P3}뗘Wp;|ME A_/2⚫%uvow7lF6#݁d9T0n2ʁ̈E&4,RiUP{ӽG^/Ox\_St%Q, YJ suBaSJ3T]+Wc~9#-6PW-į;qsq뢎[uܺ֏:|`e]L.Үq! B\BjxNi[GA1%A`؎b;olIx=*"]ee!)K ywfa \rTFADh 0B [;%LR@!2-e$Vs !;1%hKY@2iC4PIő(0uJGV#;cN}dK$y) o}|DOaju0~~A=i7>Ĩo]P.ǿi6__,7oS-xgfYҡyX#b:흀dsSJ r͟W-`1^m~g= c'gWt^*9[/Xt;:;s/$u2]>faʛY/C r<07a*yz=IDY{S@h:jPR8:؈j,$sc'*Guv0&DESZ;MiF16$\ *螁F@ҒE) H P)`I1tĮ fr6wR HADluH:%^ Oⷶ ORXWn16E:^Cɻލ~[,!G<$9E4ԻD۔]3j_$&tmI 8!lprSpK"&(2BZy^`-Bq,5PQ!A4ۿ~Y 0F%gD 05%G(g((.hHp!@#da,0DCP'V!Uϟ8e>(;DWvvATPIꝃ I) aDAL)u^UwOR!2)a}R CPiccRu`fVSV$/f`--̮0&b}FQ/Ǯ׷Z%JƼ A_KZBHk]>ߗ9ĝLK.Z+yYp C$x7,aDyud'Bp$CY͙SsԜ95gwrKfX!PgjӃ6w- BgJ" Zp1ra% > I/H=B +l\YD ] .0g2_ފ '|?UL%%B\>D9'e{\AxW1\ȡ iwyA _^g-0֙ACWbQL%uuS]$:ڢBai+A~k'k?\@yq@$zd@w:c $&AyMKPveB-sK= PpymBR& H9Q@RsQ\(i-iQKRj%T ..2XFFI- 6$…'S6ISs8 T&wH$} 6uh(׻]$<7#z)L<f  32,W;9jNHGcpk$א2p;i|!P_&اӆ`!3oK2QbH&yV 8=2WߒXU< U~>d= {Efʡ%rsLe8^r/s.B= {Wȍ/ܮdW!  2=Vbˎ$';bK(b_(duh9:.b\ukL^Q ]lidDD@#..JV{skyҺK/K.å2jFDbD3IT 玦f33cbl2GM8qwպ6f?\t *)F!7ɡS(BC\<*gGWE7j mu\{5m5]ӬA? e)wA$ʠbAJo@a\t\~L Kd|z\4 V)x'BZ`e6{&6R:6GOߦ|4ڒ J[j82=OrBVK$k*01B{;Z6 Iw}5S1Kg$L̗gY]f?к+vls̲% K&lxl/>3 3 &/ L+b,#39I:w=DJH,O3N!.OQ"`dS9jҌfV3ZYt䃚 ˆ/iU 6OJLhJsĒhLWeJs,э e'rx(AΒ|b9 of+p(4L2 \.`*{B*BK35ȕJ}3ĪHb}+3-tkM9B d94,BIEO-])=)7(ock! `$q6;ZޘȴYJ'FLv"g&DרߊØMŝ+7="Ґ*"Zvתa). f|rQ.C9dŠ&Gզ5VA%>e7/{_CHšCLƩvLJaq>=>X$ }du m NS%tvSGlr3L6@ R}](?~B^fd*F5Zl@f5%G4SE)^ost7n#;+@H-eG6u#fZ?nw2o/3g*rK-MFMsF|RSn *3"5cSP &J/ BʁUFqڑ2)dF;p*g.v93# *0V,cR4Nnձ_LRи0O[R`K -KApsʌ=AGiJh OuJ@\1H"\B2tJ9 Oud( [d^.ěR6LpJ K,cVJZXpQ9׌i}6 "q$JښD=M@ЦI(XbpڷiܟE 4*׎P[Į-ih4mWZ& )$g3D(Vy*~ӄ&Uow̍ތ~:q-R!;޹4],QppytbW)_^='//tGJ8utRd:;4wwȷ3?_+ĊABkN+ a&,A#%w7Lvn|&oΞHH  Z.|G"aF|s)Ȅ&Hg՘3E6.fߜyCIFV0 ʨiʴ+'V  rbQY!:m,$%5zj꫺R^T1n6A @s9g9#*1)L*Jz\L3S99sT4DKns) -D ݡ|l ٟMT49a V,30A- z,c!Y%: bjƳ;|rFu/i% 볃Ǥ ~f`fC`"Bw )Ng 0#x1[1 ~G=^k=ljW]8c|%4RՆ(PKuxS!RTWmZ]}z1N#;SS`au,400TO[8lX: n%- CVEsU*BZhd'&kruxRUV3ʎP3NQ~l N:D4/wsNѦ:t }SL 0Rt  pN. !H4S@s\fMőӺTj0OCDnӷH:f;%h@;D)r*"w:d'dMzwycg*S{wsK[oǵYô VKɯ]>1wmfw:XDmVW5oo<;߁4<mB>>Ζn>$J2mnCn ;φ!IٙhzZJ6u?8AuGu˔_dz'`(MOϾrzuB$,(r!SI+^ĩq#.IqY; c^$4!Iw$u`|`IElbV& 6"`@[PjNjafyQ>ׂfϵP@2r2kA^g2ߧglM ZDJxZh#/w52Gz~P]ʄwhiދvo:О\r  lotBdŝ pԞPϘvPi NQ',PF?|, ^V1auEit%pqGh羏xј/Ř cMƃ.AzhK}\ų zҡVW]ojRhm(q|G. Z,Kf3aTB=*,ЏdF!u4YyrEnq|s橻{ŷn9ry8?'&tBJ]fnyrwi8zfˉ+QIsBF;]d |_!DK6Hjyԑųo[6/owZ5_2֫b#,Y nZ('V SL~͆>mMܚrg9|FŕiZc|}ZL#'o%o(& pv׷RWzn0Gk1"y0p\,ےj[ *#c\{Q#l5Wa{q- 1 t(·ރCZ0w35A;gWRzEqET_/5f{w_OF3QEdtyVov^ 8Z8=7=U-y[ `ZܛiJA/yw)/9_!Ѫ|Kapbwg}s?{W8J 7_dUip8a~A@T'3'vwb(eS֋zbR)^Sę>][{8#1~G僡@$;dUq[i'L\@FVeU:hL!f:9#;9de$ifQdDsd?/Jݶ'ҁG ܑr}ܔ+dkK!77.]Nf,Js*Òh ~ ŕkG,csSVҦ=c r <'2T+`oWeܴ)'8&P 8@9e=O+gӯvU6q N9k?1jrm\^ .Gp,[kQ{YwbJ>Յ!Njͷ@b`ڋ@8 &MFf7E3ւ쫏*ͮ4캪Ϯ:j^Ycq)* `1.,9[yO0}:v}PSW|a[rh%1)^]'R/fܡ}tI>颂" )Ɲ/R ,X*JbY6wpY%I@ M\VOsYEO4Y.B%VSfQ`􁭨 ]}S$0a\af!GFP˹u0rYtex.'nkeF뮓D,! u@B);< - BB#s2Oie&NS_oa8t^nVFB YGGSoy`tA0hEg  MUpp\UQbDm=6P{ޥkhm.;놘ө ӽ_?sVjVqJ1*z(ekfg4jR-wg)kMWK}"RwdV\6H睧#@xlAOx0;V܈ncڣ-4c *6p%Q/2\t#?ŋ 33FNJw˙5' Fj \XA=i%*QqBlp,7+HݎU:,Rv!|.2^([ee)M&(Ƒ#PcʈR nc6Č3P;,hfDPy3ƍ r2BJ+ Q4JIq'f%*cUnc hf9z2+fYYm}z.9йGVN=޿x5#d5܀] lȽ *oȪ}ftW\$a~vAanpRfvE5S#R* ^v㎛"hJܴ K*j \f'W35. &74/#۾̣kjyΘQI5Wy7pĆ/#lI!f2#"11pvж4`71@ewt;lϯ_Z g)Cyn&O"`Hb~<%ЃY#5r^\κsy}dOC<̷F=%goWWbFM=w6iw>;o-zh00i02+=cqRwڑTgL bw5 L}Wm=k @ne3dtʜ PTFX2u)2,)c!}'0Pf攩ϩ~ڪu(PD$sJ{egxuicʿC+#ͻNʨq}B]KN.vC騕F}:n H}w``+9AtaQ,[;{NwT +闯&ɤgDO+)PHOXVc[QAxJ[7Y.ZU];t4{#2*v^I4{OwF$I|( ^ӱ*][, 4ՋyRZ60yEkrs[\X`:F?/z~9$'s81ÞnB#ҠB/ԇEgaJB/i8 RT AKK~݊?h dZK/B, 2^lg6yg1t.ʘXJF=-FclKf;Bp(3Tnҏy6cΣ;~p,w?\U5^ ߌ P: \aTZZeMWKCNE-Q"3I/CmSc۔_و@n0 ݭS8snȹ{7("UCź=]{ ߘq3 14R%9JG蜏tΫ׼yW6\:}9(0.8ݤWqQgz~}ki!9eڨFO<#Nm.?f?=b+l'^a0Ft'ky;w[8|) |agP8}Ah&QwTG)ڱ-!ɹ4gT@@Lb>?W-T!ʒʻT\῾ûs1и[ۧPٯ,^VE?ggB =7- gWm&K?DXr'1x!/w!N1A˙V۩v*gamN U2Gs.3H8lƬ LxmJdO XrSZw#jIUڨ ILa WM— \^`B`HKE`)jH\}>:o d E)20\KͬzJgS̩;4W1U,5SgGQG*8 󯷋Uwr]|?Nz2=OǧcoxĀ}"u|܃ Mο&Y:out:noMe$k65^ jFgZ -&"ًZ'w{:=R 1} c\Ġ'4q@\)w nb;HAyb E)ٱ]p "YFbl0b0(]`*D:ZIr*Aj@SbZvMӗ߫rxģB"4YtXZQqƐܵ >oOĽ=v~pw_+9լxtg/v8sKӎf'XH10` ;ZjɸxJWkQ;M Rbn,杤$x(>uYH!\KI%amUjt8W!1"HL rd8`c )bZl_2p$bAjB;{[7%TLLmڰ(֥]EuMpM9ZN%pX?m%B< Q+QcAnr;lUJe𲌙T?bwe4,,+Ƣ=CЂQɼGxť|I4&Y #|ֺ41] ?TxC:F @cnEC SDmh47$+@VMDjyH \d NҖe%0aEU HP,vj s*;JW_: `F>^v4,DV Nj0oD[@4 ېنťi9 E@eT44@U(r*(H@K411iAqüF՟K9{y 'y +v% kǰRP;6TсJUk|A:8Y:#?>痎1,3mRdD1~*~tJV\sߝzO h[ttiM]LGlOL֋oJԱ%OGOS~cnq!3ՖYx6W1&A9qI s$d =7 C弝qxPk9TȋSAj<!řj1UgGOu<h'D8p VkU3ܱs׽_'>=pHX /Y1u$o|o,M3 S r,Fgz ɺĿ2Z>l&`z{YػFu~(|/A SǞ8w]Fz͵5{+{˦~h`έx]# 쐼h7"^@VBۈ28}W0˶_6MĘ'ͬ_!&%;un&b7.f*3ކ)%9LiP-$10g4}"}CDTp4sM?$,_Hۅ^c'}`v-oo\x(|E7Eq]YwfUuWYCt^JHQT*A:C.3JޕqdBe0K˵/0A03A%kȒ${&$YͦIǰcu;S`E(hFяq>隸DoH܁^M;bOƱt_>sY1d7 ,ǮUŭ)1Ƈ\.m["p)%&FL`BV<`X`q'X*qlDa'CYқc>ʸ~_z3ֵkc_})mۗ%B98]F[<4 jItg5T91Lj3dL.zXSf($5 b+ж ޲Hqs}U:t?+tN6eEd%aF>Hl bT%պyr aj !a%GRI>^YSJXy**CQJ!\&h^ZC lzN[J;m9D ݺD3I):Rpr0j1m{p?]a€/mEV-$9M&su0"R&IO6Ar(~p `m\<fwIRq0IomxAͯJf [֌q R' ذxYDvl{|0ڂޟ9ן{fY3*ޓnO1{69 #3zI"~\ő>;P A9 q#qD ٩lzʂNJLв'tKlFi!"d0QqT0r`>Eh>zÑ&`K(A﷋`~;SݬiDP,;tEtHc{GTo߄8鎺r-W/oKOuDE[ɅǶl VRt괴 K>{lvu-\s<.$lr 9I:ٌj+~u@"\!/./9eɘB"fsA&J[# g%UNeׯ.sE+űjي/uKeғ*RG=tIW~޿?z)/)o\3)kΣK}05i&p gd(Ks_(W05sc_Zr<"[V1׵+|lwM;R?+H3q4sjkX!p 7|>fpcʣ9!gh=77G=}sfњ)m4?<Pyܙ/B$Xz5c X#3dE$WcDv㇛mwCVWL:Lv2罳qMLMmqPZ?;'TZ((QmVc5:X+5.54lVrf%/^šG{ga O {~fz,nAz}"1\{z+FJQpLS48!:¥"{9Ci&,׫+ 4g%( U z}UlGT#v k & (zpiuKJ0X2(- "Į^8ʭF5Mo?'V nݎK}0oA,7"Rjl)(`)FZ2y_w&oĒڦ`MmQ# JÎrX *4XN%&TKl>b8t3yQ D i24mof{dgw:WWͿ?:5cXv璧۝lN^ :uoLL$Ve)zT:1$t?Y5%qj/Ӄ,oBs%>zA( Yu#}P_Èʾ=3bZ&c{dESkQ 3ej&UFT{:Z#+畻aoU~Pj̎3oΟ^m@dSɵTaWD%UKHW*+>* 9|)q_>wQ~ݚ[<{,.=먾dshVվ-<'q?^ !+VQMGo/~l&2&\[|ڥpW~7a{+CR_ŌU,2c<# уx!\mѶƵ&JiCfl~|~%5|b}' 0 1_ȋR 8ܯpͪIf"P9þ#VAy_u$rFt-CYǍR]=Ҕ0B O2''I“ɮ;ReC֘mmpVXe)\Z3*emͿi~g}CQLa CfT1bݿ)ܨ9,|8{Hw4Iqښ1zljx)EqV>h&'&6Nekc9$ltK'I6Or0^`Uˋ72xJE _N9|svb)a߀:=S*؈/w̗S$ ѕ>ݛ6v`Lᕏ(oݧ?9 d9j;Hg*DǷ*2uU8j -Ѷbz;??0m+GSږS'w?Ǘ`IU˸iC`[f ph۶D%/hTLhMo]O^v _J#6p N1rqϗPofmw~`Gէ?D Pڹd OSƑZi |9i5{%d {Iݕg/>W|E /cL^VohiI6N> &'αBQt`(/}zdqP) 1' A)gqN5DʁtnW#TTzn]߾er~֘f *.qhdʫؽݏg;v;;?hMF[}vIl*}৳j_W_όWl*7 wqB/-Xzu?OW$JA4KyCL[ZU>z8)#6gwPϾ+c&=3Q罟TXy?v$E4K*+C+֍JW=Z\ RD'w&m~J[DS[r"%S{Y{SvQ*#Q-)xJu{,Z(&BX)F ,#q,?=Tֆ,Ss\x)l33-чVNSUppud ᒕBpR.H+ WBH6(ALPr%QpQN %%B+#鹿:\\|kBgՓzލo2s9|ߛWK4.*N")Y?c'qy2.^˸<{Y]m Lb)% )@k('QEk* ?Ns`*02|MW ,D4hQbK6.iny&rك/ɇNc}.ڌlHq[y`S-TZ #%, $ >KM\3LU0)-C\B(r4AN29`9E!$zEuD*)L#+)QjrSzF\xDŗARB Km!K#N<(]98*]?mܔ|I@._xbH= ;_xgf: rPQ^p䶅DDW*1wFex[~#0LƒS~߾ ́> ygjJ$璘x5}R#~8#yxWF ~27&^UrAҳuy?j~Ų2h gb}+IQ w;ۇxWXtS i9jPWPi6OֽSو]NJY˜,~=,y_E_M(Eۙ6L8pW@=ࢣaH{XY C(+ 33${lX[#>dLüP]30~IL10$"r_kHCB$&hˇRQ f惆H)B$@V8\~Y]$'""Ǩ  WrjMLX~bRC\O*ż`62Rw5! C4%b*=:$z#7RCsH)EDs %9A03]A8H rDqpnKT0$3CpڬGl.ki}6JFwJSC)| Atx1 g}|6|d^@_302fs6gvqnvL|Ρs]Uc9(s %Uƙ EnR﮳BaXg ΌJ1M7͏ּ PcfHTX~N9bkb6B³ F76ҁuҦFPD@1I, D L'i,o ABQ(RLrd`BN7Dgӕ.HodCÜ!SLkmH0_R\n v\iAlȒdWMIt71jտꮮG&Iʹa N0ׯ>E0`"Q<+r$2 #a U2U{$y| uV2i8Q*(*K,%dh82fVfYb>8:[U^&K(}h 8Gݔszw( Q+X N g'8ڰw㉇: sbsQ,˜xղ 6R2͈Q fB"`ye`Z(%9TQ [{SH|6*f|N|l> +|gJ {8zKӂq۶֏ncUg^wwԮLO&d! fTGdY~Qd|9'\-<2!PhWDRA -븄Ѯ+{@;B}Sa%¤K\+k3Prk\aPX[PKp+1h;"1ǽ Zx-!*a*TUj9[2!DLv jPݜ5MfG29.ξgJzиԷfjyܔ0CZK%f2Xfef7bpr0ɬHYtw΄u6;ԙT%.fYex0<5 \'FYBK7ȪO]\'%gu^HO^ vRH5+rE9ArBH_><"$AWnL`}rG; :aH:'l[8ɇP71:>y>2h2OG~|/";g 'D6ψ%#+fl2M9b ;sg/\o/v8qOz"!w7vxqDKL&5 殔K]?';O&i>|zoK+W_~cc, HFl°%6%iQlTnNμc[,>9ٕy4.xy1_`Zg.fN_[FoR-7Mcm7y*>fza=i' rZvV9HJ. yu#ݟ^%e;?Jةf]} ,.3=A;-!)X)@S$ r\LW!zE]NupOKC ^hsM#lMHk-> soD[a8]sz.NV zAb"[ |4DárL'Z \)<\ ғD9DtoqB.FdP0ekF /#<VC$1=}D8a ލSR3G[v{ڨsEY%}svfjv.՜A3 -vrs}rCW(o4b%4=9%䚷.M{>sy6=}ڐH$Z?8ԣ:: 6+UC,pHƕ:PjT[gG~|۠=皗}ϣZbӒ<[|pE|~3er'Vu%wA9{{pJ@^=s[;r} /Ѩ5uOc!b1Y:nΤ[| tkbhԚts-vM+BKޝtOnM?(U짵k}Q9ӟ'ָճzh3pC賙&ty9Bͻ&1*K1WRcW5U '9Nyb&I6Yq HjraSk I-9!E)]i3,`B0+Yc*#@ШQx"J)sSx*\&9M"… Q`fȂ8:A1u>쨻#%y:=j(:a^ya_-GulܥrJ֎A+2 <32)Pdi 8qj@ʘ̈(S Ec $Z~*vjO2?#*Q^wQ#x/@A.ASWtJ a+,BϿ?40MμcJ 9|2 FBx^*Kvb^w=SO3 t||qEfz itOLm/xM>JGX)b2=i]PggȅG W/n 3o6ƒ}aiaSSԝ5­bRŤumh@g Ktz"8HnB3vsGw~& n.-Dkc}SiwP(߮n3z\SRdErG5%0ul:[:ݛ ^VW2ZB̈́N=g`Op-"ӜPe4VPVie)ΊLKBʥ Q8a#6MjO3(`u&D%N1έBh%4c  drKiUpP)W!KYXJ|sE P!z/^ZFz)Rw9\?צ 8nY[bˆn+A"u=An://R^L!,qI^IOh`p՜sjSGљcQIx@5 aj׈kӑ}\@0V]77*.Vjzl|\K@[.7Q))p y@ČrOgfjy?џgH>62{=`R?g t dX 9ôRc>..J$d΍l ;NLţHZS7r[fT3mw 'w9D߀MF떑ZOfE <|o8Q9$dX `O#0OxnM,׉EP"2Rō#`'|h5 c&{ڒ=g1O#ję_`Rl ,)#47 gdg9)m]9Myq\- 7k":&K@t#6?}6rnqxC!F*+o{= Y ݉CZRφĔ$5uj83$Oua ӣ!$*5nr.8Uc)+f8)Q9ZcGJ $+eհO $XSIaH0D.ip^]}pƎ8* W~Ï`mQo>Gx|},6R;y$]9gԤاk_Ƽҗ2tqv݌y/x3=) Nv^*Z "e`<8e^[+Bjx ";=W[L7/Gwd4Q fe&cjd&vȖ2CXH"b!;hh,I xF^|ۑеlL`}6h6,/IOԵ&8Rew QKpWIq7A6N6y`@ؾYDD ^ c.ҙ˿YT-2Ȩ-(E+0M ҺirTlTZC5zpO3laQ5n,ZpuG VT|7(6ɆH?ůX{ >;$o{e# <QWKѳS \1lO3*EݯʓY~͂fqZ1,m~Ut^-h}ǡs>ْʵRU}?~|uGDΐVx0%oZ t5gnlլY<5 ).VV=F-`RV>_f:"ϣ5`Q{{ݟWHKL2HX"W\Jy gJœrカTw7v08-6&{o,Ny`Q<]<{&wΉ‹a)ntsKJD+AXOc9[P}^(^lœ[OsW"nrtJ&q,$ 6'7Ě4Eq1QyG%ؾ|/h/A^8 ژ@1v,3.4x}|%,v֋I#5w=ĉW Pw8L'Ob˙wZ1JװB^WbI|x?2cEZBMsfyQe›t-7EF]pnHݖvym ۴Xi23Mi}Ct7w;Z(1S3 *j'+y3!IIIOęx`4&UJ!a'E. IrƉ0%r'R`Z(Dqmd\dIRo AHM>76F"M%"z"wDzG,eFm$:lLIʈ{(ي=]SX(H<0+/AZ#DLJ~/WZAPC hU͔h'6Z/+Fڐ4ѷIfn,A -Kx>U㘠\A6BN*.zEXUy^*J Clk6g2RO"lCҥVI _o͢պ&Nb9sY8i7D$*ίp? U7I3ƙi|qp粤V/d_\> ܼ`cޣTP"s9nI?nPu:\)U H)@ Maݲ z17+ΈE\b SLOgjf.u"Yy#z>k:K1 4yv}t9͸Cg̸ϡI^Gz6g538RY{A*!VdOFeEg#J[ (ѓQwD ^K FEmq$ SE )/G9w5Ց~*?)uӢfӃ8[ P۷O~C І>J_eA+|g hUP b9{eTӟ0f+R9Nx3hC&hHQr_:D+̘$T\ ~u-~3k^bt#<%(ΰV6VBHYq T]hz; ~ Sc2jDn}|r(tƓ͟Th7G9Ѩ9JI&s &2#cT2>K4xAPE5Gcϻ)Wx|7w Et5pO9,NGCctC\|tZ$P>_Α/O({0u+C t̓Uu/0'H !H.w LO@e0~:T*Whkw*d@8KY6(( Q@Vi..> u4c;VD6Þd14[JRmɹ_2ܲ /vQ'R;段,=%L}\pۈl.ҽj:@f@\6 _i٣p3JR_UAq""e۸yK褈H3J^7=q5khed`yZ&E,#Tkpa_DpEqu3X/yyKN{%x~Ko2Q@N-Pm^_cx;ήıU!jHк-˧BB8@DYRD!WLm6kw7_ּ,:Enfdt*(2sbihAM#;OAi$]j95N3{6GLov%GD¦l>?Rn+[4Bg78}s߮X}AGw ̲v[^|Q7LMhU K6]NPTr&m;.6wSUʅ.>{.֔X\H{/cB`ܩod*PF@5alS^K"zg $D[`-R2 * cUȘMebN#J#|}1o?upT?.afv8#$\R9+'4L\GdK=:N$hj)63FlʘQL%j~t%H_FhBj!nJyR'^߂l1hQ:TWgʴt/&ٝtC:d_CNu^2!ǃ0go 6}9{iܠ8m$!N]W}JmqX8ONxŽ 3*b l&%lߖz榌Yo N?b+fY'M(҆1NQL1 QygܘIL5ƛ16`u AMN!C6SLMS75'gb~O 2ae'`Z }-L?CkA6 ]^L*}ֶܐ^jUFPy4)-:+ ef}p> bȱWA&H~<]_KDi99&ZJdCLmv|WpX_[W3 9y O;| !4vҤxڹEqg ~:ͨ1L ZBiB˼ۿ?r+LUe4?|ŇILȩO*) v`G9 v`G$:E!*"R`c3"!J o` ay>_WݴLX(3h;t#ZGTai2s0kr=0]M.qmMbX$V[VJB̐ڒ)ԦPdهWO({;{uदIzB\DO@tB0pD3/e)VWL'겎H()VmjS* J/'S90gW7fwӿga \ Q:@FZS½o..?͵}g~Y;OPXIoѢ.qK4ep}F HΨ@:#'|b9I'm ׋[+ V3AzmSt,OFˁrP%d$H|%|/^R!l*ۡgjkl$s[VU\%jQ7sTQ1kzhSmm艟j3Կi6DF'+)"TLf9M ՗r`D3KCLx #V5rҀh&$W * C˲ӌDTz $o'旂B_(.vn0wCR+db%Bвr4n&f1f7wKbrA'c @eFGl^MA sESFyE#(Dj.%E!p%dB7"7aR0VjXez_#AGL$N(Mh5Z ġI4ɢK 1-C5mzu'ʂhVȓu%VKm˸,AX!:.jR1cP9mweIzٙh@3=`F{6<$ZlR&vۋY%fV-釶H2""/Qi䗡WaY83Tw !t9~OKNzbCb&b/2R ܎ؐ —؀3P $&":"AI@x΁IS:7CL^}Ad(6o}. ߡoP}Ü)N}cUe 4А]A!L ޝ UPp4RP . [ùf>|^P_bn>œn4^\ʞfx|/wMX<8!b^t~Mә ]!M.\⅛qRf2`jt_qe'Uy Jf~2%[dq0n#/6h rpD$D"m[7z6 q~6٤D昅g/xifNOʛ(e XSA"1P_4$ҦTk)/%NxAi+ҽ):+Cx _Ih/#Zܵ^5piZȮ1Vsz+kX-a{]RQkwg=04]ƙAwMgAiqp׮uai(d0l4bh4 D ˶g`-٠gT-: MBDى 6D];tb(۳˜UWf̊P!]r$9=pġck;a,OgnRyg@uÆrNu>T#*F,AmXJ!)p0kX 3.\.]'1*u2Ta_O%d)sLU== ~6o~卄#R2% (.CE 8#S3NJcWwɏVcaBݽj '(ӫ 6n:)rEp^io(f !X |Ii2֫h5:Q%5:jL֬[皷TXb:qmoC|b,KiRnN[μ$ SJN / g H= >^+NTM+N"RET½Tul(*߶p.f0b|:b2(U4UG1'@v{15(]ɸ^$ɬ! !7qF $2̕ByⰬ"$5ƞGoT\#-kTDJIðo(Q[(aWFW^%x+ X 7ضdU{^<%Yw^ƘC@ UV=('"ĸ`BlB"g#3Ղy ;7%QP;Mr}]"eM0(bX5`Ag Dޙ0J)VL$??+,&) &Ȇȹx>ُYKQpG<>Ƿ܋/xc=gas)yʸ"y!)L2b*ъ u4Y$W@+l_[OfPEi!l^t֮6 +k]5/*5\e mԷrvdnvWM cl2BPX?C?5w$)Ni(,!r!CNʶiyHcC 4b:|H6[ \xK:<݇ dEڻDjF2($AJ-[ǘ6aA.13} dHhv/Tͮ{fZE:rDIl$`toJI>q0R>Wz%z~"{} CL%1Arȯ ,rDB\tϒ`!w%JQb3G;l rPw Ώx$YhK2^*́]ʒq`M R,/ Aβqϴ sPgZJ hR7< M\B݃?`:T{Ii 'R;$ 89XU3 HH TH a- Ϋ|*6^Rw\+Cc"s˃%/&dJ&t8 DVTx/'"c]g¹>]ͦz%ou}Kqzgcbnu[rAYo[8}d)G9-Us:2wQ3GwY.%%+O5w_'cZm'h3RT;ԊёX K6=௓&;i>F8q`9#tU!Ciֱ rӭr7o+JT!֢t#o,ʭGO(hhyQB0]1'KK≆Kt^-(>hNg;ed܇tq_}Ly 5Sf2o7eN"OC` ,)sa"(L阶]/NTEɏWa8'蓑d!<K/^@_9W9z/e:^15qSK )TiC"$MLh-ؠ\}/q>O&uT $Ep%1ßTUeHzlڇ1a15 [He3A 5p3X+ Lid3IjkݨȺvjo*umQF*jll˄"^3[!eXQ#YILRISWd9Ү@5J4 c]1GFb sՠ%D5pV8eFSF4S:)XZ*K)ki ϐzIh\`2]ekSe-Pȿ:ǒBvW`+,LP(whB٭ y]/W(=|VŖ_dwnQ]Mr*zNqR?o%` &R!q?NA"Rlx,so!¢vҰ}s(YѬltqsu%]GjQܛ*"_^ Þo(1/8D#"pfE. ׌v5L_[s5C4p z1N#-x",٫iƨPK'PY(ԁQ m ~#/p? WBrҌ 2~E5Sz-Lu:{p#(V5D(7>]mmr򦟮ru~#_1ηt~@>lEgc2~?Q;Qh&?(%>^" %Muߙ(r=_p/ܯ1|k;C_nDHk^cL9֚*J#E)S)$#DלT0B'rV*;|5А(uvDZS($iu+ΩT]V2J +ߕQ*wS1sR@<A*3?MM_HU7^^T;|ާQUhgRXft'q)֪-oQMDmF#ci "(C5<G.jeatsa8vL`R<[U >vUn6Do(*\IKw+^*KKdU I64 F, Nm{S\+b>xTLOnTbx|w}ѐSgC,;^ޙ!9XIhɟvt7q&"toGfhYhU$ ܊P>UsN{%Y¼JE5J8s4Zm9,|MBSk 33{>-ӻuqS]|^yÆ@ XE /@HQJp@b y1p+iV{w$/Aidk yM=stttǴ4oGEv+ج[ ,#B!6ox @|uY&&Ŕqƻ(k lsQ6@XFUZ$uiIBP=8%k OHbG\SnW^{˙YwNkeW  =зVzã^u*UjG\VRM0 F SK8!¹DR6fƀ nA=*ra/^gȣ_J'X?C>r42On}? ?zrop:A߯kįfd*.(*A.r89tA P8^46{6UG0Trۈo`T.!$P[:L%B0[ "` 7O*o58Q!w*{!3 !R))W beOAԙ`Sc[2X'6.(T$_ |T~{r.^!^FL]p29E]DF.ENn{ q+!Hv^J$*)JL24ܶΌyG'A< D䒒ﳤ5~՝bvr=_yAGyJZUeJzw%%%Wm<GrȂ4sżU OhJ;YE\Lű.Oq0`ϣ 6Oߖ53-Yz8`lCF^vGʏ[m4Вȱ8@HNDžp9¼y<f=` ӫ݀lyzr,1^qYB!rB6`,X)Mr> 5}jg1R}>h}NU|e~О%mdfIArL0Bg(TU MOU8 櫎-ԍ*_h?8Nr<o\Iઓ㴔,> nlu΢hXR:=\%HN2("b@I BS͸2G~B+("kHzS'^Lz$4r-9;@ H"6RFX wR- kM @$>8NƀZwAI{y%^SneҠ.Tz`h|[~"#%&Dbo%5$xMzPNk>$9孕QāpbҰiܙ$e2RjНAT/?yÎ}DET_b> k9LUkB?( kzE&b=f'< *kXPbN<ŜI{Y*WOj5Yd(NY%"_^bޝDI AJWHFANh`ژB "^Du+<+R>S['̗VpyO&c!AQHIG gs+'\TSY3N(ANo1b e=x9 Η5%i 'KffOX&)ʉyK9oؑ$y$ǚnQxyWјyupL$l^`woSd:4~w!΅L.dIX%c_d\Y{3f'l=3RޒH4Y ]+K++4]6t% s2Y"w ɒo,YRqjLt^`9+FP;FXo#YįRQ~(,Qea=ل膫촣ӓdTJ;ꇨ"S ZB:%ۻD. @X3))V H-b3N1s: J:yiu1٢58ӊS웈}Pe>.;Z2ݙ~&ŵ8zS!hQC< Omru5bg8e]Ͱ8mv7?!U/ܧpٗlF1Ɣn$7U3/p9;&:[fo9Wڤ\=?=OG"6c$[5 DAov|㮣F;6rY7ۘp^ oUZ~K,7<X WڧnJW6j+ECޡ(H*=oiI8kQXupMgկۛ4TAgDs2|)JTj͔x$ DD_ҧ g!?ђ S&D3 R|{9fKfHa=pWgzu`gIx_{,4gٜ߰miPxE0"5~t†AP .ȶN⳹&FN\*BC]xZA1`:G5I Fj"UlOKv3/(Z5UDm=GNI}H1:<:M1HÒJ>ZK>H/HMU,*c8tݱ-R.+LL:"?Qw|B?D*lʼn\'cl]IxW~{TL7cJϞyzrQi6FD%%Z'k7Z#Z~-k{u~QI*(V=lO*ݗ"ݦn}>/Q'bJR7{ Ys%rWϡm\6I,>4BYpV7˔{B9nh2MF}ueܪć*H+IysZ(|6K+Υz@1=rjP+yxÛ]\<ԃY*H &~Z`oo{62脯VWr)hTP5fǤ{/Λ/~H'\@ l.bZ┅w,{&sꟵ`QBAcM1WZ4{w8\vs5!}f&⢑6R54Ѥ(%nSmk",ds?*HKzf?Eo0$҆JaԠPMh}p?p;|v  J2I^}P|^:Yz.߈*V V(HUaU3!*?xռ^HLs_`| R.0b"`DR/>%H5U:*cb |B#]-0QtTo+Pvޝ|>l!HoCGsQ-2dfk/'HF>;>StP$9@豺-(8F`ۆ6 1D(k͎oowCx-Aq3}ȹu♹Cn'5hAD>a$a,J0T}>jp*z#Dtʮ8G:Oz̈uw:dԦ"zŬ܇ۦm_!-9ZuW8x|}t{saǻ698isN9/pW/;(./ eN!&c!*o)S̋tx>˾%=zwuu59*ۣL~us6믥G/W>^JatQ9z2 $=&\k /'#8Essx9'mI7簫xP@kZPc4%ֻуDGaFQ9ouيCE;F:uxd!Q[uNZuNTwI;k@,IF#=x2;T 480d$- 52FGaKCEK%CxZw16f|ƌOژI7f]$M& q(&6h{)Bk;\,QQ OR5vj |eL͉DI92$g _RxٹD_#LLsRsdZz)d\!ӽYf5rdYjȊdؔ3dX9'Y+=u <9i3ydFDĭc,Ȃ2""u\R GZ"e@b9=rNVKOL1O5cمX/ɨMZA$YnN)iARNYT'I0HYDCMD"7mWu2?kooj,f1er9"]'ki7; &=`-_wo{϶䶍쯰rj+=ImdJKr(noQ%")l%vyF"Fht0D` =_}@,ZԊL1^=@<DD "u'J{tAw~7PNo@H koPvP'ExSiHh@O +\>Uj-53"{[~z e0eϵEkO7)8*!ϝ>ڛ̝4 R w."buDwr`%L7cRb~#E% _Z0rRRܹDoΝZ`/媷_ ϤfRp* 8l%c]E,83;yJ"]A,7_$ s@D+YF = $F|kɕ / S$DU^pP9¶%P c'8d~.W=^mtB MTy4 ,#N%TI"K%oRU! 8`vsm+A-J,x8`$SȧJf)]sC0AfѧC 7c':)z+Lj 1"jRe(1L[H1 ) : $i)4*lO*)-Ȉ{nsf)ey`2t'׀l0]r bUM.r8͝)U4ywيM6)qzNʃA`=쀥rAK(QW? mH<ԀV{MYw'3/yB ] >O&FKr&l,rP[ɻJ[01|Kb+y)K\F֬0Iԋې|Ӕ}!֛dꈢK< YN'mwfp l~ ug šXDgH]ZL<6A'#爢&[.aΑ)3i#ӒQ-Nw-K*?@}$E:MOW'|nXh 36Y[.%7J'?@5ڎ;$eF(Ĥ<͙(q6vHIek'7"&mu3 SJE[Rǃ gJ%"uKt eЯ[SL*9¹fobM'ݼF)ĩ`MTTe2HPR+@fn@93D;ȿE60N[DNhv!Ury{,J؃ X'3NU꟦Q<8QKrz^gftʒƳOb\ v)+O㡆p^" rX6xɧi=੸ I m~r/1p%i=1BnjK+*XJ+H%\Wb¨gȈ%}9Ut \.+\n-C=s\;9wRhђOg@\? SA6O FR)JNV3J]&MQٮzL,3.`uyG]mL:U6fXSІ]sz^vx7ݧd:2ӧ@֏dLB*(W;=JZ3"ͽ{3zFpMR'S݇К~o3 ^vH9Zh:+*ov;0h,w슮kE;35|M1sNc+-xU\0Ԧ׊+AF+ٟWٟ;ԑw{I\s!v+rhB1L*{I+<* Sc|:P]K Xsp]&~wY?ɲ|76~a>}3W!_,sÂ^d?{-[1Lj &T5Ht6UP2BnKۍ{~_݅g=sV /4GEf8+gyJie P2Z3`,w+C7~4uhx7]&!xv|deF/F!-of,C(;D7ϲNE)}VM< T5^ꝣr(AT7<MV`CS HBĀDyo]ʕKHd\J+g,Ǝ#%:d>v'bp%):XZF :Y[< $O ypUyṐEEUzbfGo͝Mcw<`W&p.s%ss8ckuEgT#3C<" PV(!5:Jvkj(X ! % %B xܠXh}¦JѭTr%ADs 0T(\ŚBăZ5]Bsѝ`fڭd3 Ǎ-옼Y猳 L<$L40IY2BA <'>KqtT4QShqBMV;R 3*%_J„#$$cDpL4(gQlMm(,q*v}%8GZy&s dPb  QH)6$)EI'TIN тdBe5h*xC QmyĄn,'J,iHMYlLL~;Ch0ӧ,v4fLȄac?v#CtDchsfgwO> ΢GzҲ[+KpCkD~pdWU aXVEn_sBm^K46G})Pek@mDTգZ>,gI=ߑHc)#49IT "pvfNyV[D$]w'jhq4X/ᦈ1@5nEyFLE!Buo G7/ƻ)?}E%q@$l2#r׺{A?XnU >B\8ds j~ל&f h$oר񳹚se ynh&:ZJNϕ_N?Tuubt(>Y%,3o5MϤ(]8'5WEF'3gaG?O}?*DNGnjxP 4.EN6s%BÛ8Fţ]Qj QQO O{ntkga_ac H:Ifz_yD_|3^0Y+aѢ,MwʑگQGߌVF".n s# iV3y×`q3]K bϟry~_|?%}78*U1 d`\iatRN-a62jhEs̀`G\21  pš0vϯS H=hŜѮЫaEc}ݥܬF9!q>\[+!=; :P_~TE*ɦq)=k#d9QQ @DȀbnX~w_O|u))8Z>6 rgBֵ_$ ZL8/Pޚ0*8\z /1JLH&B_M88v"OHi e)#(T9P$X*D9a~1y]l:—OؑF1ǹATS۴e»?@d/%վ>6:?3 KfIgLANY!vebsB点շ[i2/c {&]Z8}csEO'.W*-+KzwۧXnɢ5˗ m䩛P$\FѵC^~eb0,m,Юk<(H;:YQxs\PZV1A>I l[}ءw*CoL& )e\Q`&shbi³,% 0 PA}|9 bq<~ տROjz >T3/;s n` ̯C2m xX3ak5a`D"كDr1$ҋ}9 .s砼Xy;~2%9b$8幙(EӔIF$@92iFRl139;dM!V}q"la \r|ڂH|0E}gN(9ej]Vg*w<0OHbu w(7 1Κbs_!]JEKSbz/즋WKKk0򕫨NYHۄO5դ|O53}W,Q:%T.ɢ$hUpD-JJm)֧Z[Y&թE)hbxGZ MSwZ)r8 Іg"Kb,Ḋjq=԰ 458pA bwhņ}G3Ԫl9!4+WQ/{h0n^8#ə2JXiXuيcQ{ҫ4IB -H[BQNrrqbţ`SooE Aگ};W(x{b9XNXFQa`?aeGYʰHQv&n*(fDx m)}_lOjzLTq<>da]'^r"u}}ThT^R.]E'p'Jqb¢Iz'z-| WS(N|ە! x̚yfM݀lxӰ|j?v/BȔ5$[-eue2ܗCjLl*8ip XnK QdRȊ&rjfv iDu T.+ X |)JS8jֲp/ϡ}elŊDn4[}\wzݿq=+.d "FA,3N$%"$8VX0ʳ<9T4 ~jk@alH Ӱ"҇udEQ4DJXm~ }- iS,Lvk,׉sp}HBQN9X9 )AhwHד6"rD FкšT\_[{֍/#NJbH_BV+K.M}G}$$ˉu3ùq[ϓNu+֟\ JY.6ޙ{>>nnnnU qoX%#aFxIKFKjLD-ƂQJ ѷGu޺u*+[M;:o-n%BEmz^ht2;\Bs7)DoFxt]qVbò4Y} ^J= >Y-Qc Y%|!Hb`}k^)Ҭ촏~V^qAWmAӨ{# T1Q+6 u-nun雠fM'L" vjwT0+W>+cCuZ E)EB**.ŪU:m#?Ϛf\ fOq#x68W=yAQ,>SGyĚ+X⫽<1U55'?*xrU!dPf{gc0ȁ}<ԙ˰z(X_UœmwW.:ؒLv:S_6C!wr[U WPEՂ_ϷZ}9 ~LE4ۼa~nkg@kmw,CwT0.N1(L~sӱc7OntQ&C ZaPU2u:D(cu\5ʥk|b+:1AخN:$oսIj:RS<eq)yocG…؜XøspP*@]ZqũP|Kb͹[zv%=ƳE=p6ͦmb (`ήO?[gV8bnC΄MAs+U SE)Tlw~[rx[,2%U| Xsie"PYm!Bgጴ[7 cȉg`μ0BkP)pP&C3}0=pQ 'oD+QĘnKE| o-ײ)x FS:@6pcr.*׉ؘ^}cpdz8>EEy)O|f7rBl:`1a,oxQIPs׏pROx8pt*\syUcoE~MaE_ɣPB,FOb:2ZQU&47ߊj*d(,Isrp GnX.2Im5Zpfw:Я$"lE5@%d^`!0#$@J|$ W4Jx 6\SvoP*a>x[kDV*-9î=IIt)GTPPF1l(a>$ƩX<$'PPmP;ݨ4,Ey9ZYQ0.aH"1i,TOA%lepdmGl{r G9rOj%lj}Cl[(BiIf!C9 ⇀MxQs6az/is0X'aև7?~yC?lo6b:8 w5[R򂅵^>aǞ3) a淏>7GoQ;G=Gwpkstѧf2To24AH0 srr3N@‚xH$1o{42F1uJVkiR#˘4cNh5 u!hNF (8W V*^v10cL0a5r m&TSQQ`^ /[Vw9V*`m.TE,ɢt8d Iw@JIIjG0|wDZh sGD{ :sݻa(K=v6^Gy*K3D0n IͪKI攼}gl +[Ƣ2|K1_}=N+ o;3Y7GdpقpA$@о3İ$tn<&wX͑N7>T5"HYKT]z*ܒeZA㲦~L}[G1ZD+>^M,:yݹC2)Uܛ~9z6j ǣ,ŋ[7+L R` DKd2y:' q>ݎ=w,N gEf0WQ򞫮laތ=94[uMuRa8 SkPzZYfƧgQd\: Ӛj-?w*n]>CAb7/3 h#Hjzh5,KҰ%5B#Q4(E 0c(h&Iiic UWj Kՠ(z KJ$ ;TisBB-;Lku+_\c 0О÷*}6$Y$6`vJ ZH]bl0k (^(_M讙#x<}2VvAA{aE\{U˒Vih4Ճ'krWYnokzɢE%o5tf7@D+q w!p_;Vkdd}<+1i2H"%=)l?p6ͦ(S 9eC=7\/nB[S?jQNE?{޳I ~,p Ȫqrb1Xӄdx̠Cj}2 XUSr&t^dGƻKLu_?Ώ{5[CiI+L$ 88 sI-^ٰ䒦\ATl%w1JFf55E S .7W:륞m cnz봴ĊPJ1r\Bc/1qNrwvhs]}nAݭ}{ܸ K/9./[SuS`vSڛ{z@JIPp\iՊ grZ|Zj?EW<>P20AGE!GxFW/_R_ώ`|=ɤfד[{ ĆPJ5YE崺q2s ?&opbޝcz3:':48[TiMoRV_i>әʫ{W`-15ѨCsS^]\ڻ"kd{P*Ƒ Qn}~xڊ1iR̢LCp89@&}D띢Q6 >q".6j%k^vUq*d#t96R|i2(;`e!#[,bk" T9CAqf`K}-&rIX 'g}[˲ʙ.$vdjrj6'J6Nmon&"I 3F<c aXvswog1Zd0\t*WhlKYbДNJ*iXQϭ\̥`ֲK`%8=[[L.=Og_~~lX~{n.RщIQhjWݳwiQF @[-+V@kLi:w|$<۴H ۺ{r)RgZ@ :K%D=.r7/c=S&qOWmtr3h 򭷛\n. /f 'ߚgZ \2u ];NC徃VވA4<0`}c#ˤ2c#˔5 X?OorcڒdӦgo\D Z-4b94u+zwG9S1Frսk^4(//fA@WZ$25e!.Avb\ٿгq>p&{9p9~C]bܒZb="Mo0$sb6,䕛hMU6Euh׻A,0M$|FMG[:p^w޾) }S M4˦91<2WgEҭ7uASpۥ."w[Cd' !C^3e}Fx訳:a!DlWv• [.)&U՘޼[iDֆrͲ ‰oUݷn#WhYNTHz*W-ks$胣# I<őlwލd5FA@ ׌uJ[?omA"/ 6l* e (ڥ3L$%~5bmtB3%ź1A OQ..E/ V!wQ;4*-$NS?Ƕ/fHAw)"URЭaXbǰ#a#rDEf5 BkWư0ZpGS;e fETލV 02m*Z|*k͖K-jq^nˤh?\USOf7*aZVZg|,W7͒E>$~_~Tooaxuk`bQ\T 0b6e餲8wUGWRoi^K[⋊ KUjVFC-8՚*Ty6fp̬0F@;ftnoV ɻZcbcsHJ;Ċ׷[p.GZk Rja6jUeDj2[(rUPܶ/x]oWe1Iq FzWz14oUKtmgI+pAP&͸EZ!7M5E\P%A^GGI:!zz/Ϙ֕6 ]OS<\x8h ;!gk& I(Q^sKvͭH$@j}mQ'&9*M5%qx$LXbCJ#n3G[zXSz2X?27գ]}Vv-"Ktgdc‚7g3H=%YM2 Pr!ZJ)9Ê5(d"#O/ދ 0, N."BIN.N&D<8O.԰ b11jV/kxnQ0zI!-/;FV,`jRͷ pn5A9i[CӶT9 Bǂ ?j 0CN6vP {#&Tn h Ujw{@JÚ_BX RYV{<%^S.cIXp  1eZs=ÀYԗR+R9FfC.R9;kO5:v}:nFY=g1zvc'P r IvW!#~߅07h>OV{B֥~2o1Jt㭫gIf1< p|}Xp=6:1,XlVo~݈3ҬSn&Z)Q#K Gu}5}=xwjoA'4_k)Tvw}7J`t|. (uvuJ u``ALdåfTj9xܖJezbɥf][s[7+,=mfBm0k{R)$/ WKRHJN/@Ry(\i1m}@htkHl}Q6~mZU;bS;9]8UYaufa&F uJ un B g}G֟M8-o[nM>DkvP-HBcڭ.9C;h})i֙ڭ‡h1m)vAЩ*ڭ=V̴ٶv!ݚ!/|>H`J32RdksV]ds$iOGIJ5qs&*"j_Z\_thoSwhtY@LE_ :-Vpq1 Zw滋>\~~gߏV-U_'U(>8 VƝl=>9[,D]vr~Ut=n[=?AG٧z pjЉY"epv;b<7PivnK ; s_I6PC#z{ֶD?8CoCI`ɣ3X"Cb(,q|wXp4I9qRۈkw+5-^67[EDG6{+V1 `Ӿhk(La9gX7`NjTcK((E u;$mӥ]OqbӞ:/[$q' /3_*AHƲ/Rs/ژrzRPyTh{C E9kQw *G.X@3(E Eh WZz+ 6815!Pc@q KCq-#,NM#EU&Zj"0GssƩH FqQXs˝{g ] -pJ Zq4gh|~NSAFp#%_ZO7S(y1iE^-)IeV鑻]{8.)LӨ?G޽%1$7W3#2.ƁoFb6_LJh}'6Q/=zw! .F'ltYת}=5!Tq,FkeJ׈t;ƖR[hԉl=:7Q gza7Qi$xcLcG;"VsR۔w}5㛈dY.׳ Ea0חKog~t1߮'D0 ׿xy?Ftt1_^XϱFz8;NJ3?cg|VVjsl݂]\'ʉhM`BImQṣ%9V ,~ud˩O>]iK7bn\4日6&bµr~ ##SU=^85Nxxmֻ Y:Hsv[6g<{ l$2$9=>W~)B4(WQ(p*:ɓ凋REN%gU*l=Y{d(ƑﳇC!Ί,aeCؐeءacjlsSY\YwZCtt;'\Wa &U j~ouvPX@}긑4X" c- & o}Ͱ^=ګT!'΂4fdP(3{̫N#7 $BviE>HG;S쎳 ͮ[,ⅇ.i1&S T `K^ey[)_Ku:%[DN`~ bl5+W zvMS2V+НwP\|RY/ْ@A7󎰟˾Ž:DBֺaw 8O( B; ϙ9yx9LKd,>5(HÄqA3r+d꿕Vu)@^! Hyg% /i7T:jb$f<řR "y$آ ;!SLTlRF1&qNBh{uԹ[VIpz$8.m.:*:N #r Rq9)T*@‰dD&Ǣ-VYCX%B@TJ#h0 L4 JXJWT8f‰'O#@8TNCO9,NL!X]W̘|{rl FpyM*jbj3Sm. 8}[y' qڌhi)V{0<"l31ƤQjM"KhdaQbi\3mR6q)%M].x笳]~Kmwzj>"̖XAR"y7s{px0ƍMI{ ܙ/_2S6q4S⚄1 <x;r NvK6GU$wc j%w#ZUֲ(jɑCrRX\yô=nB#)xWG'Jtڊ*8)jLP8ieNR5]p4NcjL2 ,EUhgfԧb=^\@h̨{AB u1- k- I$(itjvp{ ԅ?:H#HF0*j$. 59+ 6^rښ FX `#$§2 lj`$舖 $Hbg+@&Hh?5MWz89f]fP9zo+Ȉ*lƳ:&fe5 v%N;$ޤqG}+}{oxl$R㨂Xue'Hh C7zop>Ui4Zq͑n 9ja^!l<7MsS#;\ن9yhɐ;0 YxI? mڦBBKBIܔnS])v?׊Z%Jh=koߧaB}Q+ ķxWmycV+Y .8xmۜ= Q#g"ځ^:{x@M??C@[pL5=Oi8Pܢ9Y{D,LAۡ'3 J$xp^AʞR(66p8EH585`QjX Og(Ftvyc4~bq>z3ߑ!jtˆ]y&y`_6dA'Ovfa٘9k옒e$}$^WH͞d=m]._HE4GXKǀPAwLllIΛj6#M>h,Jp1E< 5ޣg1_|DD`2䨦,&u i b1L91zGMpj`G/N-E;H{:H`!7Dm=gdu Aa/< |1.vD KN%Sb -/P,^WP4''y]/gq>Т!9e0J 6+p4Sr: }檝ggn8\D t8":EXs2_JBD1:&س *eu{|pWO",d;ĮrGCm rZT"|~ ;5HPdAVdNDnggoN=& ۅ^E1kn2a:`ч/ޕ57r鿂+XG ǮweQ]ǐ P8,$YH ͐hteefee}7[p`dӪgƛ?hnE1Ҥx|X@/23 ÑM>㔩~>ݸ^>5 8޳.rN޽ჸ$obxb2VࢷZooECHFՊ޻-7EҪEVFYA8: ք@i߫* xݠsjʼ [u|ٹD9JA~?&—KdmCBQ X*RN>q~ %^_*@;CRCVsNclh UFL7iK9wYIC,)zZ/lRv^]k{u[#irD['ܻ(ȴK/9*z1t`9b-,jUz%%cFY# w@"A3N-Y8J.J@R# ҒR g14!=!f<_%j:ìj$:xzeN_}t ̮?~33QdiJvs/a s? 'donoo7}GCJo.4'LгL^LH߇i"Ho} _0VF M0R6/ N'5F^E5 ~Sw> n|RH~ bd4U# Fm-`"7FxR鉶ŝ LD#4=)Ƶ^Kyz@%ݏGel0 CL0ަ˳&ek BUN_Mτlc~ar'oNs3KV*k3^;y]`37˂? PT8(JL;!bmPT^|7 V$kC,ؒ+M:,q4A03a\9Q[g:vIz6Z0ƺ SD2q8n FkV)5UWDS~I).*|1ȫ3iL xJ IX"{n3<M>,UqǕq'L[8~C)zY)L<뺦L 1uM^:"cO-f2~'mCoPOiXqnyii_ORD {eaȫ$:m[kXŐ^vFf}z Iꟓc\V z|5L itsT DX i90IU(vx@ٹ.,']|0?=&w/k,J_i~hhz*Pttyb݉,v k8-6±DJϘa o.@/R\PB9vTD9 +e3k<`&u0;KS6܍>'KM9pgP!\F/HKK9kH),KDT ]#I jHnMj)/) mNaw1*|#-vgL~ǟpԌs )ΪZ_,[N03}i.S4qwdjf$' ^n|?)~F^_U/=xT"B>B0Wp@K 6/>l:}/H|7^ `Dٵ Kd.5ZJhaf)w[K%k;nJ\srq/ row?_Xg)yK=VztJ?+ YLh dE"-VXEJ8V@eFHO9)T,RhLХ鈩+ X7=Ȩ2Qi33#5 ؟Ѻ eV-*a2`q#lk|_Friqnp4T0mgoVӒMۘK:&܏dϛɴ7KhSҊdW>lwcb5}'!m/ͷnVWH?iŔs$<]OIb-ѭr\iudи.[F6<]twy{;t_!v܅TmNա'lt su3H̘Rd7ǭ Z*K9gMe!ҕĐ|б+Gb=BRZ+22Ǥ@J8Tz G-kpW-|.KekߚN0qL[VZRpQ##UQ(㤠ZJ-5,E"STc5=eܚ2Sι$Doh_ByZˣv( ^3[ >16Ƹ*Y$`˖J.P\Vֻh  ~6'Lz ?==<܅>T$z6*kA/;㦗JEYÈLf):WNa?͐,e*=wNEZnv;30}0%c5zLe23\΀_ݪL?\١8w/Aӊ8m+qr` 43i .=t`';")8ppS`ܝ6gt6h,`ģUsYO1 f]t"ex}¥a6-JMP>͗=;[sʴT~vQb(=wF>(]pc H-(.~,_N=[FZ "0hv8G6%T WxKWk6exNyK}ir6<{ AjCĞp>h7qmmvBcC~s,13^^bɜ%H MSj 亴NS5N/0@eiO ry2R߉AEQR5*FKr3l.p_T煂2D~s|ƹݴb!qt|]+9Pe0s,1W(g`(P5*̴\)/nH޺ܒ٦ ӆiJ%5M#iw;0jΗ%fi/m\FutaR&=ҧm -d1OkLZ=E+&c YSʲ:[)V6+(AmOJRt'EXqAE2#&К D7Tn=a%葐9y!g<&>Z9 ؞/˳OQ&_c c2&}$Y6y L!{Cwp-ƒzqb$14/2<1m*IN5KL7֥LT$b,nxO(E3Jz;vm=[YOXƁN9 S==\,pTp $_σ3aוtQE Ep ;fRIA)ە D]5s"꬚PgQⵓ-S~R6r"hj*ϹTST0eiwmj5wb}wsߕ?Y]-7{T*0Y>ueY͟^Pi]XPid=Kg5:z5B JK[?J[SNp,[{fh;os{۪Dٯ&[m#Xl5RFȧu(#%BGI4w `;ɝCԬh;nNNbTȞSUw?M1Oqa=_Dz:]VAbR*1Ig>w鏄 f] [Ǣ iO q*_=U_oFB)+v))tx%*c$Q  .s |7*xC~{3x8fkh.Mz3X 42]YÂmzV!dn|w.(,@5ԪJ*Ěˠ)9R@x 7 J&TQJs;T+q$uW\=TKe@@痔՜ܻ(H- T@c4l"sZYԪJJ1*U~ښTLx|gYA1'u*H-QԖr'RXRk vzϼh 5WXt1銻5)%3D5}nm>\,BVot͏|nCNEoэޓPYӰZ`9 TTntݚv4;!';60uހ{!+FQ!gh>;S}rE`NwWbӋ}<>NùNbu2- xOxuW*o:o% 7lyTneQ;k9*MsC*+52zB#58/!PpDA8JVJ\yaQ TevE6+ULjhi=x R"TXT˼qp3L_ylrY4]ILFW`ZY\Αw]\[!0YgޏRr%.L=GGE +hhGec^Av\@JEyAY_%KBB\`5~*֦9?68I>?2Ar0:[]p6;tSL5S\pv1bqhÜX~vl'ɶ}).aQ# uOKƞ4x3oU-zt?Z `qjSX]1Т`su7yN}sK5Tpƈaga`ܔkS!b(-S"LA!jg DiA8Kγ][G+_V~Y s y9g1[k4qߗ%iF=[-[@3R7ɏbEycוӬ15hVH"Ѥ840|tV>Pc!=9~?w1P8+3\)%%ډbX( 1b0SOLxpyÍs1PM1h{'0VGG|r7{} FK=~ 1IG1d37{|RKWNQ\ҸS'{|} 2 םnLJ%mJ/6Nx±X썗~Cu.7.zI,Y%*j)AG&g䇷oO<ϢW|6_|OT}ܵ4$:j[if7wfjn5H %8Y~4?yVTwNUx:^f`LT㝑țp+!_՛?3Y,ƷӻU>> Jڹ_-@s`ЌPTEd^]t3z_6ߨ1 }}@)FBqв= ޙ8Pmr(X@H*sLU[[d_㍲g1Y-q?RErv-(+M"B67t1A-rNI+I6'geuy,Φh]NYvn9wZ-Z+'mrUTzc wg?GM[m( nBhHWbygBtB2&fZb,bnDUG3K 4e3cҩpcc{W3hLsGjODPX {͹HKҭy]16l~%&ZU~,P5?NoYD'rj8'C;pDA˿sJ |: R1x[u*+例-`&+nIq`)*}gԍJs R5="Yw}LaaGD %5:"U. ^\ɐQplXVخVz<"in 察uI:pWp 5\Þo+r/sLŧhP6Xg~5 }|$Џ n%oI Ыh4IEunj$;7VTOfðd:ꇝwfïV4l}_N&kV5̉0#bU\r>,?w.|ɷ'ށlXI$: bR̦7-wM\IӅ&TWnm5|LcL'~שRR2i (pY!:;=HjtN[\NM_~Q+Yq`NK1D4gt1"%_;ᗢ4S4HI[RhYvi.*TE f=7h)nJAzr ,&J$A@6j֞cY2{4- ~ϠDf pc %vJB%`0`5`9x -A3aS-B!HP^ N-#&>&?Vn>? 4{Icxet, P݅}$uq,}=kܢ`ۣx^1~K’]c0Ta&ڞ*eȆcbFQigT'*QyeK'֔jTp X,O0#h+G#q%<^ϣ\~&B963Yyx5^,VS7_=-O_̞mWWF^G؅Vu(@,?o&oz$g`}刀p#ގ FD3ȣ}_ v}:reV( [J KBY>=?߸ǀҕޓ6,5DE%vOl"=[l/7jE{;?>_#-eIQ3~/a>.~_avw&63ZS(h!LzN޽i'+)T(ѳ$F'䷈.;X=Dž&x8cH Wy8aUשּׁ* DeFOY kLWxݴwCѣMfӓI;a,?U0  ֲqx8,s:YDf9e=omUKucukW\qGDG8^qG@^cCRfa"HG^%pO>5Ѓm%|k汋tJڼiOjӣx3wJaZiFalH(чN gK|7I_tz^_?0hϟ7Ftr=]p釛ӋuTZ=:gM;j)b[u0:sӿ,vxdZYMqi8]. Z n=.7\Y"ǃ`L(Hc賛%EQ<^"VmJ t47J##JHFz@42AJ&̐ uFa4!J6$K-y-juҨ߱*i,ϕ O<\ǡ>ZZ0.m.+KR$ZZٟSJhjmzRuR*`lɋmՂvR*)<+&ViA#Xre9Yed,45}9Їl3r Fbxp9E-xVd=P#EMnO]7{pJK5Ye$ASÜK$wﱥ:;<.w.O4׆ޡ1rwhȾ'R>7=%K;z:!i=\ȍJiJְ> 3'JA_Ղߝ qt9j1+CPcvI Ӵf m2[lh6+`8/.czz сeU{O*d`0np56p0pk'2V)1<eY`96`ldB"c!#К6vAf#O-D#EQ޷h4985+ vhe1+֠K[׍$&I@vK MXfS.8$cGf!W Gʶ 6A1 +V؄D^.A}vC8J#pX*X #NHa <J ּ={qu"z2s|ݨ@'A`3@ԌO;kچ`mAɥc0+vxׅef7Ph7+hJ^9C1_iĺF$EY]ώ&;FF'Fl,4chw'3]X m=V!}1oJv$<wnM44i*,4_r#ՎX@‚{v+P8h5 t-S n&Ew 'e/h`j㌐m;ߐzƫ{~^>,,iO5 TSM\EC ={׭=۵ `y 4C} 孻KZi^_+M')e{c{$H9xoioe^3L]z0oH%yɋ8i@1^y;ٖ d"rhLy>(B[QkŋNgYwWBne(̈́yGr\1s:;4móOޏ5u)[Nǧف˃t6{a{Euk_SDn[?>/4ok7u,ƽVF=&(sx\װ׽id;#32H2[U_ ɥm[1Lbչ5ЬM 0Ǻp˘Hۋ|u=z\\Pv=G5S +c#/W;Wԥ ٫iYQ{Og/N#v$VǬ\rlOfa7:аwİo)j؞4lTY[P[Bݻmk9;[ԪP#".s*SP%f~U̻R7Eɢ|Q> IqZuz?_fT1]9\ޜE<ޭ ~Aw+t=\ѪS۪`x8I~fRg\߷ovEvo~Z8N+LI\D{ɔe/R?n\ ֗5]3z_+:nCH\D{ɔ!?n Z}6ϨҠFtv{ƨǔnCH\DWdJWʔ F >*쮊|Q#Ocx䍔:76X2jYWye'&gc& o`27˃J`Q 7;[ le%/r#```w!wwY4ax\jV$VU\1F11V+|8@e1SD5”ϫ w4G~p`iSjz 䢱INF2!Y,$YaJlb;|56 }:Bli8+C\&Pud5Y_1 ^*O1ID34na61"dRZO6 m|4,!{@ 1 ddBI+r`i\I gœJJ%:?Rc '5d,pHZ\(E# >m\chd]}lK !2&#Y1+)x%sdd1.e) e <=ʬ(YQNcPB- |N< 8F@COGRVXYy;uGW7݅;8+eޑ'K3TЃ^Fy:_N~1MG|/L(aIkn,8meۗ !yBR.i}]wC$ְ`=7[n{1?S@A{6W^Õ{̐~A6C 56ϠnI=Q@Bj!ܦUp>E'7+eW$ҮIk_7S ̘f;ng`lbkEk}E2zڢAkFjHǫPnĮba=A^{uo#ggޔW 3; vKU޸Н q`.is>{ͥBoG( +ڤKJgU:,YDHIJT\3 R@QE=m*bN@6l\*u=X+3㩊'M3&t4GTscdA L^`0Y;(R/HyS72Q&Ju |JGK) 웰6q Bŀ>%6![: @ ڐ%[몲M19~ZjzZ`(/r\b~x @1c_~J򽌹oGϷo^U0([0_|W׫nMYv}&x/Hm4Wvq?p2w3 'W7kY?V`lx^2IOΘ>F|`摒?D r[fTFnye(9a-o?)ڏzƔw%zЊɚFg> F׌MW/oNeب=\.5 *yQD7̴[M~e!4Y'%.?"fkz 08v\CNO&HM4-TF)QOԴdkZ4ܛD7-Z==wiQ?Ol%6-N4-?m@[ +wq4 U8ũ24 h]x"}۹xm43"I=lļa!pM)nuzۻ[ bL'}bV;wK|rXDSlJ3yw3$ tRǻocfOҋ1"[ hǦxMP@Ҝ@TۢH^qK?+ 5ˎs64"v"›0cJ܉U%jmķnZU9L'Fӽ8swTA4Ui-J/uT[)YV9Hc68JfBsY](hJ"*5w "qV jäPDLw)3X3%y h.)ꆷ.\VLk܌k&5eL 5)֘ɗUwQQ2VAu9dW0l/vpB>1 *'Cn 6R0C-]-{Qڇ(EedAִdcsWb.aI`L.gu ~x/<XcucϿOM5<)C}@˫W[7rOߜHZbieLw{U߻)zb^?чTE3" MPlgI7YU$ Q^on2Kl g@Y,l˭ou Fe7@4ۉ#^EMXmW*[=^Zs.jW6Z+lUVR:u*:c?ԭvd=-;Ux!`1[v[_S|U#֦a.MQB8+ֶM-UƼn*^ڌөmnpl gF!7~R=x0c8/odȉײ:C::/\|yt^M1๻ߜW|!sxytUY80 k>nΫP$Tr9BiɟaF]ϫ9W狿} ~}ޅ/:kv[j[q^{wR'?>m?_9[7*=FoHbJ%uUej)DWuUY?qOK-fy}6ѰS_b]AR7 1i]iwV]/4hd%7'yτȍw0rVn- b>r<'X1l6f${=敶RJɄP,F1ͩ)[1E Crl~uNٺrƊjJ6u'(8mm@K`OiRdZ#G蚓"]tV4z C\p)/.|fg9QMHƇ۾ƈ%C)c{uA 3!|iZhJѨIWG@)|92-řQB5߆OD'fzx FϬ%RD $'k.r,)&ҧ(6HROĭQnz|5|8z%5n#5A^l7\}Jm' P;rZ kh}&$^sx 2iD:pLpGT@Z\Ci"`96&#OբJ.9%Rd${r4!ց+op)@!VLHL腜R.Ja-\9A`Pr1p`UYF-CRe<#C>O!+a10H8>_=xv=`ҁke]2MR-J^YK?c*|TQ Е5tYļ[e;yeռi 5ByɩaF$!}oY(yVT844ye 4.yyRf,! DVPRd9Ob8q% kaP/%F;/ڐ<<$ϘΜ)cdB?>փ>X`FZrQ)ª,e#4+횚gM#Q-\dHZ/JBJ}`3,m_s=݆stcs\Vu](p-D]!9Lu҂ZX]j[6}9=kZޖ Rz0r,=<_Lt2pH% c Lo`q;Ѥ$}>mS9fJ @Z7I JXP*5ʕ6!ǜ;*!O.9Y)1pafTOG[`~k*߹ru{ǎE$V=E.~Zs=%V(aԔ Q'ӝ ;TDN,Xq4J;*q<p̓9> h΃aFA!LX0^1-APpw>껾Dyq(zve1-bܛD^M(Vmz=RԞwu<ͩ&E%kMB_;{0&;vmn^ݜU֙aβV'F0@#C׆^:Ad3 my 4THtILOua@JjGٔ2M2K+Ti Ss"T$8N ҡaD@,cK8mwEyW43\!2RF)NZ.kGU-:N)NdUFI($1Y*zU? dhܼ {_o?Ws(}?Z7_&@m] YT3E^2JDŽF{-I0e QI]h"lkf)YXQ ÷Y zR{{{ZҦnFH3:ᙗ1yy[rNw )4B?0 r62EBH^>݅\7CsHVYxX'&4 Lhy+ާ|͒{+Xޟǫ~+~aWKE\\ _nXW7,͇ktݎy}y0q,uBxd)싙@| ^?/@=eتJs2wÙwXW-l< -sh1NP~<ȑNROÔVjrѡ.aoѸo];W?U;壽%{ӓkq7Ed2RЌ'6:k?`jɵ% ׌?RxS/G M{ DA@&6Gً' G Zp'0H;ك^FI(F?^I!\X,w*L*[h^3wzͫ_\2 j^]1.Y& b\ 돍+겨aSQly4afSMpD!y _gvHA/ L4ޞY`f #O[,iEʞ}#I.RuH{ڦT̊+88^LQd 4dיi(%gPi;ִEaUj k ظB.xY¾yȧ 2N)h\`N`4TNEmtw(p W7MߓRgѼpIm 2Gɽ mNj'FaKZs7ŒRh6 adLS־>6[CALi )cJ7Ȫ0+쁻>* $d@5۫S:)sT"gtwŭ/Qz OS7 A@kzyp{XP k`>A[r7XF}Y!8|nЖbImH$˜LY4{{[v6Z~Q?) uzM/(F^+3Iqtc#*3&$+ުīɚŕIr)%y!&)褉,¨Sܘqj"J:q}bhI!NZb49ZȖ{GKk(H4 )m# 쫋J)J[ %A%,ӅPQJH?٭Moke7 VO?Tk V`RԉJ%Ɵ?fR@HJR0#X=NDF0AҺ#i=*ſ>^U˳Ս]~ZtZz+3;eqҎ9};8J<V3L?#10rHGՠ@R)$ ZҤ{f=JzV]l^rquC~^/J #'We2Dȏi~l~])lkm&NǜXGg#,x(X Xv459kL2.Dk^9-et* +πA~ԖAHgx@Nqa(>F" ` YrLAs}Ei%i 3xKY%(.D $K<$8Dц,mA :\"V䝈іrV*rYiTs1ɳW |G"saO<ˆܨ՝APsPt2OЏ<αn(롕{$BPdo9CA9Iiʄ@hm($NXALӊGЋr(5[xj/^\o(?^|.gbECx8znӉh ,'Xo^XT5vHֺ#Ֆ`:Zw%aN@!z *JPPt=iEԗW v}>ax28u2BcBvJPhܨΧDmv *cxvuC!JxФSA2'Y /7t=Tt™x$h.BlbCeZr\;LmBpbhwV*Gǭ]|HX-iTw*q>nZ9H*ֈPP[ TRPNהc_bb?}ceɿH,XI4ˉz(<#`J[}3ozFq͏C ^1npH M= oތqkd&.XǬPiEbG/4f)FD%)$eN1If (幙`8QUغ%ٸj&: ǹU1GP9i*Ȥ5 Fu$\}u!PrnTf)#;B]X ץ#:['9 )pDX=Sk{(V`CщB=zŤKE'(ۓ3c3҉CP nH}TUfr1T=0wOB .;tp%;Y3& DrlB*:NB1| ;ᕬ Hm|f9mt1:x žao\-㵼^nۗ6xil9pk: v3SC5ROSߚ5+ ߚԎⷁ{飦x9csMdɊl׬J ^n mhO bu4c^f A1ԥ>MIo06(Y $e1"mZgNDKmO<FC,{pV2Db޺h!FHՀ2e)*TkTlA]wIynFݫjXewx>n+.(_Zc;<Ct ,rdZYgx^ԬSʪ$XC7}Hdޞ|^Uj܌vg=-=JVL8Bqe "3f* If!XRL $xHPJՌLQ \Io8˾]c8 (!vȁZ2tP'dȁ{&BМgz# M'tr_b+ ,Q6dquw(.j&kވO>:K^M8#~rLUri1RF+RVڥ#ɑ]-0T.j|Hmo1ɿLsh~uSocRj֚rdaoZoV=/[s ^vdVqu~ (N)m⯪o/ =19۾z)i#k\:iӥr+q0sAib78(u Q@[@>RK-C'C H0`v1F&+: "{6䘎h*anrnO *>lΓ5{:Yؖ=:ͭ5l2mL3,B7_f_WXAZyEtLQ) *7YP|Rlů#cA]dZ+t#^z;Iإqx$RGG%4Fts{-݀%+qBIl]>7<&*@gOds ʞw8aMVu}5 VAR)`i'g*̡yӇ7S2ߧb`ll`ӺC=gk?ZM&қ]|.GWR*(RЗu|拔{i|)"o^!_62Z z9:WjfÌ| ErF@IdVT<]%^)]IC_b5eȸp_m$Xb.mG_[B0](Лg5:fP*nQ0幙5 [73)ڮ֍ _Q ,ŕ XEZ;\ٹyp:eh+ SnJm9J" @b{leTMÙ-󵺻fIB 'b4뒞:ո.zzQ:qTi? gҥql)~7sP˄]vuʜŲ!M#,:: ܪV@/gq mi)p3=xHݵ!ok5C5fkx֥ĔI7FEhݖo=teZA؞yZ*cF9:نY0$=GeV 8F?.V)>]JfbNMfPYrQ໚ȪUkz2E|#EAg^ö=qY(dӱpX}{_y(PQr0`{y_M uyߜEUf!!Nt#b랈C4b nr :"6U4Rpn tT`8& />c]o[4nٸzظn<@5bg҉uv;Ff~TEo|6|G+oϧeb|46=Rw_s7EZ;Ό:@̨svXm?Ξ_OffZi:{|d5q>SۗʝܙO%Gu"GNx9y8g}X:A9#"b Dbgb4~InYaru~=.J vc$pgT)`@8] 6\7ߺRʊ&cTL.1+$iH't2s ³ i\b+UtI5Cg i%?J)VMoxӶ94fp_J);`JO+-gx/@;a }Ѡ94C_{)hKJ'V3r5?A 8B˄ܫDטj'T{%rsJV?U]H律'lƇ!4 yԯ/7U&S:VW_f^?}VYo.i?k1ӂ[z/x,80k&|GRKnh`KtŚ)y&M;䝐 +G_} (;2~Fˋk7_i=XGnXzF4./n U۵^a&/ | cd&6EWݹg/{˟R0UBmbM ho;3!fW׬uBT"of ,L4Esr$IZ9%fZ3XQgCcښ8_A%)陞*Cl%:UI.\v$ƔHN,@r n2lmnחp.wlв(66HgD)l{filxЊ\ u=ʤ$y|\ Pn?ZFgf3vufk#y ag.qsW6֎m{r jCXZB k+śu^(CYsĖҡ,>dZeEG6F+uғRQX g;WDl1DBJ;0$Y]q[W`ujQ- ʐԲ'pHݒdfI{Ŗg:Eݒ3ˤKt1GGr:!UnֺR;wމ\p VJaz۠vnPiQ^AAOvJ'4hsB [q_AFhK6^K"FqI7ͺSK,>~ce,zg=%FZ̖sO!zG=Zg_F)v6wr(+n7Lҋ Ȩ 6'r4]/)=hcYK&}͉)r/Y"/D^flJd%S9ˌqV[5y61`$*oPō.C_4e{2rm%^bkZa(腲!ICF]$cBx U2+^;u U%g;A`2/jz胡O'%Cpf闫P?4lrf [& !r~uu5߮~"M T|V28!%i >]"ߝ)X&PQ=Qe"e(b:d9ߴ&9owyRVFƞ18[:(fYmW[29aTgQX< W L^& TE`Te3C0D׬ +@- &ϕd ˚EkVviݱy؉UhD,0lfŧS35Ӏ;u-ec 驪Z$]bqqiY\YM8M쳀|zg}g?O]/Gi-r ɣj.W^jr C>Mk]]x+ci!9l5oN(jS#)I}S^ɷՅSkwlߪIl˴$-U!j=~.'_Ayߴ6\sݤTj/ r+=~*pC ozcC#p;@Q/^}de*7O 4FOBow|Ο>WaɞGX![L-q-7ɏtiOϷ v]7pJ6VeYJ[sk]zU>'ux7];Vf2TFk,Pv&,tk*]f#*a`޻ò3ZƵ{NnlNnvG&~渉" v[Y>\eEooigT w Vhw'WsXnaEK!_?nn5Gn#9RG_` Vv&ej;7բ|S냭(Y3!Q0[kuA }Gv#T376Sк!!rmɔ.|ݐGn,uC%Oxݐ~my=Ui{jܖj+XՄ(R!r!BJMjzYJO\J Kθ7+o*r6QYQI!" ڥvd!ߖn=G6F }֭-4eR&(9ߴL7,+7Y#Mо;ww{3 ;'!Ot, _~[O=O6܈ǺyAHn+:fOz6)UۇVGX}|l\F?DqX\.I;6DYĒcI%l`Icʡ䱠|S,;`<3rgS/x4Y,P,-kf*B_&J.X֬ T9Yzڦ⻋5nqY!6CrV׿/*0 nÈvJt]>N,I7j 5G{֗w#Ȩ+p!JJk1r193nltn] !ٞʟyԏGXx7x#Cx7΁;\-nD`nI|&J90V18Cp)PA2cP+e#KX tF*–Ʉ abD5G);tXa|,0fz*.Ez8pekZl&,vX?~A3JӪ9/vd%j&g,hU0^M-(;I)Ta)yQ%x k rYe3GQnwďN@0SAyȅg;ՇЈƍÇ}*^Ib gԉ'B[NnRsH7Vl{]/)ǒt.Mg^y"i=N6LRPqk8W-Юfl\y66tE我12cT9BV{ ,i"YPzy6JȓBhŨ0-ŌbFЈ}tƹMׁ k{眠*dRJ k)gVh޸.r8x۵x6.Fyg3kD&Ţ!A\QNscC7WhUɅԨ&¥1ά*vS0-uy-Tkej7Y)^_LA"_ڴX4<3V dM[̀e>j?Τ(1SJ|N`ާ6u1hCtBR%e'a@)/v6^8cwƙ* s2g ՝ 1M`Y?kpeT+~;:HưgϞKiahd㹃o7'Jd,MX) R{fVxnuZ񨝦DX;1f QqEh:JD.PU1:T +[G)τ=T =wDVU*^in4o>!AZ -E,gL^>e]LBWq,#1ͽIU%%Jn&$4*Tc'(5[PꛋIRw>T^u\b5!1đq&^*&8+(,Bޔ.zݜw@Xq|tCT!G.IOAaVH|bF>% >β뙓tҤ}G^ sI'rP":jd;FW|[H&ka/:.d.  f5œ~I󏠬}̯7np-rouިkZ{%&ߔԬVL͏%m^Dg" } m~k-W{o"Ѥ"5&HO! $tRNpaFn4q٭CB4=Vzw_dv)AYŏa0 )1$مb[* @ JNow\X3D+w /T C% $g 48&MQFGNFNᚿR_vnAQgPf&vfɉ5!x&mMfRpX<œbv+ =zO@/J$f;8KN L`'' ɹS5AӺE;px޺;RE:8lX== "ʣM=񗈇ϋ!G[.bc "12hPpy L= eqw\,'?k77˭wjnNQH)2lti⣵ % W亾~hf]rRBHpdB~U{i ʮcX,eh;?1a6o?Fmۧgևޫ/?o`,6A㕯7bpjZ>-_|Qah_hW+>ԯ3)9a5ӛOO'뱝^>%5Jɵw4( ƒ-pڽǠցZTv Y#Cơ'7WRd5U5ZAeQ\ GsfH+dK^)2!*Zd#oK.E9@9 b6BJUT;u*JrS"jNzw\)R\bS:=Sz)-S%Nsէ Uuo{G4JOIVA0YI.kD+5'kx굴oc6+ݍ7-E T_NUg7MpHtWrX(/Vw~~o/I"aC\M~k}Qna=I^>\-*[)a*6q?UF-㹤sSd%;3EYiQ+΀nv]uD;K{[^jzw|[pa {tfڙZ/}Tz/}3csNz319qX)/S *_q{s9#E~rIC@ ڏ֐Ӓ]XIҁ`9Ma73d^H7BbvKԛ5fҼ(XQ\D+^@JB(9 ]9+ڔčOzb쉽<'aOEZ UWqedA* 0>- &bydOFn^Խ+~UJK?M{^?}Q}ZMv?͟w0#R /7 +^mWy5?|t02Ƴv2#s\/.bv_?5~"qV{CfZ/^\k\i@HO2Z0ܽnm}um{w)&2[G_B/w?tmޭ/E0-aC nonYQ*f(A`%YȲ@)s[nT8WT%(&rT* Aɼ(+clUX^F h0amA(Mgzo_ wo/7Ȇڙ{!e<Zfh5cĈnCB&cțZŰǟlAN aS9;b$[j˱8۷#*PAf障u,TprNH燭ࠇOϏq#%Gv1zbԈQԔ 㹭_…?Ӟnό}Œ]CPb0 ^#Cک}AXT;A-wQ5ϿWZ YK[ZTX-r^q d. .Tά)rS-!YH(\)Qp_n7>/HndXs<)CTFɪfJMUK-搓5u^BJ(#m1xL G!~$6Z Y+t4,L-נS߻J0A]HLx$ x IS\9 oxΑTb`979 LJ=U h&R'UJ 062+|p @P5Zk@̖eY ܔ3D-ȱª( B 2iUM%Ц&'nl!`RB%u^Br59BUW6Ŧ0'V3Дw@2 p$S= פ&-m& l5P$34qO3^0)Hmy)0\F!VXw;52!@L3ZԹ+apahPIO"JI}ٕyG(u$tԗT.Jr Q<7e$nFz?-r1w.7Q/_76tw[1Z$(6tUn/-w4◕ץ |VΗIB+%zmِxҋò/Un^elR- HN؞FoyRZ1rTUVȢ$'6[ak"!kJ!,GIAHi|>8(HOy\@0SZ֩&pІSΏx!f]Z2kEX)T)KV#T'qv|\T 6>"ЎOd:SLH}ٕڴ9@;g3 Rc6nK qAhPU(܀gY$&ʊ45ieRyœh6uf;hD5E#@KUmV90l@L0is VrJVOR*!P}yTqdzu:X"p^i 4lnf1FT.k N.EvYbzi*8gMY!ؿe^'}G xoC l63֫U]۔*ol~Njwͯ}ܬR-@9![̠NVyi {|xKo!oxv' p\s.&" :?mvPWg{b܌F֝8pƞKaoՌG!}I녏SK#'AmOXC2L/94w_) ਤŦgW)J{lb.)~` Y|dtk 7>" +SY~tRo1|0zݨJ"#9v_˱)SY3tҕxNK=XHcp܆@hn; HRDK2Τg6ef,+2hHBX]Ώu"VbStX%O !Aoϵ,;ȗ^Ieq8ȉpAg!d h@9=b#)e늬 'i04y2qr^{PnyԩUP_?|0\`j .s%s T ]* 4y JΰΎڔ?{}4W쿯.VmqFpWiq!pK姹\k.Tly|g[m:䏇ǟ:fre@< 7387C]nTFQY(A"}B*V$'U)wY k~UfMk fgZ# Zsݺӓ`AWK"A AjV#V#)p5IjWsÈ$f8 n8^wh%u8?cxn}[p2 Iu+)!u64| ϐ @<ʙ.qQrZ&Dl5R9qcE]*ta%KJ H-ٵ+&:B<6hs[âTb3D+9J_B q3)Rd[KQfy/rSNFQbyZL"@sQV˰WBֆJ`EsYqx*IwPBbNtuiHǚqݗ}ddCPdcLhL_q+H94@G$#2-~5tqIHd;q\Hxtɵ#B@ m7a)1 /,9>œx]%ĩ;] ,̥f(9<&p  CO [ZD9>gJ'IGvSV~)pU}IF Apvт,rn| 3cU6WLY(tK /dE`ٻ7n$Wr!d4OX]v?\Am%ǐdE[OC6٭XvwX,SQΊAS?Nxo{  8ϵP"t랴 v[b1Dc&TYKTY%-U`(=jm P=)2ΟHSWHA BQxLhPs>m1[Z  >ZtQZxdd" ժF97h>?V^jlTKs `%n29 PQ;''ȥt-l.``P^G/yG^ЧV+G,犖A)Yik-[|4 u}"sm,H " D ~^nl" )~Ngᐕ¤ƌJ ׻;+{5}En@L670Adhg8}zW JddhT{ Hq |<4ѓE4c<]Юhp7@|W%̈́@|@tP)!ˏ73dN B90mlùq"E1, Nͧa?1ҹ,~ooB>G#޶nyVQ`` T؅_pOG 6 9%_TZ4S89u Pa >0Pn` @DĹoo6 X5EP ?[i3`7A bQ,?5Ň6;mݟ/zBhȈ/mX o#|GƸuY,5j^я9؊) & *.mb1Bg&Of Y L U*%CcdiU•C?b)Y*c lV3 +gmI-Z\$0AY OԂ>vWrjI"9ǟDr9&RMaj3XB" xO7=˕q-o(7󅇟1x |.7 \ cB d)*Q XYЊ,rQ._MyFE}{4LaA9OCQJm~ jLւx jGhlDJc'rd ȑl\%X]R3U*@)^Iřlh+n"I;}i{ZFIM)DY7M%XhKClXh'AHSk"F@eQz٪2b J -*k7Ȣv~Ra} J#tR6uF1%}j !lIn,@xQ_K]DQكy=NJƇ4>A$kX$ȣ&|h#:`?6ɾ@!ʶsgD!n>oGQ"fA}/ ;anG pV,fh u/W K\Hi50xB n}zu2lu z8떌,_IS*Pj@਋%j_8t[WEq_f}{)$ɍ,MS,Vf^Bqi?3iYIAkH2?+~/˛\E{EL|xWKw @:76~ow,l,JWT̕5SScU` ~$o"+ʦEIFlJUG@ =Cl=%D|%1gԄ#V,vDP9 h%*DeNaˆ2bllOk:tp'A| Q{̅<8D4mٯ'gA9z`LXfTL)v&z߽Wk gZʣ`g ޫC s!)08jz{ TF}t;ZմWߴWz hC窒3]ʊ Wr]Օwfp~c].ݧo nCpy@jSaJfNr/ݿiլɧyL!~I+7̓.'xY r#N366B|l5x;Oc6A=G\h7Nzs̢E&-WbYޖWΖ[>Ip1gZ+D=Nn؈ 0G<#AZl%>8~:I-Dr~W.~>i|LwOI::6ѵxHn 4J%à\h~:j2p401QF0 np&KSlLpeU3a! yU0w=f eU2:Ut4eI$ D@|m#?aF[.쵯?w Snɒ<57p1Gz*#I~5qdhsan[ ;|f 4)y]Bo3qQT'x\}R2w:<tO[N4r+!Vm ݒۋm%ͮI[$psA E/mP%D1)Y~mݲzj⧗[?hW%vQvPcym>|]A}bA@z )tTd!_&Tw;w;ä/Reb:߈nH h-rXWnI6չ˝jyݐ)xTN7bOS9wKo[rXWn6X՜-cb,W~o~ϼ7ǣ[7Q~'-Ǐ pbdWAe;4H'NROuXKkml|c"|mœB .>+mޕt=Z%#WaNF%#2hnX? yM (Ks/{~f?9;ݲHLIOgZ;:eK 9433ARuʠ}ݽuoM5tk}#Y߽m/±JTgPD6s_iAjaxO^鋯"I BNSDA%[~"64uuv~{Ǫs^zu=b*SjW&mf'4󨘖5Ӝ8I;n#5Hi5" G= 8W$Şd$|,9oXM5s@q@DgF,j.(G[3r9>ڲg$Tى{Xv.C>y8,˺њf/7Bm~? Ro|5 Ou|ɩдo>,Ayn#y]g%H{7,kEEtԥ}{UM,jNopI:فI_#Zv28}c,v' t8'4⟯/%im+y˃ y&pMkݩ/UPp jk0֬)-W k% Acy|yJZ~$39JmTټqEn"Seq(z2k'Jgp=D'g'Q)@h/A#1$a/TțZ*騂X+Q#@9㴟.Č/o=9E>㗫.~]3FNI{8DN BVMB+ONϐI6x_%V`FNߢw "jPtOSqKh/E9gN7>tA0uC'sG vP}ٱA;ZqrS9a.pv ;şRDJNڛ'1EKMMPh5δCƙn¶+P閌L8jHo7JEi,4<;gcD᥅ cR8,]tс;@:jH)(Y7wZ4RL.n_)ÙІL@˅uݯ?Ce(~uHV(E9F(HDL;LL_LQۍІ(X{bv#ɍOY"$CtﮗP;خST &iN`{YA | 䴩-5(M Tq8+ֹwiiB G@BJl_ iZxhcmku um| @F VLPP#/y(s!1Y^56Uo׿l,'F֭@ 0 m*'Q烛dA͇Cvs~ {̯0[/~fξ'O;,|]U:܉ٲkuf[g=+{X1ѸSj"|Y'O۰$""LrN80iHTXdZI˰2{XD X0%.m y~61E):@' W_֍N|CNFZMm-yN~.rqKNn]i?I A|^`/c^3-ZooŢυ@}̶msk̵ F-{=cxb~D<]?4[N, i [!i1 K$UصiZyǞn2] ^pqQQ*\:qs5o{hT&ch0KSQ^ظS=B8|,;lh;Nqq'4lx80jZڑw^'R^H&2rFG$W|ciΖ7{/SL{7:m|# D!hiQ6º`Ck7mvU| tdNlˑK %-Clh'CIMJIΗ--i32*E|kI q(WU"-x_YIP=eޝ/[vc2+Xڀ:ta[r=b0]aeP"\(H%Q<>mQsR+D yr5wƨ52<ժi=aL 8kxtfg=+X1 Q0ZNUϨ8ntW ѤsBIt )) %K6ʗ\%R\0hT8rY˘ѼYÐu+6kL#~73НYh* ֵm7/^M<@H.}qgaΒ3.4!"f`[BRe"B 94B0 l l)NPC8A\̱ b.$VbS-I ";!oѬ2͆Q;EdƧjw޺Y ֲ2FY&d6Y P/dΓцЪhXon|JI&; yH->[H A?I/ڶu:݈to"s "} Gx2(s CB* qY/T62 KTC°:<P `}fʒ _{ʽ;_2ܾNHS=KUd/[l0Tv2 !7,!'%$9X1˼d^ /(_0׊ ti3sX}탇RN1ݺ>#VlV`$ V|]odm"_xYC#;#ϹuϸVJ`Vp5 JegQànX w##M? 3]F{ARWڎۅ$Mζ2)R ٧8 bZ(TN1,: 4"Ȍ5BL?MA ,lO6=gqTa26 k>Cv@)XB7R i8Qe'u=g@ |ـh0["&6ߡA =xE\[!RF$_yBII@ǔOmQlg3Sb`٨-8K@K/d.S>l t{cd&)PEZK '`-.)*Kxirf03^ 9K]I^uٯne8J\Rߑw/:;et>V[O3g>vYڳ zzuWOyOwR k؞j㧿|8 Wߝm>̷bMOIfm[շd~wˊ6V5>|u@wt䲗~hи qS*o[h;߼6:"Flaa..8NW/]|p{C :{@2kzxxZ2qmM{SVpÇ_{8_>.BuLSe.wٝE.>ҏSpMyquV@,:V3Sm2K5 /D"Z*le3+)s!ݵV6ρ$IbeGA1=o cp/qy\fFsJ3֞ w8?K%Nsx`쬞^k6 \ʌ;%BRLO$(A`VQsi$atX-*@JHQ1w)2 rI:Nf*_Vz 3A,Ʌ@̑bTNi7IȍJԷB{*eXm@Sx^:2:a yKw7̲0*hFA>=¥nd3AQJѐœBݠX)9q#M: +-T>TN; gJ|9o˱`7l05GT<,@|JP4_NvH|PUbgSUT8!qC3Q68dYsehNs^YsËd|YB*JLڶm<9R_kɥN}:u~}SB-BobN{:e괗^v;md8`:ߪjӪNRSN*h8e۠F}Ԟ?.8x$8v2)(kk|㧳[#3=B T<`~yrJAXYI,rEٷoq' QK:xL.4 AYڧJ0jx vMV)1iIBA+O+-4BhTu%FJB4x_Eʯu7YwK٦DܘBd1Ms iA-]p$]ݴw0 iҵ8x/3~Z*iEFт˰,eWww;þ|O߻cG5IC".07?nܰx[<΅51-[d%>̀in׻Ϧk4+ٴBMU* جt ϳSuG7%3>$f\&u ~04lj6aC Cԛ:Df6B.2su&`%?-1x8gFy | 4uZlL񞭏pìSu }|`s2JaN A `hDAQ, DB;ņ 5}Ru<{(So 6w [U!:nݟ%Y\ kmݗmm݁`!!s뾂RYkI+ ҽ^j62_*_dHnO2@ylS!2xL'bY"֌-豤o!dz[wN Sl鵄5/4%YWf "/B^9`)%FSsh =8o@h-Iݺ>KWJNξ=!R~~BPd{na6Ǐ2dlʒUìᳮķ}E.`i4`k)?(EȗB՜\\7.1xn>-NahYq1&N"T![1\JC=.έ?|8ϯӐ:yf6.l VN#?W"t=?vc+EhR ?xJ˯i7@RANZ*n!wzmsV&EJULiiդ'`k`(%CL Ut֧ a-T)Hpa<cMHJbLAb}M' մciE20ظj S&P%Iq8HL56P;p$r]i '@=@is$i$7 @99'5Vy½9Mt7сo, gY%*th /-ɕ(hb* XZopm?Ԫ,\ۏoG<kw|^gķZf"^`SpZR!4(Q8k"`W meVR6y&Ih ]a:rǾf|:GatSr\εT>M4a,4RukTPтIEQ "YUj|Ԡ\@$zs ~m S]^Ls2N/o *c ™[X~Z~=9ɄCseZ|5]q SSK__}_uq}j\Ө/Bz5sw~(fOԢ!~Mp&{}0+}aom,TĻyL:sXk ^o҆cϠ 2!~r]˂DUflprwyؿU/aA~}BbhBD g$-@wB–%iPذWJIZ 1+rX6%s:Iюy&Szw'(euů*)vbC,9'jdAtK|t8tݡ*'xn|{^d!" Io)8#hm4e3ecЅDenr[3 VhȆ2ߑ2#nHoe'F@{qhh=jj-cևx6M VQl$OBwwàRA'˹g,='_r *^8=uh[c67&ZWPT'YKh~q8c)Fk=\zą 9\4JNGƸwݿ2nJp|H+P=Sb 8,.o)e;;S~)&bO?d|=!0: .AФp5M؟_ on̸D29RSUzJIɎUh"9ɡn\/I:A;pN-PZ긬ҢtR(O9:˄cQ b2#7oPj (rL)'@j[ 5ȨNQ<:QL2.}UT- (`A|҃&s80cƖeZGì (Dm)9M Ad*{4؜5>TB8G#)mD0"ARE]Em]/D0mi 7a.]{Zu[J^.] @'oT9XlD{9(rxmW2zN[P%?&铯H?NW%`\ F QX-1ڕr9 q4c@% >mϰǙ 88OuVikc7}O[m7 ` a/)?;8)u* w'|q|V,||zppw7#,D112D{1NQOmI@ІOS|(<`+O.\*9wv?><޻1΁2&8?YhfH|~֘ѕR1{pu+n%\ iН<,hBeksϲWqڃ|)v{lAv}^mf,}ک z}c^Mx/Ձxjp, 3.[ץ]쉍lֽk9vfQ֑d^`?65ƭ2/.2jDR?o sk14=sR^^l[tJ5OP٬ j@rv|y~_h6 `}18'ʯbĸQ%B#KV x)Ѻ 6)ɒ8Ӄ$|L ՖVxz@Kv5?3(vmC7}3OɃlG@HnbŨ i*M*&@7iV-U˻AV(1")y}MICg'Xkc68`[aDC[p'~Pc=:QR[#!\nO>u"Ku KzB !́,OJJ^yD3#GX; irml(Nnx1qHw<m _r#c!lٖH [s|MW^m>dNVѕLLKR *>DN4+24lj$D GKpGVO:?áٻOJI=rt<8#@uِ@ȷt$YI@IC.{5N6]4jա2F>ppe>YAXk>(|;.0g=E&{wt42TБC[F=2Q#64 ȼƱTmp0[}37¶ODSk}E'FݭfѴ5Yto:T]6XփҺ#7XŜ9PF/ Jc @B+^ w柍㉔V|B U:~N_&g RvJyvϳ=J(6 T@jSkc0yq-2t)QȘHi hAC6)"NMz'}WINS Xw ^?nHv EpD)VK@*& b*.("W^T^9捷D(Д# Nr(Q36qe)* -sJQᤕ*%7ʴ `>YFM20(tk׈Q5GpT$|nvCLs%8e$kHI2)AJdQ@T+?XbtHXwK{č(T7fSlMb {XTxf0zJ 8! /O.kWbe|5]_>vy}=.^a],./9 Z]V~qW3wggiDM~9o חȦGX ~>HxĻygg?5b1n]_9?3g|W w)0:JVqq^ͫzyHVXYG6{՝U'/XyòuP.Ak] BJT3ЃWlq/4t _o3m|W<鵞EE-%qc qw1*$2^OL:ocſ;q~ݸ;L-~-i}c4\0Ȟ0{oFL6*!Af3wոmCBx]*!sv11'-OǺOKN)|\!dDGIBb@h O %"NQ=^,qTD)uP(CStk'T~9єM+YYBll.&:Mtԛ7MGm坕Xr*) [ 6qu #&f*} #wX&/ݧ7 _)7'{e%J怹 *d 'b Ҍ/5+ :2`|L'ۘRE{ew!ЌQ+DW9Cf "+>h@O&P(Ʌj4QpJrX#[`%W#b<6r G@WW$G|nڟ7Gq*Ve0A>?0!TŁ o?'y=;IZN =fm{֢c1s]Myqw;[ZrZu4^|4wgu'.Հ+χVˇKCLX aC`aWƻV|Lٻm,W\2SR]nSRI%"A vdٝ%٦dQEm?I|\ ֣aq=a'#$IRG;"xu\*ئ0F>:|ޤL]'6Hob\1/f_VB$%5 ~kzTb&PpåaT\8˸3*Ø%eX)nRTZ/1Ei>>ces* ġ .GX ]*Q9tPI%S%5 ]U+fEQX*Fci^W{xm@F:y1>,# ǿ(1]Dc`7:+5&V6aZ#6jj<FGdbI>sl=t 5)u)~:;͈) QMy]EW]hJz%k9G+wNֈ3D3Ĝ*J{+'K)n>;rR` j_S|:<֨Pdߤy5ڡ(%]Ο_9+93v(]͗՟yh;QʹT5C-f(A.)%+8_{X(o~_}om:Z)%"xPv}͠} (S/l)0+I:F ʈ)u[aߍs M:1\K3HU̸sz Y? 7;?A͈QMLA8x>bwIc)2x> ӱwBxv6sQn@20xߠ3w;YnΡ̙PJ+YƽKՏQb3wz5^w )~Q~y-[,٬݅( 'x:<EШd7I7L/DD4FY-dlK>|oF ܻ:#L*1YѼG]S04)|GˋbUw \6|k͒Wn}>QQD5^ifFH,ݱۯ5Q|, /KŘX\gqJ Z΂J|8E%{I3M|e:)/ŬehReJy6#-Q{R[3vΫ"ql- `L~KoRb@PJd졛>uW=Z@כꋹnkQo[IprU΁r䠆Nb"/F29ï˜]ۓBpm9H<涖z1᾿J1/Ô< Žs ?-B҂{׎*fG7rIN3pO -@2lH'lRšaTpa2Atir9ʅ< !Qtl)*R pSP&xiSRP&(Y2%)(/@2--tA:SnhsNS DN<逓B4.9đv[R2#!\p+\ӎ@Xj-: N5i@7%gEc㔔Sun:"":kBP$p-Lachdئq${NDkh\Fh`% Hd$tx#$us9 5c7& j'NcJ%?˵}ٷbilt @V]a%Xf~{ya_g]Z\~~f|qAE+VJ`!yIS+ cTg >/+jEoAjR3^>[vUo[<[~B],]T i,]9 e<nz["d<ۣb: IBS - )Q*4\WFȂpcODUm)mτ,V]6{SBYw7k+EQt k[n>֡B_W>k)HM;g{(xDž]?/!z!L5ۿs??#v觻K?{}~,sRz<28y\ߎ>Z;}o!v=q3^3!tV=S`=ӏ!q(tázk3"wBTY\tOZly<;-v@|Lp>)ط!DvОΏbL8Z;Nv,]өэW=>ĸώC>:&*Quf:֛Hk\j N}_k_'{aOz;g5fR` LMhocݖKK#y@qBػ7ԇg,(p]Yg cJ]R6@ADx}+_>{ދA1Cp=ht޺?_IO``BojnfsO%=K)=l\C?]6h?F>eqYCТTUM~=Ӿ{w^m]_uvJ} ˷?-ⷫ~$uP>-[|uD؃p+7~TiTOj=%$23mve6ՍG?d{KDٻnfvilE+İ!" ZTlHeîK ă. ŠxܓNVKꇻ۲? 7,%ur s[#%-r\Ls:%5dx5W)5MwPw1")>mT$OL$\cӬ9:Z)=VAHN8͊a<4lx")U*i1B&fR$gdU6OY lo'9ҖCX,0䥲H HGS)59ZQ*muA,JN^Nj%Rfcf* ʒF+Mf,tQ A,W/*xUKEJw8_h ՁM`w%@ 1gk&AA(J{k™_͆r*P'7!M:[ڽ>*Hw Xm3$+qDoBVn.Kܾn&D/EMI>j$nz2EZ2l;7wk*y|z=cy::B7BX4!ƸFc7" L҂# C/+5F35t Pc9'aj=Uw~:[h321!F#PA#/6F@l@XoHF_LԮ?x(nX}e5.w4U̢}4(P3yyرJnuG>}-n+We vF4ÙCf7g?1| fm<ܹ pieH{飓(KN|rマ¤[c@iThEY/nZ(w뙹]~ ;բY 4ϑ]jѣyI;RɓޯCsX2V>K%.od\I3[o5ۢTU Wɜ!b[5i~@B% ḡ ʒ6,9;k)aq̛MvY d'FQ0LV9 /<AR 4د44mfwɠE$XtTƈR*,AK_i,uacK-MG6G"9M9\SmJRb Rq"fNTFwn 2tUvεiwu2mG"ЉGއo8 ëI6ݰ{QK-3R`TeACUYá"ea9)D 22o-)N7eNow2TSx597FyHs{v^#F{/jdx9{RQ0(BaݧPTX,4Tۄ3@[BF?޼Oߧ)1;Qӆ#oz^ n~^z{M[s*uh` ~F3ӸUP0lj/@wBfoj`H< pn0laѥ]9&ܩuccnu{#E;=QZx<|:f:'>n7柗v^ .tdfj>ZNjգ[Զ_;?U:QGOI& B E{I:d*s{BS+!;F͋qZ@iDUNkL# U^q͹13s5%./&H!mj'{ ɒ+l)-ᄹqM8FDj4AP*A!A[uuh !)o;ٌh{ !]2C.j:SPrp8~4/4% .]IL(o`}T 0Ayji nIhG=c^P/5 BIYwmܑq-N^)OQIQf77a97IZ&G2Xv?.5N.>W@g\[d:N4d) MvEҰ!;4@)/|ihCKMph4M&FK4thArC>;!B#>]} ɠX?=1azg-tPcp>Gqyٻ!NSx;.6׻6vqOm؞!rDL43OIҹ^4s- 8Y4Vݟ;EYSp0l읢x2ia%Zӵa;C"w*8LK Q,dFǒi BZ V$g>Dnp! ^}:Ⱥј#Ka5(9XQ_V940Ec 88q&$1D&b%O!ǵH0%Jl/*db 9-cIQʁƂ#*1:>O ?[f9O7 !9Sءy- ˁH"R>䖂"\#V7Rx^0PKO Ms7̟kXZ )eݛ. yG:e^ee54#?X3iZ/ 7<;k7 ܧyn0w;G7S 7i[UfR|9$a_7Uz3!hig1L^Jr<0;n̡DpwG|~rrN pt*tOn#MRC}g/gt2iez][T\B.sP8sLm$|v>K_? O2rs{3ؙم׏Zo3yOcЊSZt@օA[suEE}_=|cǖ]W`wYwZBŭߧ:չ hl˒a^ ]ڮze6}ݙBs nҘDjJث?esZL|%$\Ԍ.JF~NUnEj"҇9Tv0LjرװFت~F*hH,Kpr'k25ŲC cȈ{v?RR(It--[{Fzšq +ݥl 0deGb0?ƀ!E'<3ɞz^pAX:7  rA31Z2bK&[7:C3l-6vxP<޻yoM:Ff =$HJ@=#? Čyjeul.zsi.pLo.n>-G]#=D}O Yc:F@H){vyk` 9Gn„01@3'D*2 ;wmDCnQ !˯AR(x 5ódb_˗+ˍmV?+Dc?L Ҹ`|اL9Մb}I9 V)y՞,֌ t\CtI4mل .:>%ںJjVS/N3 Ky ͝bJ*8}4db qO$ol`!Xw%uۣyaﴴ3wPzm +'gɬ  S(Y>}=iAfr^գyZx+n[[̝֢v  V-(/ύ%;v5VvFAΤ䧈>7<׫p~Ot!0d8Y\K@( K*@Yn֙J0 QP,-yZX \[rB(׷oe-/F?~]{Y3JqĶJ:lS3zNcd!UcZ! 智k "oiFc~o50>UAQڷm\):@#l1Γ737F"۔Ȼ]zN&1pTp)&Y17ӑh/FOL7_8+4eS' ȬJ\p@E^JȎQ%hBш"bL# Uv?$'vI.$ Pu _dUsVηlOP.ANAeyB(="ї4OI3 c=.!N~l/:nOL%H1ʍ)0-@-aI!ӴHkyBQb5b1ߊzʋw&?sLayn™뒞%*+}6VJ p.v`# c3G-nĬ䐯~-;w[./[| *oM:>T!<D`IbCI`Jd!ȼxMqõU/ϭ@)f F6}7ƫ#m;[ı>}{^?~3C=EXrLTH n`@"0D(k r;G?OY-=%Gkhot] nuX?/}S3Ӹ~@~UZځ k Bkt@[ylzQ2!&p@rעM@D3^U)5Zb+b ƸQ Bvj![=!0$nat^-,ڙSr'c~d<-, 8<{cHT Q컙I. km#kw~Ɍd!Ckl<Մ{^L{I4FA$n%@ Q[^:r"{zUؑ | jbyɯ1/phLVbԭR֋:X,V @$vZqt &SJ|jeO[e7w-H(˞ #r`?wn;J`(0c{ghv%nn7Ѹr~&pdfj$I,ctLj7=` BZnpZN3=ɛx &!omKv^1weWAEB1fZ9fEy҃zH$^{lP]1E3ބ~ 6 F)uHgGהL&,B M!r6fpٚ72 rvtʣ3H\ !"朁!fO ڟ9=);Pw&FKZ @׌4'1B%!wUiwҁТ7MGLe1ז}>*sqAJ&=9ZZ&|]"eQ'{uW#yYpWuyZri_=GryL3:8!V󝻊?k{EiYEKO/>#ZdGB_.:^~9% Mml]uO@5Hnꐨך!^#u e3B琔\3"dXA_"qo߽} }wчBj j\/E ^o!Qi^9}`0cG+, $;4]Q|QUkzApeFr t@E+|PUF VƖEhb_=Dj$R>}pRSGJ2!,5I/2iS~$#x3HӉeNP2)cVݯQbFk';[z,ij"-$mB<#c ,2v9XX E ag?T гX$XTmf`!USF%?d`XEn*-vŔ+Zs,Kx/E/Sp}|.ʼE@e9?"cA@,\dBO)POZ1R*Z9H4&,%<Ǹ3Sf(t[2v}ia3u;Bkp)| :Ԑ@-; P{c -ȥ li-1jT@3kO"PGYF|_|>^)ይe^~,"[ӆ'/Kа2'sT_433)DPHEQ޲v5>G-?rrp UG0̘9 7V;BP68[֥v!+r;I9K.Mb_A5%Q<2<%VWT -8%>@CP?)w7W.N}*p[OMH2g- py`/9}%DLzE2$ +o^C2+WdUm7NĎvFV .K!ӿMXȪ^Tܤ _(˻v`ĂdQUvA *ۡϞl,Utv# Q|yS2r^Kޟ@{9U?eN'ZɄ/(IWkPks'JGѻiIӟ3h,0JZ%U,TdoY*'\xؿ4>s0cc -)R9AX;S^`Y9/s\?Zѕ֔Z,:6Ϧ- ꃧNIzs)V{0T_466IsZڠ:|FQ\Y:v)[˹Pa: !2eK!nh PR}FX&0N$|1,Y (BJYHمEpdHՇP F*kE6?] N-RE@ AvHO*֠¦h0@mȀ0LZN,w}XEoQKmXs4lTHL}!`ʽM|ΊBx!-^ b`h I]G|*ݧ'_e5쎩YAb&jLP: (p;%KX/w<"nt]V6UY{1Mewo Ue ye\]o.jd#1(N?h&I;Qn݃Zhl4 Kd9tǵ@6Af!>]M=Ofyݬysy= VL7ײeIk}zr)&c$*5ש] ة];Wk8%-}}܎,;)}w5#s7{w u&wMHHS;ku{PփO4l3%s3y8*Cf$'@1A#HxxѠ1Un ~Nٷ'v'okۮ0eW71<濯<=l_j,,'a[Mx+>_ՏBr{@'T'WہR6}i~+5!$C(f{͂]MK4;W^:EZ7&֗)m| QQYRm UN!/=n^u˃6mH)d֭@uBCsS~\;SR zC&mѫ=\w$zu"z0ɵ߷ۦyzss<=&=,qw>}m"LݧE$}\ w=9/T[UQ~: ع!M餌kSOnG+\9y֋&9|2H}O"]%qW_%ZB鳈bz1byd#$ !כRQ(j1]kLJx8N/A*Mj!_\MZ a.B +!TB+3u8W>aOF'MV;9L&m*[ Iӑo a e6/cpP;ce/ڳ\6lzIH~ XyakȨgB"qI'䖅D;sBx#j %1/A=e1]>Tiwqq֔R2s1DXD.&qc,2$څŀAQPS GM5VȌK(6v,0$i73Z亂LQb/bF9zWr\2ؒ1DgN"|ZB'\LmQj͚m/^؆(TUا"T"j{{yu-cCNcdβ1Zh*OT`Ϲ,Be2!Y yaB"=zf*|=R y˼C,d6X!iXxlXSp- AƵ8C( XA(k6͓5R u+ŝvT= [81_T1)svܸP/'CZL\&.M,b &D3H*-4|(9یȰ٧ 0 (Q@ j!n;t| RѮ3U/K=zEvZkX.e;%٢ quK\-Cjgҡ@*ɠ@HKO]員@OO.d F֚Z2_O$8\ x4~U;L 5Z0hToҀe4X!y쪷 PޡZPgOXHC= U;(+mTǛT[Wg-\U{SMU$0"sT_4CO/R`~ Rׄ2}&D$"u !߹)3Zwilh-eMF#eI6$Pi!xvgZpw@D+St2FK&B9"rY2( l1vf8.Bx%FD>u17nj^iː a^YaƪL\[Yd-yʆC'ɮ(đj,PY('JŤ#Ld"IZ`-lm gpir+tmݙض}zfw-QI6mDH4x ̤E L\4%2q}gȥ &zQ yh~sYcHr O\;M_ K"Tu I-ZF=\@Gu"zDO)r`( Uޭk3W9PCX$:3?[ T_46ZgvCb{\R#"T4-6YKOZKi5@ݴc=kJ碴ZfyI[ȃ)@4@ŁEjLliC0TTI}['(l1C -DNYK=LZZßOKI`ZzZ*UJUeOdQ}ѠZ qO>q-MsQK-EhsT_4VB㟶NRk`sjsT_4&.Ҵ v2ҴiGuT&VɪdIԊFe?>~F2CgG- 1hN mCuއ,ٔ3+}83,@S r-I,- ,`iKuԒpLV 7?zϻeYd))-[ey!:[vܬ/|K@oXD7guQpuGn9˦lHZvq$."61?L-{ ٧*䎆yS)xJKwtWw?Vpdt"NVd@~mI$;v.ڃ Unt9POr8fGE:GTt%qc33!qڳBC%yO|,qęVPȕ*?!Ԗ(i~vov6H%2QElkv %GL6&.ɎuBwr lLL>J3S;d !JChҦCsrgIRJ8s>(ek ]nTLz`gzȑ_rp0%/,N 2/0$;Fd6Y?Ŗ,df/JXfdɪ2;uYiVJȾ#x`i+TDbF : ` n?3΢ O )HAtZZaQj)")s/BZl21NxA$)UWC0!Z "9G\OR%$) 0o&.J]I^ki=E_lCcIr:.˸Cu|.co|Y&wwe;!%3+}>FːGnsr2/`2΀1DeK8Z-Q`:*Ks G:M* 5zfTwU:G 9G:h kK;~5 OaxZbV_CMTG?ʁ*W&j^Bk9.9$* BN,NԽNIYcTWSg 8v\#Rڋxf J[4("HI&GPH]kNSOXOĉ|e=1j&{bDz'"k)?"Ԅw;+җ4xXҗbTsM~sÏ#DkuWutPVVu׿,ԮuK8!QvukLF&O)śe &hHFuʀ3۸5ҏnk'=/'7hqxRKN ئ<%Q>#-tT#+a7"-<q[FW^HZb0y@?k[>Byx~~pU(wΠv Aٖz7mSH*/Fb@8R\:ezRJPMdǂbš(Ƙ*f4ALa&`5`5&rnBu͐C*z{n* _FJj~~G~~N\-WFO "Iqa _~{ׇinXnha=0B0۽۟4b 1feSq Ɂ0e EnL]ctJŅߧA>D >{襄ŃUݠ(8'K5D^RZPFќUwTԞYe`O=e5wr/tnϧe _$;$nh}N"8X߼- H%K14e L8=:INPfɑ6YBpJ+SѬX+G'qD T5|I܀/ L;Dk΅>b#D@ﺝ1:vRtR"SI3p3=nz3Ru&pөpn|h$t¡zZQhabd)iυ,E<A"NeLx Ze!`5Ղl]v+Us_X$l8+nODW4hݸ{2Qu{L~zs nj\:z86i-F-L +U a !tGnI/aũr)"ƴ+NPhYD9/Qa*w諸X%=}Z3Ɏ؃tgݟ?)er@ v_$fwco¹j֛gaPlQN8}=U &J4 JIƸR4LW35aԋFv&L__Qfk AJvH>sT(L)0gSM)ÜST[܌1Rl1=E> .hf_,Ms#t$'WM`\̌g{rmsM7#X&m\jqx5j9[-o1\*\nCL_|`>x3^>*eƽv  {Ep(8fI:un+.L[xE ?K,a?b8k<(3 w$sلaK7@7o10a5\H.X—WR cZp _sW&@-{+[Ջ,G6B Y~?͵ZyAk]jbHK{'-FŃ "Y>X{ЅB*Aj0zY}%ttޣ' ~mJOYaN<(8j|O`Ctk o"ϰ $q$ S$e&A#my%af\$S$T^^p/ d!.ʽ;s_G)qL9rs7K~pkU"t6, ntLI\]*ě#I?fKgHIȅl RvWZJΛx.π}$= wuF4Ğ7gt!t'1 ~.b1k{oCQ ӖVw2FMA}MAǢ ^zrQxmti=,F-A$esܥQ[s2(m.iIF{TNù=\'IDk9+/pU$J5v\Hh;g6viݝf.]U?Nȸ3r֐ c,Ґ= Yn .KgWs\۵}R݂T+q%q 8,ʢ4JZ׃&G#ݠy(JXHþF!MPrDeS3QkhD}Zt+Y̓휗UEJS|zQpj,MZm^,D U\!8(Z`v(S|}A!>T, EGQ[IH8i+-9RIϴ&*K3^좋BUmRQL Ek z74&f,*1sE_QfwG1 +$x }W.s[fX"ԐŤdu:n"!9f1J'VhD q@r0Ht2'Mg- Q;O_Id&#L`LS(Zpxγ`rdv[OW룑.ܯa?BΰJZd\6Jj}(J_ OLwgGhV9ܥ:}Kf͹)zo^ 1煘|?T"HJ!cPURWdRWW{. %jjjMSVZj)kl/ϴ]kP(9 o:x]G]])YR9a5SvO& OCƈԽj]&"T͔:ay ˒4t@@w w(`IQEz_TbVY:8B]HdPRx3L̢F6T)`j'4'H@jr#xqv']HR R z]\baj<w.D2){ HEVTSb-|}lA[g$tc=td ::y:^N`_ol|}C/7K҃CYD`p:ۡʙGP ~|TB`yjѮ;)>_XF ,@/=Gs滱7at>{|`ύ YJ M)J6aA,sfV`8< vṁQ:4=  c0ܴ^8-,`Cp|; ,V&`{Dv@#،0K?x-~7q13 {rL7#XKNc.OZff4^}oZVLÇrv3b"%>x3^>qoig'} `^ AEذd%}2gx> Ӱ"xa?5jU 3'O({3jK7A, ߼_Ć~TBO(o:yohb5Y[-U{uX40"~z4[}qۂ@RɆQ2m،zJm<74B A_0R)Q̓4u8|Jֲi)r 9zrC-,m2pOS%SG i=>p ̂0y 4T*FWvn)UpTb5)+l,)KRUJ 6i3x9 gaG{"sR=F_r6ܻ"}0;pIz/@Jי z*`Tv)o;C`.`C}yi{*&l1CBs[Wq|h}f`&1 #K (qJa0ާgښ۸_aiυ4n 4\}X[J9ǩ\ΓS) h#K E99:}GC fD%G3F_cUFv7RVŗéglؙt簍{Q\^ʮ¾1,jq3J ;w-h/ϱ~pQ1( Ё ]BFB%e@Ec ﶗrC"kizC;#_gUd{xnF }vc%%!=e֨ QXeOF3O3cӱ,lQ) 9*s6Jqa/CN8i5 r|PHh h90JDWl.Q}G:J|S]$VDmOՈ}A:+ Gv2z Q;g7pN Ld31:L3LC7|n}$`|RG7Xhj\F_:GR!@zL[7p &G`QdYJlwY}J١T7y&v:;Y;ՌRIM s<+=&R{giyzSuj~iv?D#jn>s6Jқ?_-rd/4b| gi~BH`o|i:vyvq0TqH]~25moo0K[X$!/\Dd^?3zCyTBb#:Hn{$䅋c4N(Qw8_S]ccްk,&$f$kN=oPB^CuIXsC;~=~bݚE-,u7K[0)]ɖzLb S`JKyﱶ$a#Єc3Suڄ [O]ZE.]ܬEȍ6I?{U^1ff,A^&̯=ː.~Gbqku(\BQf(%&ۨI0dpgk25̴r@B+7%Z*C ӖoBѰݟKch9kģn<|'t P5;H;F.VhY|Q> {qΓ,WY?zgRҽ0Zzrnyo֝LAO0HL6*['e@Q:NUX e:-5•-F Q:!zIŒB㱻YD:Y (wQU܂fRGmML  xp6"A s,CآB%hBG܃uw[!Wr;Luqq]pl=ӊ^Ꙓ8,dLnʇ$p$cpn(sa-SpTjlϷ"`I[G" QQ3sz`UL0Щ( Sȧ1oh sHn-QOQ&i8`Q1rRYxn&DX~#:U`UJ $)CPLU,Ʒ% E chi6M*MOtm#U^Kֿ$ f@^t?݉KѠ0vJ=ݩ<Ev?͋ǼPP.h~K)'6}'^9Nn;kXu}ga;>V]'oy_Wҧן?>{ ?$ vSX#Fl්{j(Ҟ/ۃh솴;;Q dϖϠ{Q vh"{-^$%˧s-'P]n,ZM3ɧ3Pi7&5v|g77=.Dh&>n`&prQi7TE`QfC`ҌI>҅tRF8K<=T-DKKSHKK/}|= i<:ĀMd3)[B<iogU1cIA\X*a*@BQ"Ycya,Z9jBWgƱ?؀12?RHH?g#%CD#RInj(()%䊕8$V],FQ,e{BtoF۳\43+%jS59R(nfm%:)JQ~7il4a # f+-!Q*]I9aM?\9B+m5t_*88F3B(v1 )i U1jAݔq bvmCvy;O %1#PDKs bV~w!D,J1/unh1RO8? ; ]/k{6pF8r:8 5\㗠뉬Hr JDID(#EZj@3Z[[2ZRTBXa˖ƔOx `xZ9._6΀Pa!9`:h Z0nVE8R!/wB(PTLVTDYe"r^Ќv SVPR **`s, IrQeExaE Sh~qNgRG /ˆv^ RmQ9-?'C@>b߻1C[X1{"FqŠF\JpN[ʱͦCP8EQH"5GfjYIn"Qd3T &tFL,UA+G7J/, - NC+9͂yX ¥h@±* 1+*/FQ#G ۏ4=4Us)ji17a[Es~zRޭ]o_F̘~;ÿw~_')];wSh#LN әB߿=a~XoMNHoh'E3ƼD˳ɥ]|`t&*tnNPֺ^cՕUX׶yFR|qoDxp6-x{t$p9.| X}~6l>>(5U(F)G}0AzSH*Αm/+WESxw+8OQ}A5ԌScy խ2b*0,+DK!;,VWS5IӎFEzؑU:K8oTfP̛3&V^CyG7hVC]#>o?}> Lvx@;96N> !s\0`^$O o)Aztv (P^p?hx7R"*ӈD)BEx" 3mq:I͍=  FDƇov*ғ,Ǘ,.)F2 9H]B`@$gg[  $!'ylBivÒ"ܠ:2lF¬SsݩP\p 2s |Cwt̀vO<3ݭj{mNOxp7qvŞ/;cl5?&__]5]lNj[V| sQ\'3|dhn[9ѹ8JBIFvJw5JUύ.{M#^l5M+ -ݪŞɕ߇Oϭ;ºogH ^mAAgkv8?#dϛTSQ3vZz罞yTOxu.: h~c\σcfWc9~keZ >g~M?Ɗ嵻]~q8DG)T]V#x7rk{?~ו *ߓ!5V;=I Z'Cxmh7Fv Ftr XS]B1 bߎj$Z 7pbV`x>j%=cW cɱfSB|\*&81 GG-d䊂Go#(P> Nr0 ,Ʃa՝ R+@}b", @@(WүwF5$QGpeE\JW8rMр 0%'¥Լ^5 kҐDOOZGudӌL # ʥaF+Ǽي"LQDY UIIʪZT?x,QUyn~s/_$gQ@)PT}0]:&#AѓyϩcWiP|as<&`hpBUvS<]؋7?yͮ>.؆%te3,F= ; LJǍ^$L i'YIP<$/N^,yꝉmXɜ֖RZJ,$f IK`3(,E o)9yr{%26zM1y`E1y$ GH2a'~{ gPL# = s47C#QI-n%gL0{z-'o'~ /xfQO' V/뫞"W_m1,ϗĈ #- \>_+`*2]USZf6-Sotg%ϧ뇲<"H`BNLaAח߬b 'f'*<ŵ~XX~}tM-f:ƹjW [PI%RV1(+f%FTk11eUwV{Ea8=yCA7׋㻠zo.+m *1d w1帿0QY(94W$ ߞTPg-QQ8FmB Be'3=eАo\Et {[!d_Re8F viS_dF33h7):xO9"Ý/?1$2O&\>V_tU{mӚeAOߦ9eiKcUU0eQR) LAK ,?yDAdh1܊HmxX| qmcn >O7~3_ɂqCڕoAY)Ȝ@3N(x ң:Y[gp~GH"u;iy6yQi y-ˆ=yj bt^ٞ|q%2y%zr)`Zy 7ٯ 1(T܅c] s4h0>&ʰT($%^e9 qUHbD]3_7$ש8Ӓ`IPN UU\)& a f!Trj0x>R̫ %-ndRBeI5Ă7Xq8n+TMD/7zI~]NQgˀę׶: ?뫇ӫ"zizԯWޥxԣG[Fa(:]0Ъ2%HVJ;ͽvjA7&A^{dZzM%YKOQKi p!՗TSĂg-=I-tw'-0neZdJ3Ɩ-2^JR #jO50{ 1A F\2r-8+#x:5+jDLQdǺuM AgӛpdD@\]qFܤZ2*ZzZYbV㍂uCj!/򬥧i) rj%;N[KRR5^/=m-8cd~)8-]P:ϥ's)I]h:譚k* QPYBqρJ$6#眫Ji*J -Xa!+DKA%srRʕ_I60*z~ƑO e .ٌHT$.dQ&I?]Ie7on+jّa\PAg&SC|% _i0^ 1pw89[A}=αjM^/sOԦ .!Br[Ŵ& y@bC57D` 6 9O rFe0Y5)0DS0%G3*+2/bkS2yB갴){<vs]}Lq| D=6azS^:f nܣvqV$# h((zˆ0w cg!]W_W]QzTYXt7fkS{=u90$FQ7KD=yx_Aj 5z,hz__>;Ȓ Q%8"8Oac4q|涏OF9lLvMy싽c| F =E{oK2J(茒6P_ll<:_n uѤxQc H^Byd]ʋbܵz8T^4=^;Ua|rbG<"zlީtzJ=^3ybFxOÓ 7`i0u U$cQ* k*̔)a^N&s uFƂ;y>+fӫ/hZ&$fD@fIxR$Ȳk20F ҳF!![ [XZ/L5} ZsEͼpw_R d v_[P7ofC}vjlD,NԂ|>2D2]+Y_{:s² ׍O`wwK;jAY2,8nY1m mR9g 9)JQ0JP$5ǣ6} <}o"|>\ϣ3ˇRE}kxLPnzE<gl(l}"TtBXLZ7&eGƁ ;׎ =pD{=GP]%c2(J9[Q ulrg&C I^}HNKaN!"|h5N >?CS]/ӄ۲{Z=}a?/={1W~7=}~I.|Ղ-^cPz}Aa◿ t6_MC?Ү FH 0Z7?=e>B %ʿ~%}/YB/ ,P*~ A2J6V%MMHM!Ӵ_/S8=Zٯ[\&TеDǿ]L'>7;oުM}  sB K*Q0IuP T8O{_W뛫j׵ ^Ǽj *y.Õ?D&ՊmLgW\yyiB;w s&&8GۏNB]ӧ$)*Ha&a?y-p:4WUF!HWbZ&ZT$r yZNJAI p}ե9; rS%vM)x^}R)1Nj@ܰqҖY{3iʑ\":0!ϏL{2VShC5B Tk $sLK'שׁ Ȑl&>+4KEtVD[^I++*G~AS k> CR_mK@z0YIc>>Tj[j7-Z)iVDM:2fԆ/}VuzOJN(5'i+M[:)'aiu+"doJeskQC,.e-iً=$m!r|ߝ H%BA(.<~]IMYZo\ x2I]2Q~9'r KpM$eicK`@KrpI7tI!4'-i#Tg_ihz_eU/./ɬ6h- B-:uhM\ ,J q5%u3"*RmZsAӂ !^\ ʸ 8}K;.o,f ]a&aǜel u/h|nВ3*T\R"B_`AQ9*UodPZ;KLP88ϖ;N.j}م$J P\1PC!4U@ 8XvA[؅  TEԀ2W꒗Nӆ ( ƳW')Pi/FIs&FYa+ţrPyAJJŔ,VISa_Q[o3@<,̆ "z/qUҧm 8OY #YnFR(3޿5*6TwpW %"9p CX+ WdѼ+*Do^U X&hh`2NxWz ^K ҝTZeKfIwib=6m &([SܵK*^WAhIW@vkuX NQ m 0eIE cC˚P2`gϹ?+D%=[,#@4e#ACڵ 8GT\9)^Ux&xpo[}iWqDq] ^scHF,dSZۅ=8DĈfU @.-S@+&{\;qCF6D"Jf KR##eS !`{LlI8g#{ZTދɴ GמLj{6Ua-If-٠X (XK=Z7G @^v=mMXM.|Yv9}sjH-[8088ЊZ'IJHU.⹘&+y^[u$}6R[5ND*ն+f'\-f,թD+mqii+zk?֘` Zg#hStt؃M3FP [[qέŝYwE߸^3~`!{\\ҳw 荞7F rV e]*LJ܋#.yLmTqITœT1URTþ?/浤q+w hYWgI`ژ2 [tEP=<SpG滫PR'Imt b!2,DhsPX#'XvSJֹR!54f8Gz#hǰƪ}{ᐌIA{%r]쵓^awb g:]>| D 6I(Bx?ӄd@Ob||z%V[I,$F oVrVWyҪ06H W;Ow{Ix6 j)6:'*cy xn)bI̞6ޤrf9~ 4d5,WWt`W )}=[%_ ;''j  b1!Qh$ⲱSؒ)w˦[ %l@5[XUe{h 9zlٻ)j&.4N~&2nEfg^gqⓝ_޸U(#)/bvy!lnMy5_ :Nq¿nNMjg=G'ޖ@IU( 3ZW$8{RUD GZ1FIvm(9΃C 9\\^҈ b2V4F 12'UbW$HyAc˃6Hsj`B"wb =mLr|/;y˧WyiUAéyK1;HרfiS:zm-hfNB[=Ta*@[P Qg'x蠘%X\@) Ζ$ ot<~:e-T4 PۛF.,նt=TZq[мl[z{'9M=h`à$1]fN8a&7P$-K0nuS) -U(JnmӒlɞÔdL^.Qo߂0ǫbsf>z&dOoO9l5 /Go T} )B.H%w's@@2/Ոˏo  HT#t(:*;kY]f.5*1c, e?8Fc9FE.S-KKϤו AGm|Ig CK!~Dq ee6ـ e\{LdY=2-@!Mi6ZEz쑄6  }Yqx61ip,< +' R8Wiwޗ\z(#xńa48GUfUQZz@Z焢UD\7qQE!A+(eol~Bk_*D R8ֆW:! WUDYqMRJVʖ[Fm.Z[T[%f+9Gj&`H@{ę\ |zh}Ѩ6Kr&,7{ )P("[nsű,TqwuLwRY3l`Z4a^Ò@zczvQ~6 l6d~ б4R3wrpJ)ҡo'Z4( z,m i8=١EP2yiM7:B-^M܀F7v&eLw(tv2\3QLd>QL@:~xcOA?΋7JP̮h}d:xii-61baQ"m1}Q\5S,.Nnogr)x7*VϩXAw_?ܥ'ICnDެ"R];A|ŮOiE`zIH?{Wȍ/^|+L6f&i;k[^Idq-jɒV($#,>ɪ"YU/E iȀ!G?@l5yC.sS㊔)L B-)WvjfeUNy3mH=lv)+WC·CWŝ6W c>p| NRMeD˗דKffxwR@_qKu]:qKQw%_ZX퓻CjJ!Q9gN3 &1)Emyw7(C=(\hw)" ]MMg%:RԂt6_nn.O4+9+T-d_(6ِbEMW==3%TбtPsЕ~H~ 4Z{t湪lC2ĺBcA8x.FQk 8@C TA4 ANmP/LYM  64KK)Ԇ^Qk׉;F})Tێ8[KNv xj ަh\C2Pp&<$dPp4P[m%U8᭻ㆋTՃݶO| 6ǀ=X?LFښ.;Y-w1d4?>BOa/#%T F #q=zξx'5eùL-gINuOQGD\EZGD{Zmd8 ߡyԭ/GݶiHW|>51*5@;5WhV5J^:Dł m3ӅȄq!ҋ͸^ /)/ ǟ'4Dϗ_# m s\&-UV•;OݴMNݴxz"InZVd}d@U2t IEE^ٷ$[5RPq6VhzzL|PHr}P0Zqh9rro=d#%+%dQ P- !zjm ω-x(hAs!14P3lL:J*-q1*)Vދ1 uŵLrԔiFykJeS r᷈X[R32*mh.nz~6|O2~w/E}/PL'@ B(T^kMA=!5@MV'Պ8:8 Qkc~*#>Ze: ~ZmDlԔ+,ϯfEk׸G5B:% ;bHcV;cHCC35y'H ԋ!5 U ke%rSm\,Y9Jdn3dzU=僣}f{tON?u~l42IYNm7<ϽORSFѸ6ʥ %$'CU7J\E+Qu:) wxrkQ(IT/'bg+|h<ܸdVSc >mv~-hBءr7BGg6?F"Y$pq,6KzQ/ aO~4 &h tsa1z“x*KZ!!>(IS9Gz.r)&icP(G[ 8ʢ, j5^ h(BQOcl:hB7E^"N+zP,r./XA4U8Eޡ#!Voj\5u` մsie(.ޑwa qu>1sr㶷R# rs ı4QiְI)͡ĄD1d~=Uh0A^.mIu=od\.e=a7y?9j~vvzv1NƔ™Ʊsa/sCKУ뻼{7Yo\+`dץBBjooǿ'Pz6nP!!gұ\( hc}y'Fz*4Dnチx[35@50iZ1Aa85z|}#*#t"f8)kdhf0SH5H4V1fƒSI|ב'P+ij&Z 9`57фkcNX‚S<0)A ꉠjSDp gKh=|Y׉񫫊a}y1^|*m<^>Dau?wIPY>-ɲ|:fi(-cU{~ ]4^mPo˞OK>G;V>6v].-V`&sBT'6Bf릈_^x%'R G.5u rd/&b-&A8*ܳĞr}ܔKR:(.*E1Ǜ?[{5'ݎU`bLe! ~qoW{;cnǟ8wf˻[t0+{1~fV4`TV>E$ d]61v9vB|9S^B3 6sBlqJR L+)/ݝ0!+qDÎ6!*03T /' 2&JZ&P~SO==%c,65Wth5" %P":Nz\W6Q7T sJ;M\:f]efpv8;[A)%;󭈫QAwZ6,/2#am-"\F'a=?| EY/to1B@jn/$ͮM$z?I?Wpβ˛InoOCqeG⮆>s>؇y>7 ``n.#a{Lf~1r 6 @Ȍ&.d1sh2ߛl;[A4Sx}hvN@AeLpL>26ǻ"ha7{f|ےȊwġxdd_>$r㒻d}gG| r,Om^|ܭ9׷4t$~An8\e*W!xhR5\bOL0 2ݨ^lqfw*K+HG_r 7r;Pς|Y*icB}qoD)ec];z=1\]N֠'R}J5B^{G,-5EP"a La \ӠgLequ̷ɳƱ`gdLpCJ\RPpeE$屾LXcp"W Z@4*]ȷ翃Xjjw忋=#LkN\Hɝ~AN1F+mڑ a8R^(h(I͜WLQ<ϜX0b@0U/sro+W`$5=!cSDLDdeQ0a>ɊV>ipe̺.)@uzѠJv>9=Yzf+Yw7o"akO[I۵&{iN'yVq1m&,SaW|]Bk\I{τ|yo*gr^ZìPW1a̬#.BlG>khV6^ ! #Gq(x~N  :&Lscl1o;;u^GSd_/pr3 jIU0ղ8HqHIDXs](0/P|³*u X/t3y͉ U{)i@<5$DF2 (%y@>\?L?'M P:ù9D"b -UP*X>ZX+r#ij|wL4]Z_6\^o4ٸٻ6$W|p_,zcb5#|̋j 6pqNoVjY E6_ULqZPy\wv%b]#Wx=!ޏG3GP`e?.QUPTyum+[zd,7ff ~S p`B% Y-n3kL۰e2]h?.$'\S 2D;búU77oojo-POl-"RHyc7|c[@пDȆ)CJ@:`XwS˃'W1tv5x;eH܌˙L6 X?-W!?t\tq*eƃ?Q#'Fo=. eᲒiu ҥ-yx_*6o~U1  vH'.:'D+JH"I#j<vN zT)Kn]I Cbc]ԛXœ#xAR1I廒ӓkVab t VQEn%Ijs<'n'X"uƋ=NWDd7iUt(k:Kon&Z.#Q(!bJjBJ^NnU]+%uÜ"9(ʞy;UUq =Z% ͏wzs^3y}1JcɊ/KXhl9Y흙]#E3J~TC.L7fQ2XDd$eR9ĕ6.7iάCZ8̜R ksx(5ץ踛Z`6}4{a`Ƀm]-~Q+T"#3Ӻ|Hxcke,O%n0<KmNGAlU%ۋ|*2j ~ >AtJ(/22I "7Bde 5q3@Tj Z\$V$viiR ;})cO9;gja 8?0.]`ҽ3$?7&h)d8b!L$2=ߐ9<߈K)W8Q(\4I%N䐎5ϐ2 F W>v4O> 'ﲛMeˣY2 )h^l-z+j;MmSz_l4 ɘlb LRbd4sB\&nt8V% yQ jղ />!/I w<{A)ԡ?]ɢayg.Ͱ|ͥ& 7ȿ:'QWɼ+O[,ӎvOk]T$jk% 0jxMx@Lᖷޯ)PvۥLw%T*] xG RU9 UbOi^'pBY|<7M (H&tM֛mل/4Q+ ,O9JS!rFQ(;ɍH:؋"b.˅U iKrowW菫l6-в_t5Xv 3Xl2W|ܟwaR(UY$igc7grgH6H思FPˤWb1I-A2,L@Ynh&%c4`XW`H1$C"K-KUΐ ԁY!4fRP"Mαę3ipԣNvnVCǖHTR01g,Z0簟G ,R!L4Yv NͩX(fׅQ-K}Z".o}~]̌QYyz@ӃlTNXa\߽}y5Btm~Cnl~Q0sW@\GwAD$$7` &8_FӃbD%X x/]#/C?!_q#b0]a9Ŝd=^NYa@\ً y{S$>8 Av:^-E bHa+sjAqh܏se @>V]-"ok`:T=٨Th19Hkp$ņ$(b(@7t}q]G/]rՙ׌=GyɘřW(9(8BJ`rISj(Ҏ,:( Rq~ !; $C@xOZ> ΰDױ#1.w@1/8 #eQ @(  TIXQ/3>0` %TDk k@f>Pj{MRIRj܈ լ ʪRy>ٗ(Á;6- _w_W9(a__ϒ@-#`i"1MڪCȜ6d$-!F(>)^2D0nյU;hE*C.a7X?^#PJ5)7,3>8ɶ}yu/VcJ[W>'=a'S=~rgӗ;RӨ$ntR[tM[8c~Mkl0%@!X&ʵ^j!p $Wթ0}cbA,J.I, v%Ju}C}-)窭E> Ff{#Q J׃MIn[.*טbjc*W~.kʿ 'r c!GbXX:GJa76ܤ-ō1B+V$Rͥ=?CR_,vP`e3K;3,M/I)A!nYbLD1.( jH"R$3fdlzJ?zGij`KpnUFHX)Rђr2"S(ǖ0l)$Za|9dQ1bo lqtS$zҌ>|i@WJku}%͢TGԕ:gX'],ƵRGҡPpD$?I%%i'Y$vf7qMyOG"y`||hI+>a89Q2 Ve IS.\橕LMet'SQX8Q`JLpNrxkʬ)΅"؜aH,si_J WF_~eO ~_ bdۏf.77P 1X!m 8X .*͔I $1'̀\RҔ`11s Af8F 1f@{W#!?'^lDr* &BgJ#Hj(gHZA V5`SUKG=U%Dp"詮~Dž,."jnV) x8l/ >újN !"|H5\, >V? |q99-PM)C=Q+Dl Vx Iiyv#?G. k&qwI .\8`b{^=ajvgZгAjtBোO%&rRR˕Z t˹E!d'4r]dZJ nR1є~ay3 :1: %**n<iGDk6Es ~bT=JM6;qک>/b:^`]*}bGO7+W©bJpJ'CFw uJF&efL=NG}ZЪF5Q W=\.*jx۰%|H1FJKr&4ĴMAr{V޿;Wd47J\O5Sޚs2Cl{۝6J~V #:~!50ZNÐT~8v:)>[W,1v7* )iHI`$]L?\Hb|IcY]ƕn =%a6m5[NQ`3XQ,>_??.YSm_oȌpBg$C fT+VVFK{ݍS"zoe9 whu`vvwz! 3zȂ|MWfJߺI:%텶=I@J 5A$;dø܏Sy=1N$Mzm3ԱT6wG sڦ:+ZӺ8ԬVX؉.7$ᬒL_Is=53aLlZiH1=jpA@"-0rGF 8izYѦof;~;{|f˷~ݏ_0;1p.mgZ,\y(I%##2~QkC]n["S j v7BOXLcLJD)z+XGlb#Ag=sLt鷟o,W!v!dXM[^Je,(Mʔ d2P667% IGuEJ;(幔Ω 3Y)dI4+kcJT>LW@>l:xX*$oO)IL(MuSߙ7ƦcY՝]1X{c-*q;oq^]xO!}Ǣq CrG:: |2S IO9N4Brot뇔ìeNE!JVU*&X(YNw8u$UJ]1ɡLUU:Aj]rAio ۳_J1ZU@I jRslQgUVҚw/V{YndaG x0_ܚh7^\ 9hZ_CQ:1__ᬦ%jj;[TW+ !Z4cݮ1( #JZRX]We^2eY;.L|E㻣;=ƇYp~1e]!+UnYPT(V0j#:7"Dp%wy.R3A ]Xە)͑ײ@+.t)ɹ \9iI\RpI-' (&H{P%2/x}vU9aTc^ eK])me&jVEl2W0X7J.Z[Cmwlܶ(&˝rۼlȪ#]({fzztHt+QA N"8] sF]Fs6C ǎ6· O2P8g2XV/|Xw;sVe/!qw-=FZїЇPh }Na(Oovl AS8$B0fۓPvlF7B]N29L[C+@.)z+1}R:zwvkrpvO׉FvKKBZCnh[ʻhTb՚M"vO׉Fv,b*Ǜoxn)z+[}ʳ:1&Sym=xKt M!+!S kp{<qi.t nzy(e~U"2q|%*4%*@΁TWBCas[VZ ͊fiJ9V%1WEzX$&)#U :+`%B%خ>2x+FQMTOel>Un[[۸C;+GFZOOٿ#uοr{1]V~zwֽ%?O ~u ~6?gtZ WgGیT!*wxrj%'mo쌝0l|sg,Ey]wDk@U$h1Zm:)¥% 8*~#vqul9΃2Vxc`v'N~?c,/9S96`m`V9/^~_w?~9+"f_V_ l|>iQJL߭ s v{Xin {2EG"c,\$mS #Wfr;+=:cjm]+6YY`A/xC1' |1@8Q1c7jSzAE8zbD>/`Z}10`S7gҞojijSOro~1I_V1ѭ&,<PԈ4mh@"5<@i$9!u_Pђg_PR`Ep$1sd5:08 =?Ǽ^m8QAZ{_y6<#-N4?Bѐ*9D AREJ4i/ JӸ M(NT'*t :Ө7_NXY}Gظ"(E^p&umD)\92J`Pmx=XT4 ( )*lPQ@Q^Bi\)H>@xP6eeZihOvB2@D*tUhWBFt ??mg+wTsA v^Crd[A;Zg\BhUxL&Z9,}}?!]גW} Ĝ%o{@M@Ixe4ME%yߺk2[ԓt-etpݵ U"8G(St+fHNj<l7N#N.f{yv\xWk/xpP@x2aQ^O'uAKRctD'.oӣmF"DG&6>\݇ÿÇO0Oj!%+@=~:vNj.;tц;Ϊ Ku˛_MOw+0f|E~jvc[\8E޶NƓog]}e5h9Owo{}_}UR7+s}1Wzv_,fO47vz! 7mv[gm]z O]1isn_Ń:D;]v' kxuT3"cT sMU0^tymVO+aD`e8ҍsq\7#xfՊIim7f oMɈmfR}"WN~{ɩ59eU)A)jĘvg'Re @ *,yEU!ߓtGsmt5랦lПg_銛 Wm1{4wBjT;;g6 5Kl5=ѝc5m'H6.7wM0*@xQ<u>ĠNfCA=pCoŕS6z[m}.){xGGÆ!TGbF'">T%/.|!mJ)&d4]:-&뽝?ϖkmY^455y1&C,?lMufu=㿒C. X1[xם,@ h_>O}_΁r8bFHv"m'ѥU;|C3VyaTEkXn~$>[悖VfLMtY&ce\1ŖAڠ8+iRUM^cQ-*Jhij)ʹ58ǚTW<:V_n["%/ իuQaK@㹒\Qp闝JxAɕK 6ZDюy4W$eKk{|m; ?PF9ەvu|mB2&x޶(PWd:$S0UWuPVB"G4,aYvt1] /xzr<ʅߤD+5ͿihO) +KM(^Lݬ{>HPЄ1ƜWU]IM+.X"\0+E.σA"{Pe8Ye I*Ӫ:`HmAYI mXB\DxCFOۿio?k,J4tw`ɁP; " ʑ]oI(.ƌ oR{nΨjNpm]q&Xg/s 6J_n)kGuW_ef`]sGrWPrI. *}HI+u_)!D8.YY w.;=LOwMZ5E;,OHɷ״E07e25Y@NqN <)P5>2ل\%"`5"CA=&57OdMG&f٧=,m8<]Kxg OkgU]l}mnhvIC TZts>x58"9X0 )1aU.|Uc}LO[2.7L 4*e,z,fj/ x*/pM!ȃ TK,bE s)DJ4I/1L1q^i6= h'2M-J<bH囹E(G˼s4VkO|\I|r)H6"e'GS4eӳ@Gf(TqYҤV+[a']VcNSl"NEc.GXL93VD^jnɦZ ` j L(" E&:plV"`NypbkPqɜ3AvEP( "Ar_k7zP,-.E8BRܼ2ږ7`=FRQ.e47T|Cij 0]=!S. e^FbA(5 FFŃ}k+<'0S&ȨxUTQq8sS"Ղc5Bʓ E;8;Ѿy-hWI+S2x&mJ|'\nvykAe?uwM^_j$@eЂ s07]vU&jF {*@IHL(o.#U3IR,'D)-sIǜTɻS"O5!RD:wj;Dȫ>gېXe Cl+G֞hW0s373||b0/_v1zxSǡ_j^0m ksȵOν$$Z梔Fm09޹8JM_15l=۳Oζǡ'.<Lw}Ȩi9pRS+c]'kPkFMc]MMqGݠwܫ .Lu o%&\5.AuKp$ǍeOKT jݿN#:C3$+ h޸ [iyY%fbL/| Zp(M?'e .3z?5.`{5[d6^֋ƹՋyPNʭ3yuҵǓ?V? Ew.ֿZxע9fL"JxVO%l#y|b=g#Yoy:҈nv,7{z2:QF~V+{.s*#!kSY k@n}WSphspz:TWi_WՂ0uW[^BE–ԃp]yI\ɨuONTR$Սk[NjSAޮ WsĎWi@&ԼpȂ&.WVD+\܉nIBrm$StvvA5 GtQGV/-tݚPnmHW.d;ϴFD A5 GtQG/Ia~zf>{T@Z`IbhˬlƗJ9yNv# &4MUʰT\w" YD ze/ntVt&˚K$&TS.K OxEu4*9i hK@+l|}ʪm[ff[;vИ~y1q}~kY<Eo?"Z9u) }0R" ϳI@y2<N>OR0ֳTXE Z(OUZXW'[fsXEJT *8r2HVX!JIAHUsޖ7nEԴkJ\!)1lsTv\fU.Y\Ƭ_)12"_DBM8/4GfE57؁MRݪ{1j-`eYZw ),0>ra f;wXgBn~#-8gs<9C:.Ydsa`&!PMv/ܮ"wx-òƒac w+udr b1AUkq@lؿ0&Ѯ [7wŰ {GIY +hWJЄ/RahcKFlU†,Sr.]_dѴ+ }XbLAS7cL0i}SRƮ|[kW10f\=-V"z< @V.5)bW({6|yLVLW~8CB{EvR9xw)դǷ)>)c-ԊR6ܛS%ږӮ[Wۣ?Gjﶭ~O;m/c. ]c~oX\c:7sECuO_/b1ӤRr62E8~==7z!5K.&q^z=L BcmijpElHT5SHMAE8eA ; ~L~1/y/kPJA-gjRK}XdkH^_:CݎnN/I&iZ.yΗc>y%6(-8l&qvسx(7~2:)6I9͎K|]}~LbƤCbJR8몮I {g%Ne?8!YWANfUgnNQR"\i󝑯KewOuwN.O (duΧLimw ' Wh[Xu|+;H&'ȘPN*Π'ʊgV42P;U.9|c5-p?ξ}sv}g?~tja@mO-DMJ8T9B4A-**rpRp{bDvh 9`\i`dD,*%p=2X4e(@ZqO ^h`6A)ahZKeB:L:oҠ/f>,e#!F.H5 fA替)}UJJiW)M&|v7##u(]9! { 'jE u FF?+@^.S8&w/N]@"w,xwӔ@Ke^9ALFx%WI^2j p<1\%1T1OAɹ ,G;%cNGƂfr8eb`j`G텱bt5⦃b:Lb[NHK 0RYHUh-&iN^M lX]Ovۜc3];eVˎZ,X`M3X&`l,Xذ", JFqHhqFP+G"6p`-DOG+r2\iJ.ND&r@@!`Z`\zr wSxWcA~)gpƽ e؇U#$D"r$jP# E<,D,`,HZ4yc $,DVH"da ``QS!`: .q$ӕh0r*]6~Ow`Z>=w Z-x߿%M i!$o?<0+_zʯ~57o_DDHQ:+~|ɆWg-.F]Tsmfo&&Y"OD:J6_!W]5ԹE6@{+fXg G0Nm[I(g`H(p6=j &Pcp4=$XҀj7F;54Zm zHqs2#!?+R*I*z M0(?ghvpbQH"Pn48+hDRZYy;0c,uab %/(%'%`PD9ؙh FMbUx#4UjAB&vLG⤌:h\H Р " 0F.-AmKa  HF ! :N%!ɤgT4 CI!VHl @KV*% QsFK4RFJ?NW&AbGY N]koG+?M0\@_d Gd%RSn-Fs_Vݺ$y"x0-զ!`樶rh&)Q$] 1FUƑ > TV;aD,9\Њ(`2\HCPй R" 7R;bʈԢS-remU #q_b'kZIA`̄pbVQ"=Rr GĴ ZUx26XXe Z?HIsBoj `(xX i% BII#2HOۍ&f2!'`ETsC k^/idbuH0H I"/j'g` NIp\ :I| ),Y0,UA.葰I ”hĒb!j`dK`5\H+R+yVX(%JeDJpؔaJj9| 픧s@Xl?,1/^sbA. UCL൨ڠ&e%-8ƨX*(A04!D$R*050mc֜ 83#bj ݲK#8r'0 I"eEK"/3}jIa96xyaV-1]|w 'Qzxʚ9PkxxE"M k|Fyy'?B |=tEגq;u<ݛ:į3?9>LӏU})=E)#}Ʀ<[0y+o=SLG9L7?V|tռ/^Ļ'R髇}ߛ'Fs_m䡏j;"'-\z ȈS^kl(8]4K̊:]˿- C:g4&)1 ⵾AhHdHl6-DUY 2G"EALOT&*5/lV/Of) F#Ff"#Dd}1-sRdžF~}8ymZd9isp=lwD1ηW0bRLJGC'\#g]%H24%`jp Fsaɶ&ZHNf n%ѠXk )hp0# s$ *-^EȔQ"rv Ҁ׆VՇhn; HPq@5%Jɺ*o rf5b!S0)Bf217/(: 9"%d#=$;rzp<,͎Co O<6B2 Q!e]@c6H=Zk7.rj u{WR59lDMx"ƒY0f\; idTu0ŕw €C yRdPV#( .l՘uzE&$,9IdC\[hеē3r9{.8Gl9#8R[[. `U=YjRA8PCu x(1\KWclJ$E*9}(ʗ<@|C*&TK, Ė_'Ä̛,D#CƗ<2˸g9XdT 0\ #P%j_܅IFJ1AAŴ"'ug(k  i"肷!ʐk;T ҡc[3 f N喀u7`ୁ@ ˾hR2:fZx ꡹BXsf.&,gK.  s݆VIE-Y&8%-!V*,=$?mnn-PUsȏh-.P5;@ɟ۱I)=O;acFε|C:5.K7$:Zx{/P)0S#z͕a:B0ODٴPlbU`>l$;W6@lHknlukM!3X 6F Z fceEx]7HhlFS6NjHp C*нgӦVʬlqPF& &fD٬_QzZh?pjceEx]0Zx6eyKp陟s':P%i=Zꗴ~j-Fi!uOH5#My \BfLW]LxI4Q1` /0mS'Z+zv oU?-KCE6qyrr,7`\60ta/0p!yvn_gKtm@0t-սMtoӡ7\|')nn>pp-vwG.?7H@k1h1i׵Mps{rG׭7\Ճ]{膅VJhE=d_f.K?f[%qU6ғ!MVUq1ՕU*s&-[ ,MV#uv-K( DZ$ֆiu U(GHcm^Q=+*$#w)$ѹė $bh$E;$޹%/@Ha6dC+z. K8eb/[5V2W{H$OJV<ס ԟ]^ImW~[@ v" A\ Li%0%RCC9T<#vW~;}l7[k1|l:;$a(t!|>o=Yaf! |˧zaUZmHףT!~ov"mɻQS-CO/&26Cݾ9dM`4ߺ7:,HV8eza)k-jk㐅n+T>;_YDaiy9<;5z+ZG=} S%atYM@^KkͯnJD$r(o$̯"JĉlP~D"fal{(ζH9 2jof_d9؞IdƳ5W8|%@IB}Z| }iUզ56;@{qc\^0@ʐq5kZrJqMaA460dLqIhezFe3ԄTNE&PTLV.M+,ȦsVʦr*ysHJ+J5{mMhci0HGu%@ * Pj? y!x4h<$z=3{tqK2SЍWveJo~}PuJq:[JEsZ (iyfEk2Se~֬1}TZUk#4kGfeb廋WscU?O98*ܠ0]oG<׿?]>ߞk\J~ߝi kw(AFV,W\um\5]Ґݱ]4:eLg ҩzAuwKoHZ\/{džDesEȴ±}?Ӕ;r=bw)%RrN miK?Q,{3s=͝ߞf+4qs5 Gr6fDQg$yg?6MY8hjQN̯Hv_~_}3B'ݨLf\_IS WrCɢddt.r*Aʐyf.B)8-O8Y/dƣ[cؒZz Bϯnf-=o4~ʸ-tN*Wn N͆O{l}L֨DEHO:8*TF=glA$ߒ f|DErm&eB)^i(b/"bu\ElTf-Em0T͌PcSֳݗT4OU.O(5l$3 ].j9e,m{yC3{zڑ*nF;ZY^-a+Z_{4J^kޖYtk,5xݢUG/kwERSҺ{-KK8)[/7\ ׾j'ѲCPZRսalbZE۵DMtdCx)hn~g_Pwٞj0P,~g| -2tSpj N3yawi}mn>b#Z 􇗐 %i]pJgF<1.cF}F:j=p٥MR.;J<@1䯀Wx\6S.wu==:f1+0+7xlƚ:|ڌ7R ^`Ɔ~ ӇS.1<#}57Wnh~xf=?}lTv{ [O2z zfdW_k/?n>=c|'#ƾ]Ω^M}\\`9\56 v7iˏzǔd&l3Oo&y.Gl=YJJ?ǩȿfrLmZ5-:IBL$/߸w9uȏ-â[];RK Bdy5Ҳ 뤟7AB".p@`A9 r?C/Ӣio "B!K4/ >YnIcmQ[YkW-;])43 2 AR8-i-E !'YfrFiᨰ5+m8%[Wqpi+ӦSrTn%@NrKQP6m!~ ({BZYc`ȜczauyԲR3MQȺx5zg~`5?I8 Nk 1~;o?SA.wS=Fy~tnJ'}wnnw w~}?]/:ħn=OM30 n 3~3' F2murʡ<`M}xi^*d8IcIc~$i{% VYw)(TKfk5_BwG=Lအ֍CyOiA '|7p1bL%N wZWǟFÅ\bYJPк:2+C\@h%yz8VƷ4!_~Kaң jb8lp y%EC#"vO#I"<]񮗁010Va>xTm\l#&s=NKz3J,?a_|9[@ก*-DOi'4Z@^pjW qpM`8S_K4&85K_NOBBˈ^cΘ3L#vNT_}Ϧ~+vpֈb/ $_fsB9lն6xwR Gt7spNÿl%@ҹlP[{ ߱l{1%ީM8(3uA;IS."K=2IFANi+9q)w['G gƷ~GN癞E$&k JUj7U^brGLrG\{/wUoبnl:֠o߅(̛/׀Gnz;7!zm}qyarRc6ZsAn/!29ɽ7V^Ʀq\" xaGk۩_;?Vvz2|"#2Y!ЏBDr SDz=X Yżn!z " c ;i!UTyG/v/%N"1aaʅF1Hqbs_ \\gc~}&o]oY@fu_Wp -:"1xWccflngo?C:r\}C>g@TS ^I ~[ #`J +0/=,A"t&wf/z3H1 |Cx?+wh@X5r;CobʪEqLJF^ pH(46Ok\!s\P VɝVDJF:$-& # ( |_T5mw`T*`2R.3CT@r=^).#]Zl$O:dB!(bvB?ubZ(nlQ?yRݑRDN9JY'&^GIQYy,|(1}/*up–IY*~sq)e ZQ\yo1%B<=dնRsRȅ1UrdF`.|Si!t@~E $W5aA_N@`yjSV@ns׍EeN^L%7tmSA!H/ww)v^mIjT{8DgMLRjjA(4sMbBwX+hx9h@b !>0O*AS^v<9u;l8)V)~_1O2Ci3F:glӦ Y 3K\pW * PT0WH81.a5c(#œV8 u4qLڏGbtm h7_mjw mjK̲MP0A$ܯCa(&ciIDpgPC+x(}<\kdj+XB\XE,7\)fP $xwzqV KSTյSձ I"+e=˶-F_'oNX?F_CLs?\CVRĬri!-hk&LQڨ\aכ2IFu_fCs(Ds==ի/CZ.W>$roe$TJqpFs)_6Τ -om4ɝP6Q% y. AQ>6;{ HV*LR+)c [GvI ։626=!ҥQIvvIZ$CuYrKb;2@j]}P;]]i3Z2B~c0HZC|xҾ TQnWÝXdHp ptxX JYVB HU_aJ :HDs;&jg+d$U:D, 8̩ݳ~skpR\y-i^>mFCF1%tRmn 3UoLUog}%Uo 9 ;a}OJjUը:'IW#풧zkA}@@cԣCkeXNZE1Vh@zU$I^-ExdZq] Pk|yn>, ~7{T̶%ulj* VUEi߰Gp\yi> ՒsQ.*\\/%cL$/-fo;sΫm;(?nxi.Bf!\x- /DdIXyxaB܍,!6z™7F?V [E?臑ߎMeF_\~ NU7a}CS\1VJ:8<_nJ3РY Bɀ(VfB)! 9Cibj3W殭92=1 EOϧ>FRDHx(b* 8 ~bgyo0ve|ȪV4J=$[ju+TH>eӥ,p<2/dLR5\<ꏞ (2(VaG#X_:1w׹\ ,p( X%a_*NR6KLX%a;TV$Bhhzi>#L SYo*BaIb^L2+@R ls!rZǖq*Q;iW!a Ac_IZpB.9N˵A0Q_g҃ v;ejJAu%F;^AB+R^LJ O=4T(43;Rɻ|- ?"W>x|wS g} 3D$~nCb&|EO77#b;h _EV8`3{7nFسbd&lv+V` q.>_쪧짼X\0~My~6-Z?=띁lTZ\K3D@@bƶF>xN33kp-I&jhڍ8GT;C!@R6 v,, s{a@1j#T7|D9~v jDlkf 9gDZ9jf62C 爫|^QgD9 }9Ch>ӰY5\!o˷8Br( LmM T|,h[~H2w2x،$2arLf" #0Y mqV GVJ0"67*P73#M)u3xL6D6uHԆQ3_h.X9D\-i 1f=5rgQ R)긠슝ǟ"iʁbA|G 9ɞDѡ]Q%lfTn Wl35p WZg/U FUhmq^ҖAmYksuY/{!՗V+!wi;TEܷ)պ=%)cZ<ڡzjݳhҌ2j7B3bo{pk;z1Z Cy}H;HߗEs粺b?7q¨h_L,V(IDICGZ HO'n?g~Z*e 1]v ۲-~^k^M۰J4iV5T )i$ PC)UWulT%QhMͼHOBXǒ{5Լ4:9M9r(MyIuu$K@iRT "k41cMٕv3G ", &WćgFS6a>ii**ᄕ˜zQW7Ϛ=JUTY[Ȫ.DMJvCt|B*PP+y{h'QNja'B irD0~^(m)p8 R> VHtoO =/< LiAZϭ_pi]zc@yKLoF8;3+b,o$3}.CN_쭎 $ͧ[ Z 6u]70>M7 a숲!P-eVϥ>?Ӡ! `l QG+n1A4_qxD@Ղ1_Vb_MߍI4}זTBYKU8#U$@! *.Un]L.j!9GsĜ`/ZmSK/zfZk<,okQR+^9?W}?u{zy}WpePU-NԭIPR !H7@5uJiN-iV/=?b]6; z/X:v4ۍc~omt&{=2Z7EI y =i m dz` Ip1XBnLkϾ'IklʲLugj]Zk)K!H-)Vu@;H655>ٲ #;#[p`ghuLu{ ޒLh#~I_-iaha\:Oow82 k<_+A/{J5 Vdz2iApm W wZvyp48 ~ݥqv R{ZswHLU4Yjyܝj,M>T8Ԉrsf fNuqFZz~_i2s8C1eZs6җ' t 9戰6&_!amxaxmtř zգW[c-qPq]P^jpٻa30KVL^>6(ҒXݴG?p=+׳\Ͼ$֮iM-LI8P1%FśFkѠ h;} "C]zKB>(J2{{vq w0##3glnK$^37aK&AxkOi[]Rrn9o4V.܆cKuI1P̱CԌ#F± EC`3Q [鉭\(!bc ,[ydŒ}х; OF!Vh ٔA&r _:e|B^Dz^c $29 F&H8d/a;=sun}،$r-=?|=sUvvę u|>U>?h"9Owc <WB%0Ӄ;0`QAݺơ8Pc(]60Vxvp}%3 <F8a¡Q" 0mU]װ$˙>w8 <+u axOc<|fq΃:<(}NܗĄσzMqk~{)jm -ñ5%gLzs_*қ1ab;gIQ4=**ЧiuUFi$PP6ReUJE(Sp;hJ8*~GO=Z=-| Pb@'`>b"Trz0%^{=f ~8+bq~엊8yR I%yЉ}IĞ6e@0R$Sceddp6V͏`U֡w,V?,YygΎ9o|)r, {&/`፾:=ȟաlhNy'L$H.#r>HI>"N -F9Y!tB31'O,)D_rjY(D{{ofTZ8%Ɇ3u/cviml @#6uQ"jElԱ"dG(?A;x ?AX>vWwUE6ɮ Nn"Zj^tiOR1ctFvsZ&ez7C\5#Q:qP<*TG2G2=7!b,}yKl=85[:\7L?- *I@!@o-@7Wg߇HFsK4({e'p)D@ChmOq֧+y"si>8$IsOp^hvivؐ{ӄ[S4:^#zQ*4t֓Q1:AQgأb`."I[]6.Fmue;f8Z~t:a^\?4_\<6ݯݿ߿;^_xuu-W+y⧶no]8>֕t/u^E l+mk,Gճ֘ Is췬ss#:S_Jn"%/6J9إg\N{X/E.uR U榪I5aYtT\躁=SCjܪ'_5O ԗ]s4.T\ё} ]QmUhP2a1*: Főg<f/q)Y i]2ޙA)b5dƙA,)&'_lB ԋs&zS-[kxݮ$WG_7O_mڮ|Yiul 6SH/uw+3t|Ju'ufkNՠX%PشvMuEɲi*ՙA϶G0kQ7M[ "2 'XȈJBUIP֝X 19Bu2RkyXWaJ{zug=he8}O|!JB>py{W@J#_O`'>AhB-MϏ0N*C큆v{ĬE8x3ECsx#j:* E)8/@)QCM+@5d#B#%?{WVd#_EVc/0}h;;۱`z0}II,~*RyY,UxtFhEM$g3(Q%S3~8=W[>!IG΄3")-}h)w>Ao(vE΀(#ⶔOL;o 9X֞G e-|*ő/ (N1pu XC%\`ۯYxj$r]Wtj_B b /y{-l*/}[-SnP8*3p"$4>x$I犑I]dl QS^3r2=YP5H[PqQIQ|c*b)h{$Hɓ 7\Ow-6Ar*eY~㖵M#U.e te;箼x TyE\PGgs3zR*aPTy/;m}- )ԩx^Y6hD[/ǣoJeDy.EG4}.Ic|_tZV0~BbEgLF9}`}PP]tqCAQ38c_ky4"\&zVsIRy^EKO\-ꪪ <4 u)ph)U&Ah"^WkpuhA]SFU׵vVk-TJ(i+Vh* +U꼩fS^=fiqFO˜̚ENq.;N y hð/T -kР(!rJ>a7QU:Y2#쌘@e2/k\|@ibTYiBK,vIgB*Sa8bTE}Bg Y$Ծ[σĵipg}QK3 1=) 4 rji2 nx;Ċen'# h {T4G/s) 3! _48]<%޽eka ZK1KVsTJXsB]")h`^B4]͇09jWQtyS` 6ؚp4r1TR JU\ݒ}LL=,ג;](=Ľr[6+SUj P ]7q5GטM-Kuh]?'rGpʣ+J44'@yޗ䉶.@^T1~﷏4?j58~xڻ=/2xEyw_72>ݭ/_\wfSǟ<<^!nR䗇Yi;_F݇J =z.UnK&\GIT1ՏjJ*jhC@xnƦfKL\kl%|},eIʔv]ǡIh$_^Mjs);Oiԓ$m{=E\[&!m8r'pq0AfypvZeZR*ᑅ/ Y4}sbe`x,6[dJlӣȔ1JO̊Z"|E<Hz-w-.ǣn2 #LNJQ׻lmeoJrD(\Ɠh*cJJ1QF+?8N\ly1(zW0bPp r\z3(G"Q-c'*>)3Aa4/3;RfU~VܪVsްWkS_Fssn45tZ~N >?="ZA3A gϋoo{U-˔, $Q 4DYӾ(]b7%Ƶ%b~ԥRHKOrGc56r?p1.)׿̞U{I;hj<_Ѹ Y]QQ|10ƙ98ف~z jmeb0Rt6H!?^ܳc^㕛V83;p&F(L/)% L9zїRDRsRmu[@U Yז +KZIBrELW~3w #2qs#5k.7R"5.kީIa|3"rqg^h m'瀤#z,Bv_#Oq3l}LcGJJܛ3p4A*;"x 3r PrBK2h64F2긯0*B/LAHY"q),mS.3GgԸ쒎5/7:Oxp&Tz,bpu14JJ%wRԕ/9p=pB2S(VPM,^'(a, J堵Q*3rE|j%(UyBVȹsC b [*9*u*A0#BqA):\(@?,?7!!rޫ +bB 7+fj Xj@}_R_N |]r06"b.rvڍVc֍KFn0q_Tiig)(?^ʤ.nc9sϋH 1_~҈F4"?5G~q0CB krߨ{BH.c04Ig  FM7Kn~Z,rFMpץÍCcI',?-xZ~oOonz(#=Jg S sJSrT,MAhވ"ښƦ j}0p4Bjktо j+ :V;/ hk)U76;C=(@kͺwC/"sAvR\ѫJcVd|Q'P⡽H7/k挰]¡J $ * )+) &-2>ܐ=]Pc]8SU;QѤSmA'yZ>C$KIE Np>|0O4ESTr/J@7:'GǻMx. ѯ]idY==:K2FPh% [NܪFm/g2{}bt^?2и3&5vљӆ)$]F\[|@0n)N "L01nPz֨'jϘ1OS3'9݃p0BӨ|8DY):GjQE\%0zy[rްe{uE=ݨep-zYd8*U1_؈[v?םH||?{!C3m8؃OKk<ԫ <}#N6տGE|O*_pq ̠kdheHa2 q`%Awe4wuvZś owwiJp ܅f&g/]i-ۥvhemM۽aJ!tGXk. Sa]Z8-@>#[ ͐$v7@]m:Ve5P9!7EG2*k˙UMTUCl(18T+YV<)PY%lp$Ѩp$]p Ee*7KLTRu`"[a8Е;̗`]ZWWqGdV^QՌ9o6$gLCDIBZziUy76Y/ cqɨX:YjS)/Y2ǃwȉ'lFuj[Q(mlz xl2lϤ5+iR>s(K4v^(8zw@͉\{[Tw_sRK"n ^o%N"nV9$G[9^1U Q^| 8O:kXZev׷=h;c2juW}U,VuY@D,v7|^݉r޼@f;L@)0 _WUc69Y7\il ЅXmd[w:ĺ~0b^ NbN2Dv~t\u -c`ppoV,vzr Rdz$EO96P8 g)6f!. ߃1;l;ę(9!Pb&_ta;hpP9H^WfEѻF1n{&A-l5r> o\0f޽nw0 b8_)b7КR}bTZyoJnK*~zR.ҒQ>yKR]R[98tn كxjF3 y~~sֆ2Ǭߗ3L֗ǒtZou/7iuΕ>ok ~ۥ_ rRzR矫ЇSŽ]JŃ3 3S{#2}!?ϟA$Dž"^hҴG0] *6zz kNQދi.UFd:PʦlChW@%i?gNe]C5LW=0Xv{C^FH/Q?7 u~&D5c$Jw(nK'Zp4=4GV ӪUHL ZqQz,.)z9a 1OH%:'6e&ā5S1zN3P-nުש R4we+ 3M.0%Y9d +YA Rhl9 G/R&V)  D\b>q(=ё߲XHbbU#seW\MXg4RUǕ^U@Ar̖+]sfw]n2t]Ƀ^./Jz18NCZAr/m` $ 0Aedw+yntn>l SՃ71Mbx7;piSLCIm<9, 0ģf*zt8Pa 250=4,4З}хN*SY:/475Ǟg.dD:Q?l}SoVp)3J$h_ra= fb6N|NApk>N,T0;~F~w׫"Xxkp'Sic'>'TQ%{_`I6haU9>lȉc$ut+53O=9#C(^j 6PkTH*WcU'S]L_Z\y`tOq\# H0^ f1W NczѸEGWϵymu_ahʮC1(2,LWyݺJh|CPdq p̒HL6N=L7 [s+%c%K]0FqrQ=gqk.ZB5gLD< b]={s7ޫn~HKޟ8뽛REJoR~%i=szn9I&}׬d[GpAҤߦ'@`oO)y|)yf ZE:3'Ens7JY cLˑBdu [3L37@/i&qO|tЮrp)n?E]WgٛLMm:'-3*i#2_`AuLL+n9kW}g攜 < Px*Xi&q0S'(EϚ|{C%LF!ɄZo}QhK&43(NFav2 Wo ev 'e;ۛHt-hHݸ?{LcXG]"';x_ß*[qK|)ttn V=vH)]A16 m}zzv)'cեH^یo]̿"RbcnM;3B甀H>Dz#]/6Urk`r>9!P]W)LY ˕DZWAH"7 {vðQD{p=jc(iT 11q%I)J=9,NN*" [ЄY{S-YKW5ohT˼J#=ԯ?~V~YxyG}i/)X8|O0|U1,~?m;8H#?>f{uLr&E]$ҫqr2D(92HH|>XƧ"B¡HxcsA${H(FŢa3^XRclˌa$3хUBL9 h*a4@s*42 O{QJ'oRv{^TYQFb׾S"Z*`5B !֘|x;fdAT+I1J`m@&Oړg2#`. <2c9TK$|2(=I?jPTDVAFU_J8sYZn_]*ٴ"ױ&L--poZK-@bYLc%CqFQ1B6]cpp$gL Gc]apIQ;]+C,k*q8tR(kg#eBgj ]:(l>V-߽*ca*p޶,s*NWhN59_JD%BeeвNyckS:Л#ŋO=.,/ug/Bg-Qe^ωj^ 1w?- 0a6{g`@ȅF=_ t7P߂nڔ=W=<[ .Ғ<發AnARm@L-Q2nRޭ 9q )ؤx7ZѻtBmLp[xwBND[ !x_I03dr`1Ǣ[.wɱH뱽nN(W/+MW>E%og2s~H U dg-(yO j8? {Po- 筽7ʁxot$]cЍ WhECrQk vF3` Rp>pQ= 90j\&.RS6bf!*#E?R`Zŧ&MK"ҬDj;-/FI\Al/5^[kx]LRB%eC& 3Z0֩6y'1uAbC1K__>: b3AWuEhԛXf]yɛYMYͩ /r3M"9f ɂ"B"WbAs L3i$O &$/֠bGP #\оXu‰M^||F!Wwvr  mbxE(Br]Rλ2LƊM[2( ӄ, >{|U(=6?A׵wcsZw$LY1JcP'TYkRYe! IM$]ZhtmiA(R Ta }DTIfߢjաLm^!j=O'j z#IjX`#MQNR_'*MPRNz4 o_5vSi.Gk-[/8Dl쉢rx1ߋNhRJ>\gQsgQ-F'= R;PkO-B/2b!&K+CĄ>q:p߇iB$tJid/V%{Umz?iୀCG TW$C ):o1IeQ2X9)OԀ"Ntst9&gZubu]^g@i+a#8Ϥ!#Y95EuJ loM /Kl/f._@Qjvo;I7|f{kny5 +3,S%)?zV7y|m|7[+̫{h`!)XTE,_?(BpۙyY;/T[5a䨣XięH$ c+Fs13M fs ͩ i8s}_OE.M0e5FYZ ԔVn = +w-!(B'1MkeS"P;/5 #qB/{asse㎨l7d/7}ĵ1HX(*'F8B ihQ]yU癉(&80*Da)JG{qH 5K qA ̝@$ga2M+lsM3sLI"3B-x riӕD+~~Ԟg8Goc'Z6]: M)%>(b"O3hZ9jNc pJuO82`H2P9 n9m   ij"iy|HPy' $N))=OCTt*OlC:A?U '*h uKg{ݜ(M3k q*"˙DJ5CSjK@tY!P4M6n]C9]>6N'2iSZY\IF NXJYLA9%+@$ Os"rLU&RBcb,Idfn;iPpr ,[t_p¹E񃗒)&/,pOmp䇋4֖xt_N t >hK^#tn5Ofag'gBha<[ydȌLe)YV=#PfPLЂ\^w=n G`zʅ`wb"nUlc-yHqVyӄ"WvS Y>۝`~m"ހᇋc+9| c,_ }5~*`DtG XEWbXū<11ZD(ۏy4懤Hb3f?LivӖԺ߻l{{bsdv 3QoFUvR|N)Cd`hɡ kץ b[C\)yWAj aBxR|n)as:tGd{|?{aY#: !'j fwBtCTImKNHtUrlnv{}3DӊBo1k6~o*ezp0#rP_q5[$|tx4Cw9H8sa{}[PU'!4,q/P#~iOA^ YCO/ D)}޻,q[ftˆIjΛ z\۫rr9GGr*@@<ԅc3׭iJ.)=3E]# RH'1ɟƐ4 q^|z_ ߎ -gvCTʙ)3f9/qTXL1ұMH B(ESjv:@1|Nc98D-fN͹y1*C 'xZ NB˽6֞g-6;=ww'5NTQTOdLO{iPQN`7K"q"XNswh `bt\d?r"]pN|n@jqq:t?Rm^C8")PQ\92Ʊ(ʞ) /*P7^2Y3vl`y¼_mID5f` ,\dJ<,(hb1Ǒ; $ÐWnx:x4.b%Rf9S((CnEEcUn1ziU1j0S8O.84g$6;(XZy,cB8ii  d2DZZ8,ZX)0&nu寳`(%l1I2.Tp/5ĕ"wVi]6!fEIx]Ls]6X7ւlIgCpq5~M.׏N)r՜3K[RjJv{f)=>R~E:6 ON(&ּM;PLzRE-Tؤ62jG)&~@=n6VV_A{х 7pM ;6~J0 j 4t0Z CTΘ^tIUL6U Gmw@+N:E5 5BIPDן`mS㧧Z9ڭ_n/kv#UDQ[Ρgá1ZZጋ2|2Gq0ez!fd`N=qRQ<sF?(['  Qj%8kogI_%}ݾ8̔u5>g܈acwGw=މӻ¶ƇC˧wfR>ɫT>h m:I*N `BK 1'QhjWk{C˻|fg>c rH O.;Om/޲Ff.ުQh+ϲB(5UQ rV'-ޢ@J~dYl/Y}uhp&Y@1ZVVWv˅ 2w{}]Oq^˸‚5v:e+ @Hw3Ο+jkEecv&M5Hm8&V~ér,TS~WQ+[ !_>qUzP4|09T,"m<#uN'\iD#[BqZ^' Q:`FumA3Bc4a%?ntR|ǫr&bA`So[ "ĭz֙h- Lnڪ2T,%}qY uAXbMKO7+TV-^o2 Da%R:kK֟.$(.s`\YLsB8H%1 g̏[P&^.pab:[Lm*K {nEedp,3uU'd*? PR%kpؚW9-Tء;]b-PIJqaN-9yꦪHWZ 9kVԖ]Wu']ĕkA®$"eFi'ω?>*}sH Ǒk6kcuH#;GT{ كWzˑYk +hYwڅ:pzC0(ºg䳬 q.*fi5Ry\>Gmò,v{~Ez,==U"븰 ^;v@|pbXP:ٝ52F# &=fD:eÔ.)YpJ7,;2!Ec!d !p(!p$م`CBehvFɤ8FH8N%nc::#3%!ZC <>O2 6Ԟ\A7ep\UD/9p83QlXQl'מּT,}2`mR fV_Z/{iRTaU?K@4aY*lvtKO{m;oe1K[5\fiR8K3hy/җo/yuh b-7;Ղc#|D.ki tOI߫eWA: ƱYPdN@(ts'c{xT,Sn" ]9>=r9ϫf&'я3Wz0m>:8y(hW.7 #rp恭SF.rpR+zFvEiEo g8;8t KL\HT\ױ |kVh٨P>Lݧ|{! ]IT#UǶu>:T8v-rlm-U݂.A{Pʺ@lJM Rhʴ-p[]}_ں3S*13x2O HÆ,4xxȣjs`v\ l"!Rj@UPk1551j?סM5X[Q_*G;,KcLToMXZ4\ N0F )A<|?#!͉sH6{z}acA} ,$ȉiyl0?Z3̆.|rZ*GO[ޏ9WY>R.wRǘB!C[TЙ8KW,qC8K}Xp"h,w4~Ip!9BI0>6V@=|8n, Z&ѷAZNa9/yn-$y"ig!. 'RI6% ͮ9% bq"6H0ͩzE%tAќAA7:Hl Y2;d `^u$&X'Ʌ05,dz+ %}+sPu4cF|eM6Yϱ麛ǁl;;͠>b>J`8AN~cd>,2hdL^[/!sϒ+T'2H,!3x"(?T_)aQJ j\UT˂+%R VJ|AlAV%'SmdD|r iDicNC-'Ŭ=? `DIE4DJsQ}'ʉͅCjAiZ?@`N\xg,{0TngCPxaG|1o Ll5;PCZl%Qfm.V"ȇrVf;i2#_fG9_ɰⅹ!/HTP"46 LRn*"DG;SYB^a%j*F57v JBְۙk;UŮt5GԤ=fK3#ܖ{=ユ>=|~{Ù j=|xӧ+Ժn2F#wZCv]]j 3QP[m=>4´Ę>=N-Pi1YKCjbXoC㑷:n;nXtO9q$M'Z#Ԍe&eK96_Y|ƽaѭ\^ܓ9zxH$i4Y2qƝYaW掠fšw"cJtqz8-i<;-iZFХSGNq3fATkbo)h.[:>։J7M:'F0k^h294 _:"*3'Drh&\irhBZ<>YX;a.Us:Ǫ vYe= c: ԍ- G\Nx@Cv n(io亶)l a]ewS*c0 먖:_o7nq Af߿ݎ_l?U{wv~?'/}]wчE.}?nd~)m?O~~8h{^5}XM*~ݖʯE}w@Z3S}ؼlաLH;#!#!`T܎#cHڙŐ cVAM<>,?{ȝ7k_ql Vt׿ Nqhwo+/ﭻ_{gs?pWal?ʛwWXkSN3Z -jx{}}?Eٱp` HRUM.Zki 'DUD -~͡]>%قbHDthF'pQؐC-e ZM]{}c~t8m5> (M-6Rsz;Q)Ե tV:WVK0v [+w}7w}?=܃n޼jz_0D{w|(v΢)Gǧo~->h[Y]ׯ7vF<|٣v3?|O6*S BYPˮ(]WM-T m>Y|޻~ J8T ,3s1hAFZꖻuUwv% 檶 y`qc4K>~J1vj'o&?|&0sߏzh`㊼/zyczsM{c i 03^MKW_$e]E潤j4lNe$#1h2kIrY#LD.EEaPJXM)ɑKיb&#。LE$w9 W몡!C={%Z}0XQMN:@Xa C"`Cgp^92QdYDRiݼ9Bȝ$FYZ_0h\+5K2lĄV BXF*ǒiotcܹ.aMٝvà5~&˸i\BI6u9,%ҕj[jJ k)˂ AQJ Y }=N[m JY5(sGg1BW" ][X,$<Q7Ƨu4e`!R4VL4`) d8=CI# [KM+7%v!Znݯ DyСt$)|*ңJ.g+x̜흘̹~zC nhNe\ 3gkC~o!d&ω3^qbՠdHjݸ%0oގ,`PR.DR@dث ,=/8ڔ^ZЉN35 S R-рˑQNn{Xu=2tso/$}ݯXyM)+OYqjRy}2qUzn]N)x|?xg՚0^EZ4(}5S&-Bc۔9B3,=qbwr~q6fi٨`;_Qc9qyaBr\ڙf#ᑔ2"!x4| H2E)P2oOa3zeAT+H_-_9WB9AB`o< >3~x_ %"LQGWzY8)tQ d2}B8\~P꒟E5|PI_o D3"bP'X9`+!}a_2p0Rq)mL':qḰHޔ;hDDi!|~QNCWމE7iS/Ԣg!԰^5`ɩr @#uWw"yNl˞b=]iIl ˩VZF GqD#`4u1m6 ңZPUvazpaz-jWPӥG)o8Xz0`R@V4^P9lJowາ%hQI] "D 5M:>dYVױ2R[ҎKyhJe+:#]/ȡ0[ e۔?'|yfPC\ {^`چ8#i 0tq\|w&m}-*QW%TMVzRav`5PN׮TrK!+@QJ]_Mlݳ^+{?紝w4Ԡ3!&`NFn׷Kjn;gL+<|{~{ڡr?}io(ئې3ߎ $I/"]Q,*j0 %2K^~Pr;æh 3!,λ`z7PJAX~rFܝ}`><򍷏$1irj'͡P|;/~㰅b9eykQS﬌ZPG/ W١`R=adl㈮߃PRj( ]CpR]xF5,@IG=X}:EB҂8}zVU( <ԫ̗x#)bx?sjyM;[-ؒR+RD6-׎??\P/cU}TPU|~rW]ɧT?޺̈́ӻAj& ؝\밞>ءB"_:١k.M|Y9Յ-L rm0z !Vcc#IC:t׏HVӬCoO)ۦ§?zRH"ֵ""+,J_YBPj-= xAA% dII)q)ow$JvŝSbjۮo1ѷ!, D,7 a5iG']\ Eܾ]>  fb+Vd⼣J,OmX"KX?JyW#`p-QضϟmdߘI_uOa3:e# EgT6P(;{bEyPg]kԠ;9Z0 V)]W,mAJ˺VJ{.L e"eNH*JR [= ϱ*!t:S 0q <-x .L7 e [7vIn_S\nBؿ+dH+.D"1~~ " d(<-=?AFzg흘|v~zsvYlNN=;]i:J0~1IHyNd;YEujX1kpO[e#0Cz8D ?KpuBYh'!堆[Fݚki

=Nvnh8P͇'Idxڎɏb|E kg)4u v=@K&9)Z~%6Ѱα3AϠN0<lzkPX쾰 }5ȡ*^?8!CԻ׹9c\mj 7X0 5&D;)JY+NI*Y`A`-E {:%9 GRsz|d>v5^%6^-R';wv"oTToX;a]c/ݴp2}߅_4}fR8Yvw0^&h["@Z|n>_ǴztϬ}Y7 PLю N&%Q{݀iXf64_h˪NTMv ǀ'c&bhyH[۷b`dްiGFPoo^?{91ʗrR ,ؓ^XH -E25H IyAQUY/I?RL7\ F-]̿~O!Hy,-@†b(% xH!_{oeIȾ/O}j^V8H*w &0Ae9nL%64'ڤ'IZ*5#Βҭv)BBeAJ )ŶҤk.BMN𛡒tO=T8ES ^5 m$߽՛va,}PkǕ3 ٗ(HiW"?{6ls,,ld &mF!%'^#VzP$"%YEQd:F8csXIH'ɜuXr$"z!e#8jQD$q4q ` 0rxe.a е]3c}@1+y zLۅZWcz2_sy `{2T4L>I&m 'd|fw x<\^b";JjuykMo|iD]kQcJ nxsΐkr1s,Ͷޥs)^3w 5u-dT2xdMY[FC'D/'O}oI41L)W˺YDwdJn)c= 8v$ $hq|-,4ج937\{ssuI6֥ç/;L^.Gm*Xch/߆-W8Vn\( 8M$IG[lQb+ռ]WF^~P1nbtԻXPTHAFhsRS5<1nJ*ZR0X(M=P#)TiI^T G`a>4(LOmo?Iki52PJ8_4fлLGXث/ p4cDX.Ƈ(޵>JB˥BL]_hkB+$a?KJP)XpuA֝y#?{?0otV<ĆBJ .6gD䨥xa:9<~Y_a.^Q@P';ĤDBZSTVSMo^ gln,,ajbdO\u<5,u;w$ɭw'=Aֱy^z)éayu`G҆);AQď@ AOP|/sY<Ї]?YjagP $>F?yUIadg튐Ry WFya sŞ0?k޲Kު얳N*ėkY'US 9=H8e"Qz9{R&)yJ:;+ Xgi ?_.!EP.Thbj+p`ɔw9ńܳymV@I 8ǕYM8gT!*+,i^5#Y hs2kYP ]+vwdY+-<0VzZPdamX -}2|rd-f[[N/+b!T8i F IA`Di΢ :pZ#4!Zi [M/+VӨJKc"T˄H'͹xE oo牰U j$]XM 65s%gi%ttq|J40ʙYCߥ"+% ƁÁ)o l\R06HH޹Av7%ZBЃ1XbU\"7ȏT9¬z{>Chq p8wLi8MSB7o\&˜!G.L}p-&2Xr̰ |徃kqU#H ^U9fÒ@{p7hHR8,{+:Uk;aHsTiQ rΎrAVo袒Թ'+.wÿrݥS}j yEH-%w7 0ݦLܪ|mܪE[$e¹b|q65KZjUVDm k% Us[?R+Fs*JфC)UF-GVlyZI)S)՛Q>b#.Nd 8A5a c[ OҙהDP[Dq'CyerM"/xFgI }ȂB Rӆw| ʒg͑[HAkNW)D)s^hCQ}*gUR3R>RiYVKY*c{(&PA#Q.UPNQĭR@HknE%QCB€ j 1vVsqaN)l7] UF`#aPDh1H+i9sj\0ҊݥL2!j[!@}4^0>?4b 'ߚd8#դnT;wf)E]EaR hiy`(2"ci(֜3D`Q`07AEq/r2jהVw`0מX/N0H-X^Q2#,a12mQ΅6uOn0Z=k ~(}rjDqݭT g;N^Q ٥zICg鱧+:9y׊:yOO{c;[e)޾S VF\M+p6AON9²v)>jrߢ/ZP )SPߺhёzÛTr]k}(ߘn4~gm/8jƣ0;/Cm\5=؈lED41i78;'uwj ]jٿn݅ٹa/Pw?7uAh^4[&GG2JO[zY'<MTIu-j+w`7ƗՉ g d)%j~vqN_O]gGQP Z̏IHul@v? ǗfZ9W~zuo2헱_F5_5v !0e@Z~[^(hZ0Srla9u6/*^{0?V@1k+L nFYvXa~!a%~#i>4)ŋ&1w. 㱿8o71||oXIg{[_1yxwNjހHķz˞Kg ׋, Ss ozgzHu;v^zg&Ax827it3f߾2tp';;eI̙A7MfJ=t[G {RK096޵[EURtx-ao8_;<tOᅨ]5FQC̓}^U5}5Z]lrKqm275ʝ$~ĈF6@4Q4;-̗=rug.Qut^{ҋKoΆa/A_63#{ο.nbX/Uw>aK32LXQ:~xw2!m7ۑ%6H~;?2.{8x?;?fe ٷ/z^%ej2}?d4{`I%;Pp}ޮ>?Vg~O֙Io|ƳG._F/0^a ʜN?NgXUd+ _:_y<1m=<ֿsmceO}V)l7kY:Fcjw-I!AA=njD2bG}~M36g[RQ;{k[x=?Ҳ:iԱ{$[TX)$Sj]"UJ4|Fi*rf>6 ͧGpkE)g!Bl뻣.y>ysi s?dstս^VgsBP Ais*V!( F.Y佹6WP&aa@Dsnsӹ}4u\5:,6ͧn^diQd@(0* QX(!q$ >D*!&@CVPn~ŝjXo3Qo]3iU@3)\Lڌʝc)^ YP'%$j'qXs_$3~_ɧR*z G$ၲk- V2DA@sf$(Rbj2c1a 6pTĒ!pp9s6";RD,.R 9cya^zx7>`?~_i EPD| !GIl@,`6pKfB(tyR}ު/߹WCH$JbSPR 0^۠xMQg-9=}s HmȆz#clYnGڇߝu7ZܼhdOg7&pg%mQ*6(3 XwçF6+rSâ~[‹]!H@К}[:zF[bl$sT$i!gfW͍\Z\3T?ɧY:7%} Qਂ6AzrO cEERNjĊR`NS#6us2g,"(`9H>BPyUHTA{/Z/KisϪ)U)e,:VL,z=Gi=[J@i߬x7E9}J!I|L!q<"Jcٻ:rWX|IҍTV!Iyk]ΓRF"i^VMO\979T]vQpu|V#&mlgu2$ב1{]CM.;DmTD]v0p5z&?\2&Lxӻ<9b8Uyn:ci]f+x%ٌk}I~ݺ{;{,p5NQ҉5x$w(*cFHnp%o )aDd%Uu~}GOzx}]/}Ѻ|rmBM@yj1߶?AFI [B3{zB{"(e<4~Z^B/ 8|sRjTP9D@:*O<"t$r EҒyWCqX\~#Wd}|[ӭ g DD|- ν %o h!spY"?"GЋrbDQ-6VtB[g{:,1 >"8Y9}%PlJ6g*dͪ Ѭ>T15ilDk7!#J`+q5rO&k"G 1/R.zwR)# lnl=AJ{[a7u .!:5'L"B,,Ĵ.+MV{s|+Z͌,6$~@uՔW†j۪A h,Y52E mYR:VB|80?xjn7e}_jW3VY}NRn(oڍ #.ݳuf'dՈљqE]GkZ(ޗ~o |t]|:="v5abYC % "=DsH1H/aHԠVY&XϚO @_;$$g-h[ˡm/Uz,9Ho঳,^s`i|5Nߕ34|9xnTn/jos3 Hn&6Vn՗/p~nmiL3X1 R|cU7[U )ۜ8—jyQ'd˂'dI6Ȧ55B)֍x~8ϟ~N?(Z4~YB3&(%vSi/wvGcHӉ$KvkrӉ3$Ep"sf#A'3Idp.pNt7p:3eA8H'9(7NE5iF;KY2т@eנ ˨ԡti>RS@?ܲh˰"J5?FUcT͏mlP ڑCMKv*|DQƇ4 K҃6Oa@R~yߏwf[f} 7R6M/2Ժu%*eeoAokJ]7૲vP*[ emq^j^жP[7}mcU0֘XFSZpT:8QP[tRS%Yu. c2⁽ͺ9fA" D0GM%mK 5(X2NUWڰmF2#2uRl4l4B`bXM_ Nrz",wU5:Ew&&cQ,^vç0nS(d E%QF[)|RcibB8}H)wb D|tER]_BPﰆ;! dZ> >ӻVMWe[S{w}Mݿf8j:gLgf(yty??UN(. *>IƂTTJ mD ]X2/DY^ DZ*S1441w논.Ar@8!;^ e k d<Ȝ[oXX- B+Qwq@)ڊB;|LJ3]a]}CXuS\OU((lE7υ}Fy,N:C,!ð/kN;k`ZZ~Z֒TePx,+J>B6[̧ˤbϓx:zjΒ^<=w1Pj8CVxdstRew Prj*/K7L+p:,19F|Qy\y~/<֙L'$|׍u餴E*cl(QVUP¢:x@$jH{+W! kSX]C"c|pRX>81%'C,J#rZh- hW#ңR9LRhᤲl^^kaX­ NCa1v[k6;-~!X! ^0% z}lx uf2gR5yD ~Vf'SMoO N. ׹W3 D&-;d=WRE4Y4=rb݄E񝙘\xzUN5ʩSN-_1HQ|Dՠ'XrgbbU~'1Hmf ޫ뵸Vg*Lu]/> "3>]" $ ]03VG?%Y1qMi*|3 ;?M[dS;QzKc2vT7tm$*Jp 8>AJf#U^Zss8 r j]Zd7NԠEu(nOflYLutp3͒iyQ){$$=ZHyޓW5wNfIn3#Xw̡82!N%5tH\ leB~eUyUJ>?\Vر{Z]m@, ~->ǝ_&^T#?~w+,?w9j3sˇ(߆IBL>L+_VXbF^JK^^mfއ8z d+%E J+@;{.H*Xv݃nA5j+PultoR-E a 3 #C9P\S6}M+(}I#6e?d1wvξM/@ʶS2=}ʊ} oIы;C iO5Udg O)4~L|kũCk ^,CⷿO$T dtAurK)+4ULRB$M}G\ϓ8p{Iizh%N4i8$ut6&. 66N'L5H%E6X/1Q5PkJ5RʻK4l蛚PKܟkdq@yk=ĸ{X> tJ>n#+SƈʁL0@tUM:vXNJ7m  wAN*mLYrFR [,0K9>KZ6z/|^ǂm^^6DB+]eM>O-]Fp}} WLlh]zO o]Zb@P ;5Cٿ&JW7*eЅfʸbN= IΝ1CjMqL: ?`WPImR[x#= 1-0xT/cnS HOo3)7`i p|Χekt蠣џ1u+{E1E3ZV&JFYEVrYYQ0 lWu:/&p'Uis תɤA>; Bd#&*g S8&:(>~igO8ld{L#^.c_ ثD?DE [3S L*ӗaBiMNK?g$~0GfW̋N KT;1hdD1y5izRLSmTk"IK_Yf1 W=xS޵F#"Ű&`Pp>a}.$[nG&|TeP Jʌ 2o:_JySFXE瞟a ts$揝QNa;S(dܲ)S;}N]i7qKnW%NoY)8Haz+?fmm[)}ц*rVr?~_Ve: T] D`-|^f/!NA(I:RH@q6 9Ck:uTBf2 ȕaNKc*8%]-XXoHM:g8g-;Eo*LhzkUgo,U$=hIjݠrQ [v)tu4aG%'(kq7 }`mT. b52AE2Y\%KBc/ XB*h(VeПB&P131_N÷'M_IJ<:(oTFF4peM*LwмVy&7m|pp#fO/zyqz8cdULਧopiqzXinrlbǺƞDwKC`Xf ;a!4 q@æ8 ]2. ;aut-3td q ח%p,5`)PzDdg9T;÷s/vy*^Ryx"X>9߿6pgWwL#|뗧*641L"^?S^?Swz]!Pǩ9umHyǩ6MV A&aUȬ(SI)}CSFN-E҅g9B+5 eWo.$223>fv Y;šv@gi.ejrBlFzt1B FHmXt`{胪,&(9ʂZBq5 b\#rѳ n,r)5gJ)Z><㮕Y\g{2pɳδ撬bWj2aS2A} URܔ:&ˎ9gTj;D4o_u)MKs5I%2K>(ZJM.3siDpݜQtѰ;/QzQ3sd:(ՙt?(L GkWJ(T5#__;yEڏҦ<%U'jgPj,Um{h zVSW$X_cw,kOBU+!s4T!0rnxd:^ZS׮31((\ (twAQmVs>7j3>)V̑j  Φaj :7:Df*!KL{M6=Nt$R _Bs,m8,|}S%bi6ĄL14hB>u:Yw!4Ydj*pX+m*+JliG,a oŎ(M4$JSH*~2x4io*EYQl]: Z_[(LF"c-31$MfY$E0u\ 0^$bnLܼ}K+X.AyI)YXx[U@mN5:3uQ5 7;G݌&0U6[7LLM3vH"8S.Ɓlƨaznr'BufbK Hx4F~ƭ͌&ƨ/R1:Č:^Z1%*T6'zNm^fa.U򨢷A4ѢƦ+iht.H:$nj諀fώPYyt^5d'n]ŭ_|e)O Y)):o}Nǘ^sS". ܴ7 GtPJ77abt{Na6y@w.48p;ˌ3P%4 N蒡,Ϣ+5v@j~0#X #IbA3Ue\K 3).h1@d 8NkUM-q3H`t߇krJ^Y6κrLYhcZs&܂ہRFҡ1.DӐqP*8tMe*ؠT# z¨LJz'oŠ5Zx umR|)BvBl 4&N_)+Ek`șVR'(ϕF+?3hG4-553E+b/0gֲ2j[-i&kjv8SAjPh(chR?C >Lj>L$m>@MP[g(mcBч1m cl,o6_HP*MB(^$R)B?HkoOjc5.3 @޽0k8`UwW~xw!?(( 'fg%&=:dh>)䱔:e!CG -$>.Z|5w?KiV<]Jk,wZ-mQn^ɇg!3 t!-M2XS[aĉJa;v:,ڐޠ˄u`) cv,ۮ?s\ iveBϟZ5?3kmi: D\ߌ¾ @رze{ `JZN]DŽ5xnr2f@RcKtw~gڽjhrYgZ`|{~bZx}9=%.ߓƯqyK<9om>|`0&d l>/w1 G%h%q*9a6sF/fcn:߂+]i=jAJNǦ$n*\ZD116ozK\ZlmWG`kڣ5W'pn31 DVܰ\(aS[T=YIO0λT]}zfݨ1z4IF^yZ8[䯯&.([PLW%xcb_߿`J#X~utz֟^1OIQTUr)o#,?{e&{5ϡt#>"llS fX$A7B>k}hm:RL҃X;T9 n*RR-%'2fM3#aj; 0ėaza2SZKعS|N&ୂ!7nhFzUs N hңWmI Y yUECĠ_ۃ?/v!FŪMoEr ~`!R3Ȍ;qWuz(u\F)XWɍ%pjnR6cLQV+bpE޿븿Ԧ߮[_ AQ2ۻnNАLDcKbl,ک0c"ơ?. XƀYHBhے h -?ZHgZG!p&rSIC o\(EסFx`_-Dr _:xE%яXJokZpk=|P˙P.L̥j[p)JfM%jbP!3X;--p:U5Otڗ+d8mvis8Qy.dnΠv!msڅ{͒ȝk8q_B@TNZL41m^qJV/J1[Ղ?KA0jvhwNPp`HpS\rPiBmeL F84?_ުLLnkGl$Pu XM=&Kg,*@zbq۠-|sP]_u6!w \neY//M&7HE?f{BFˑ1P{b'`@˞bY=11.g}֙V>03m>;)~L,,hF6 S?k7/1ѕ Z5]sPCS3nx~w<)_BM$n%I`\ԟ$/9:\J&Jnp^ņ`U.zؐvQŤHL XHF;L*e cya\N(9NEy1J"|1%_+Y:ҙ@PT!wp-SwQZP( ˭IіY]"C' Wլ(|ej*ŝ5#4ޏ[GզWNH83z=LKϬ+'tk%I8(>M7|.eO$KE*/z$^V:r]}4y& @E[$2 L MNۥHz62:G ú*HϽt QF̙ 7&[`[*"P:hmRD-)XBd bc; OǠ_J43kکR ReJA*-V0~35G~&t-&DiY Ic9e[o{+(Zp(xZ#$ N25)G^:-mz/"DmB\J'^gdj!MudFkQdcLV.xV@jm T A"ԞyQ:*F[A9o}r m+N9ܔ٣5 N֞Y]i &(c Fn&x̬7*UԦMac>8YcMG4iQ#C5v?\'uȓ!|Q/(gd?EzITqL-A1>5v?OAg3uӿ& h0泃&EBW;Œ JG-8j]2ؓ\2Z #Syi;;n0sZ\;Uç哈It`PERA&t"zә7#&9HNchʁ1U,Ghp I)omDLD͂+;1N,1V]8}Ԥ{%S虦 ok :UGXM"ؕ5 !daڟml "<64$#+*j߫/b.96;{=ȃ2{2݇O{]ݞAձhY}Ah\L{6\L~{.Dwʲ~bNyK'&tBLЮGRK^e^&$(yʇ]γFՊ\Kq5JB7rj츆NmrE.3𭤚xxYcjlGu 0W2"f6vbPs}q} 6?*'⌱D`9Q*YiC`d96YgI:;NshOF˒ђ̛͗kK]=3$/ƛ#3؋٩C${nlv閠Q>0Bj"꜄t qPf][3:c,Ӌ7 <F=?YxOz (U&?V=Pssڐl^M&:v|YOM1_7qsR9/c5 9)8B*>t8)0^N䛔{,;YsY&`/6=Kwuu.=޲>Q2fE* /0gY$TA,3L[j@3"u+JbEtOsF0fj܍aNK8tʍ;2hs_]A8+TSXp+A X0hwʒ;]}* @*~kvk.vvD~??fGLUuUxKCj7!w")$"B+|>NId_m)7ͱxecՉ۩SBu ^ɣ@FeD{Դ}np1.w-Ex| |rUKkeRkmǠ,n#p1 ɭE#,t}v軺o4cL!&?}uC_W^(bA06)H9hE')ragC d\Ű~s%Lۦj1JY#4zHF|c8f},*kf w\d4LFJq{Y/$\I~vOd/ .M:5SRdj_^vuM]m0ki+hKc+>"G0܊Nߘv1Ih{0ct˼4!l*.릶5z!em8.&2^ X8בLQ.hq2 Ӆ鬀vgp증GpN8 1Mywjܔw'=aMF6YEESo0f)(LF1iw܍aNq{ǩJdE- UZUo>^;cea-]m'^&j폌/#KvWIH_W@c􅬤-PnL5( W{{Oȍ_1n'Hr9À_=qlr?qmf[-+`XU,wmO?Pp,AMK4}uW-|YTO[͍ۿӷMS4~1zL/yy%d&9ܣRDt᪇5I>sis 1 *s0ɥumN@m:\z\Z]B>ejABe? >#%8Hn)TPRHC5M8xzPJRliIԦk:S<[2t{㏫xnnD_u~CUz-">AK+ч0dF6];;<2k6 6kFS|_ilO1&"8v xeEɐ M介<yr>Ո r+t6bрЍS-}$7 j1s?(KBb3LE =Q|:"4qUՙh $,*(!3llVF4nɌMVjf,T@0CkHH1FdC#Ȩ Z"GoQ+{cFʎtekyn4$n([020'l|N gBR`cPpBBJ㐨$:lDM\f?i3Il5&hh Р@$fLĞV=Z72SKmc-|V&[j`FafSۑJUdK|}CBkm-H,֊FT(q 5r@{i1qUQ䘄=[ɀjQDoufL:GN8¤3 /”X?]-z6P)]o`g3::`7K~͕5O^DžGA CZXd|A[q_wfJkB 3p+<(5]uX-yR!qg\qMR> ghybۙ☳oӟ OHԈyZHdz2 V&=G| -sE^^_tg&5w?)M!B7oD^n>,0Ad0rf*$|m?A)dZ9ZO^|W~ck 6Wj±/bU z!AM)O2[]΋L~ UPVqgn3_ێmeQG5y GPkgI{MHVr\jX?V0.|)|eԀYK|KԌ}hLxtP+i2R7yq0V-ylgkzF{ ίUˊяԼX0FrwV*#KηF2M@Zί٣UUZ*`Q~BV=ԗ}`=sT(X 39h́1QD q$H! 8P5l2QSQֿȐjKܭZ5-1c Lcʩd5;`n3Nsdlpe#.Mv3#HJYeq!nVvYFrSfU#Q{C͕PsQȱCmQSc'Y 0ޞgYۇa2u{]؄|$lnAF f?pyA%#ΥQ;pM SEDIXb;ı&}_]&qAgG< el:&,0'~$;H>tNQdqjX \#0\JiIA},pfOք!yCy ;˻CLbcAwSƲFD W#*f/2!.+DK d˥\Y{H#>9\q!Fj%e$PmHC:+Y#f#:Z.M"[iE` XjΊJe! CsHMjlI}.+9PP3M>mi<,NT=<$ӕHm?EJ>[ll@WiÕQ=i#zr:ѸVpLNDQs6a29mr2KnmaHD19mFlNC'2r|ש܊YQB&#'B69m`*'iKSd -lY<䞞VikޱMIv?zzs(V{drwYADyң68]f@RnD5p 7 ը_cn@Y7s %IC@-;N3ἤb+:Y~KB!/dC~!eV`\BVѢ"mobc,6lIyz[qz>kSކ@Μ_A !s|n\Tsg@Zi'}~v0'")D~SMݞWoHNF 13"2\0`Tc$D=ǾhTOı/ߙdki┞:메o^ )?3j^MaX' Pey!RJRNoR]el 2|e2ŕ0"0#Fgi,&YG&Ppz\0d4&,TYIXb^$.Df8V Fñ=rU̗Qc[zx!!BR%.4VON`L0 h)O\*?ߒѻ98ʎT:{d)?{cjXdǓ8Cf~ʡSUh\*a JV?ReY Pj6A~@ɳyð;)be~kt)'܊$aXJ5gh\PJ؅wTI"XhJh[8c׬>86@Lp9AU>RLi<5ޒRΜuw"!RN^9-.H/@1qȱx=rYa[ȱ~J ۱maZVGVX)-i8dKN썰Z><閜>@B;˼Ke圽j:\z\Zur ߎ\4 䃰!dY0lhڒ::-=;JA*sD%:=y<ϖWk[1>Ԇ:\z\XǥXK:dqp˜rڕ:\xE|> 4}`eVLk|&ko 楀 @yЭ0nh~>qp馥N~{j%r Qsݯxu`0>H^hqQ4ׅ'::6̤u3O)X{s~H+՝:{.?>ⶊD[BEb . or/ң_bOgzBoO824`fh?_)Ci;tgT{ ݆,2c4+}HKfe:.*K@f` [ 62YAd 6 F4u㌖ {+2zdmYCQ@&J"Y/3w$e6:Efo A!:]2֔&!>(ƧU%U(=ej+r% rJ4l6|&FJKLfDOYJdʆ?% W0Ha_;9U^$Xf8k<'&J!)D#Ĭ?{g`X"L$ךA# .X{Y|ݥ7w76YwYrBuieI/o,dDAt_>-Rķܤ_+޻_Q@i|^ޭ)O -rwd$/" W/kI},#0 YsQYXڦ[Z*0dU[o^0/JER|鶝$ΙA kNJIPlNuleaX-i{¢@q`'x,h6 x^jV?<7~*ӂjuw6k"j5w{jقg S.J ݤWw7.Z iŇJd9-mBL&4B,}LCPVBD39 (L櫍kaRuVfW@lwGca=: ZX3[ w6i͞!%d%eG-o*bh1[}}O[%ZXF'pXx@KTCI Uޚph&&; T!xCV{]'O`]>Uoǫ}qz_ Ubn& n&gf0&f2\/@S5I()fhl(+hBud@M^ŘFpyP<]Ks+f3Uj[YjS*>mDzH_?RZJ6nH'C7>=G-(D:6xT.gY@i,e$)k ~҇5tp~Hև' G ȇ 2v^rY&J|*uW!KdA/)Ze!Bw;5қ9G]Yŧƍ|pZ>-zp7BvQ{zy=YK!ճsgZ)a?+ڃSh[}<7ae+'G:b^ VKf̹oR(Ϲ 9)K$_f=#Eҿfw^Z{2*ygg*= ^yybT2RƄw>рBy䝔=Zù묤Om6n]o%ۙ??\\mnI<u۹b6LrH$v4~g":tKZ8E*h󩞉t)rѳ"C9&c(HLkPEPr}9 {+Ϊ|{uC]~ζbGGJ0;9nާw <%=A+>00Yw@>6gK  N#fV V@1Y H6 1XEXPKA4rT#[J=}0Ϳ׷nq?lgHd;|Eη[h `[ (jsib5kI^;U<˭L$[)Y; F4tO;ǹyר|^ JAsPk[meՐm|H%Di3l"dbx0`xg sfz M(vR ~wٱK?tS8R w~@}KeT0jPތ݇J{3'?!Π"=m[Lsh0eXo_^=R3z20pé N%xakQՋ@㕄ÚqjReɩ:IW$TkL->5a}Zx&r=EIpy_ɝ#T6R𛋱PJRsRRooCP'aZMYw^׎n(ͼ׎I0kO^PZesdzΨm5J&(y{2z4Z8_^ژ?آlj: +lzkgJE&ߩ--ζ'83]?,꧹lnYG0+]M*E+˿ 9%R^j!hA gH o8N^,2[\}񏆪trPw ;oOpGVwD;h\stqX@ ( b9m =qTZk8fu[ Ƒ޹s0"Ji>[5f{`0ZY-O!5//zM;Q!]%:JI`Gr1,E krmmT;N`Ix Q"P";sIvʉ:e}@H(/=PÖ æ]D#~\v>] HV [᫕bsPQ oA#o C{b4jZP.XS!'I"mDI>, ݳFLm<{Sr~+ĊLzXo;R@|:+[*+ jD,A i!QRP ?R SY9e <+\E%Pi),go^jRgCW\e2`9E`KwA`s"jUyPXʴ=y#9 NM,s?ݪ,Xm~6|ίh(# 9]p2BRxpZ Ciso0nO`<S6N}N?Dm-xuZ ʁc3k3^?Ȝ#&b#msP@272_&ߗ>2gxŃ06a)jhvT2gHFTa U6QLDC96J?E9vṗ߮*UmBj)3!J"]&_DjIåȐyy9EV> # ,Qş(d!l}P(>TTN pzUȱ(VYfBio>'ߌ3ʂP%?}=]MI0J:SV*bs &͒9T΃<,O%+mqW*L)SXе6k..v5fȎ'-zB @4 )jI3XNF/4!R2 qr=9KNvE,v 1Ml%{lCvf~vo.=~YrPk&aO  B6S~P 8i5Sia0@vaLxkWt;UqZQ[%YIh;=QQz~5JM4='5/ؔ%}|:"%.) !dek~;g?j^[`>O*j7W@ Hxق+F(JȣVUׯš0 uU`dž CEa!Tf!) E Ƿ"$`]:T&ݤO#Ys[pL; 4*ؠ+`HY-/7{eiJ0p͌7kfAL9.P~YaZ13r/n¦wU6m}ׂtpm7`-Kxçhlv6"fn;*(+4c4of=NNϛ?-5=`QQ` P=3k{_XcnrGCi[Ȫ!HOs#cнǡsˣȮr>'mN1-cs{NBTrOT~/Gm;罵~@])#񠑃*t)1u%oȬg21d:x8TDDL嗪  ACbJN/#C~Dur7THH/A`e)244s;hos<4{QvQ7%& Z5(fW NL:Ǻ^}*\Qo")Y}֜OťTdNbFXx@x`"##]٬;v)#1PHmê"N.&x^-e'XdGRX]z R [n69 )jr=ꞽ4Gv7s&yn*%!έh:cI^L[xnLJFy^{\wmJMœ+IqhtEf5U#a ՘FP#mEg7S[sbf)X X)c6:~U3=:߮bF퀒cRB b6f7-<=[Eí($vz> kI`|P\l Qq40E*JP<4$t Q;ڎ˱ȭw:M2wcW<è(Tl3Q$$8:سS Oc 1q?ar*,,gSe`uRzDݙ+'o ejiQ*]`t`D6D 1 #YEnjw`tbH-'Rqǝ6}^) s J8Tw`& FQu1g!+-9m6OZi#2f(M_k|p5mCzsM]a䉝zϧA۪kM# ݽ~$Z]]M4Ow3@'m nlqĸ9UQuրWp՘ >zWcU.vpK:` f(7Fw &3"TWuGslK1Q-VB CWiK NV,-@Vp׶=ֳʂXuPpi}gXux;CpSx듟v?IfS}ٰlnb[ ;rÇWjݮ9ï?E磏Ksc}+akw+ Z t}/ZCDH]te-']\2.+̸XN ĂEOHR^ Pw,bgb5J_{5WXw3=[_\?nXWx5[|EaT R0P892E.h64ݷDj K^j$kENΠ,\FQw̚ڸSA'gvۤ*\Ġ=xR`[wuvPZdUIU: ܠDRL"g3o?ˠMs]D\_g8LJOuCWА Z;  GeNEWB"0>Wum F:qNs4mkJ%h+hqy5CUdxJXdHzXim[|Fh^Vq} 5FC11ղBJ g:Y]ƈ&$:Brk`ƘcO !uweJ|O/kmojJ% L&0<88ljFat.aDcԄ^cx; nC:;/"H0ݨ|ߚiEB_" 0HPRL%EbAG"F[j˽rl eN¸wYMP{{?TE PΧkmWSޮ=M 0{bR1'.fy*2!% Bs&D ,E ך Ra-^XЀ>I{XB}j+A"sD4NarUb#ٖ uNJj9yڥ++2)Hfc./3e$.ˠdt0bOAr<2z9.X t<6neSA;/ѧz.{ՇM! +ks \8^k ,q<`qE,d;䆘ɐIJUY͉ae;@*V\qV<;d˽C<\射, !ZQ]ȩ0'u#)9l63) 2b,}/L繲XyX`sXhP$R3V6@5꬐u6ppV 90{,:x) #+u> ;(;a"WYKu7Xoï$bdPdnJXzPSUpzuVM]EóŪrb9vEM3uϸE,oor^;P<)xJ)qۢ!l,?~H>JbyRwid C:0@DF:H4POwe p)ig됛"<斣u$G_X.?ߟ&S`|Fg)[c%ZBd`@vFL;xaqB09/uw=(G&\8].%7@尌PȳZGa})|(GFg(%}r` rJhU. U\9GȡZhqٳx=xڶ C׵mq4_T7[پAx5+o 3np* 0suҬq净Z8H4+LhHj^z (;g3QqH SyhmK+~ݗRьЗRu0//Lಏky7_vu @x5˿;Jn~DrC6$q DxIeZ'7Qba9bApdH '\$)J; ̀ "1h6SN3ڣ_.f-s`}<>M0C4*` HRGraj75߸^|[ >_@4LÅN0; ::belCZmHL3ϑFnAi#9Z=5H5#D[^Ov!H$%~\.b~ϳLFZk=~d[Q͎d ni8ĩŒ˹O?9wOGo-G 6|Ϧd IXО{JSu]۬n&]/uڈK < (B\7hyx:Mae(V^.~3.Z.P%³U ʉ-4gfm{sO_'ЕT?Djug(j*}cQu>A4 5t0.w4ւ \1@p9td_۽?׏=tYnGqKn W6ƨE&Fy RD!|Զ[ Tlp[󕠕yR賜r﬋6ox\' 6^dǃus+bݝ}{o}PV '-GIVAw/r \1]_)_ūUa@ ˁy Bp~a.JZ:&4vGTف].@9+0gs9\~Ƅp=/l4$e'%hŒG.մbJ=2 l]?7[nܤ6LQsg`Jl?b7{Vb iplG9`R6+9Lk%_ÿ쬗Uo޵=ALqo:R-b^FU$~ %FQݛ֌bz uiznQXu#Zoƨ=78uT;"SBpx Y8F))c7͏-Msv6d<bǢZ:kʣxH=4L_ysoCH!\=}$u[_여ɒāh |*_F%̥$d H{A!ܿB:>֌aS.!@I*R朌0* W c"-Zgs)ews} (Ծbj+!ٰZO0o5fM've՜a@@ }\R^Vy) 5+h M9.w) WKPEIc)ڐot[=i,MI` ý:eDSNg֔Cd`AͭI[򧋠=7>eqSpizD6FSyۓn&q!̳2'0ɿӵҌx-!Fjx[e2]No4_I22OZkSs0㺻 U]VۧM&j]ԑ.0Q(H=b-dzmlwG>*7D \ nGV$7RbIaEʉXR#E]Tc`E,Q%EH%{-Ӕ7u@aq"j#fG>P5^eR5x']ODstu!I/h"&UwPSҥ01(- >P%xv8#K%ѲD5804#'VeohmRkj i./"3Llt g4(M^HŒᔔ*( &zcH #h.\F 6'hCBbO V Z+͝I,1j8'8A?R?~gi^ mE`UR*,wBE@U?c1OO"Ea&ú~%-bӖށĤrP!#Ĭ"[f L~N229" )WՔ>_469)a߫)t\S̩8ق1brU\$#Kƒ$@_W'V"C\tclݖ Kq:5ql'0.s]͓46#a*"Ja<5}̱T/]IxaG}C54MBQ :5v*Y-5i56 ;<t}yƇ%A2 8ludm+v?{WHdP/;e.AF~j af $*%ɖ~&۲2mW@y4$#N0c+W;V'VeH-rg4 2MCd7dI\/)^u*v3RY"x1hJٱLr[ho,d$E9[dQ`Uf,Xa'=T)ۮRZ9&EVPS3JGƌ>eVؿ ?O)}+7Z=Km3zƵao,I ?̄1ir^Qߔb+ͦ[wIVRnJ›GE(|IvXfN}h#c;otE1 ԭ25@>xJC\\/϶1x=;u!%Ŭ|ձ*۩ nTqjWd;3?C&5ckPKN|((cV 5l541x`SDWe'5`0M(-s3XUgy֙ң?+vbf僞.;#iyOZ$ ɂhLqCjoZp*:ռұ T*+hi8vf:-~j>*L^/ǵCMN2\d+kz CL >{Gy] 3l ȻExxzl PJyC{Zv`GX:. Sf†֌ vb\SF/~D HLHpgyGӳ\E=:Xx3l酎I;"B؏J fz[+:W,.;h+@:A ЕjVah\g` h6oZ:틓.%A"]APƍuSڳW.,2bhqTkvny+R+YP=,8]rXýsϖޅju ^Ս/-Boɯ(-)[=Upߗ4וּ5zҵ9 I0) \.镫~c{bτO=nri3AT> >\BJiJ2&)e(xO6YeON]d d/T M4“2Qeu6Fp %M+Z/&kA}g߶:WAS"(ʿd2$}w'q/h;[@.Ufg럎kNjF?&=FkցP&6ΰ]S_co_[ 鐧ge>J潜uڝ&[&{+0˳NG7ji?ta|OnZ+ p4.b7#6NmmC1ek6\~8o4ZU"f{3՟D?8F}q@mb^Eҏggiy(sSZ w2k՗їp3} bOٯIFVƢů64o*cԱ  e6ʢ]TlMŗ[ Vs\C?:|urb-Y cQZOKa ;K K^{|z51`Zjaxߌo Xqn}x3gw9=ߐKr;?hkJӗ{ O(fk'Go||%Cے~dGxBz¯oWڽ <y~aE';{6~UAn#qMjv+ֻ@P++U8z /O(_VO (~,MRWuoEx@Jv%Y^+d8O wňWZ;;N˧ P 뤕)YeJgT L f_2_IJK>_cvE)Z"YoJ-:j&JPۆlqmV KHt1Q`&锇aaz30`ҊgeɴI,eXCe| BiCrȧDӨ6x|=׺t!n >*3/kH-H?:_lYNKTv(l)"O c V8yi<)d='%DI;qQi K +,Nx;4 « Ӽ4^_&H.?`4>3>vzڳ8 f.ANWS}-0zI6Te_$ca Zf2S8:-mGHhZ YY"`p?=-Ě oT#)V_0;A{UZ\A2/TP1UJOgzMhыۭVo!ZlQfeh+ rAڙ=ԗEsƟ'6/=*-&.|9\\\.e__o7BO7=pe.<~b#t)rV y<>B]{+x *@n!z' W]EmiV0B6 U &aKX-gbG |}y'8x,r9_ ZżC@ Gxŋ? \ xY3΃ #mR;Gt>W٘\3tZף|رI֬m3r.*ḵ؃^c+"T"pXv1I UJ"PG!t(#D;.oemBF`vIw3F>-\#EbYg>DOFrK%v]pCl/R9Q ^Rѱ-b `L0ըP5=7QګC$Z<>x2[䐍P}[r-GX;gb_q,vd G@d! XG#7iP(= cþe[wWj%xU&b_cOIσ]b?~GIo+ks;?&idN7m_OS{ZZ3\rɄ {`>GEˌ*P^>#9dx3_Z}+ilFt5zYk󦵱FL|]EJG,zá8L݂^I} u vfO/ \O!WJ3mf(Q+|GwgOD=E) x~obL⛧osP^$ ;bzXv9I8L9gqwBvY{1=rhB|qcpL̟Гb>9œV}Hj~` {{} v 2L:)#]@t&hbàڹيv!L^61k7Nyu>+Q/m52-S'-E Tk{de6֌0"8#nkǝJ nIr̢%k ҅h>b/.FZipy/҃4^ lBꉋ&;lڋEri_p퀊͘7`owY5o\dٲ,z: E?hH,Ǿ?%z*Pig> ?&V1vݏۻ:9bN{,FFj =JDlVI-=avB Kּ'WC5F Qy]"dU'o "oD  d\ nzp%a)NPb-m.34?AZNc9fH'$=!=Ki"Z%#m,켶!*m\NZ\|"k@B. -gZHBKnQW[8ٺCpYW>DO?,%J!%+`&)iHRg( 5Sz/}i W¸ NRLC+A b[ Ւ%s n[M_-a\#E f夭`1՚@:q`@hI$a2N}DVqzʅ€K;a*BIJ;QJ`g[^=8+ 5; `i怨}]k*c08u<"0\k]B ]=^kA] azV;c%|DCV)&- MP- QKUs[B ;좿͵B PDݷ @i0"HEV1`R*]ȨlW>lUAB&K}iL)E0N:X-" hIJ(HGDfK9Z~ֿ.Sp}nyRWN *S|5RJV%QXNDk<^۹+ʨ0/;CFKI'V4O}_E4Ѧ6!BK傧 8٭!@v "y}ۯ* ^9(lUmHc<| eo "bj*rŽ_}4Ԝˤ\åYT{q/<5]lN+io5M坮 /n>X\8  YkƲX?xr=s!xhoh;yF٤`q}:ܝ+UiПLfL~L*5+t]y :ݎa9UK|ujpQ#ŪsCSsSjg|Ǐg)=]Tqpn;N>A|/a4J?n;.Q oL[{6w }\jo?CY4kjۥZL{9=| O* m@whw] ĦύaƟq{{ߧ/rOexn\BcM0,g+te!Ȩ#!K=nrjstn&"jͻ6.Ź|LMeΊw콠R/'( PwWGOoP3T8=7[`ÞEB`֗ h!S3I;,'s;kȗjJ@u#_xu%h`Mi0!;і kZE}X!"_4{Xn&%ݦ-ڿOU MTz2!PHݷEu3BTa/ N#I#I#I#u=4%P˃וNT:QiqD뢯VHOvFh{ 'am |B'=N88749Lކ*;{ga1gb[|A;T*+LJ_'9fVY@gAhNz'5fBqɥQ‚өD(y>9 ZC{u(oo|t!v>dbhJ`*0!,z Փhe!J8D+3i"-bM_Q?],S񗁟z>2%").}w0L|{.:?ӣd]^+M)m.'9@)PC~nL\>s0] <,=Qꔕei+]N3er:_<@^RjAKWq"LEh%B*@&^ ^#TbБR"erK%J6rO6U7Ws{~*|IFJ=3M*9oh!(HfJOLV"ɯ~\D苫7GTTw"ӏoЖ$yߤ&jwO(Rm(zv|}QJUI": ڲ&ArIRc8}Sg/d2.S:9е fzW 4[4 *KN ײywZeO:)|`*듻2 }f>JzG'G=;FC0n?eʊf^Ȋ5ϴ\ocyZ&k:2hGcR,)CTvrOӿnTHarV/ԗVUq窪T\\%%[wFaվQg{&3%1^twujbuՀ 4@@0fd4ء< m~]Cλ9u [vfS;]D 9T.O$ oh͐tS\>M+n@]up~i0|GOٞ'HKkm6-.,J-Kq9OIN%QQ4J5aܽ`= i_V 5[uGR]U VϽ:_$mv: )4ݦ4PҍknC"fw}æ-צZM)1jb)UD v.ebQuzՙiܔ3Uf<_&S<ؔzsջ]MNr^r }#9O8{yyBfǦϥFY߃0챽,xaJI a=M"JeNZtrF/czND渝H-e_N$X$!QނP:aB\!6]FnHCY Nhm}28݇ 8AIHq' _nt݊hN:D3!=L).~*տ|+…|hOF, hY/668aނT1*bq$x︻AzQ։nq/ dm d#@pP;@&;|Srݝ[g<v9%*Dq#.AUAh/Y弤k"8 B:*\Z)7% e>1Mhfׂ##AJhHɠR`*W ԁf.+RSE C`yτРȭʆU?S#/hT̪0ct(H'ayU?Sq}%ᇨAϝL^TМ./r,Qmjgmp)z%" !sB{e/6;h2DYa/rE^t %|-yq`z}ejrq냘7,i&4Βf>7%Fha: {$8p8]EXK1}l ;.ÖG?$-wm㵙NU&+M WMn*35OnOM~ZPr3r9RFnz{a?YׇRmd}\iۑ%_}~c:1LĞ ;K^v劜4 qUуo>mA2מ*Sc{*5%/Ll oJz"n&mt*rI7Pׂ%h5v(:CJ!J{Rgd7i˒֊ {Ha[KE*[b/uV~Z1eu~=3A`}M ]v/d{={o5MMTݕ =1^Eqvrd~5Lƅ "C<-W*dX(5[ ?rPRWB7ӏ%teSc ]å%t+1&)e<%t>%takVي3(`tr U9p7l%t9|W=F&ͳm ܐ*g?yiAv)D>>_]C7?~(PC{[ɛO+*,Snƻb?fI뇉V;l?X/>EHGETja줝 \#UҁJK}JmU-d4Q&:MJEW1( Hi7@NCr[p܄$RK %:@2JoNk5ikI@uHq)e.A/H=X䠍 VzhpqmIWa=KlYm ˗<Ɍ vn8[2b~CQ^o.X|CA,NF\=_D<<\cd^UڊVrTɈ>Ɉaoxec޼HR4v?c`:Gpс`ãd5?1ԯ K#='}[] VG=I|%uMȵr\C#@uc惭%O=VחEEo_gacG:( %Enȩ*JW5r+Ŝ{ VO?.z^%VbL*h{YK.G/ ǨhO@lobw1ZirUЪ3yJh*;Vy\LҹPah?5 0bn U1kRކDp^Zn)7ҽ @w9]LҩU-% 3%,۳.}W] wX TF= /³VYҙ{iʇzasex".{҇q6?{_}Ӑz ܩ=\8$aƠ$$[ FiOa~ [%{%=Ԫ{u4ǂjEuՙƶTlV[LUf2sn?5+COl0"2%-*+zU@) ?t\ yu6r^&rFf]?\z[1>"//gE&`.) C1X]L &NE}t}m#>{e`nDLCrJS]@o;`ޜ]։~6!!Q&"yѱ7}vĭ:w)>B]IdAv)J2:Ska}TBFOuk{B`')3zKvR_F)4Іh΃uJh D@*95Fљ裌6K:.ThF!1 '"0uT=cslmŨ쑜TB7ʭr N.~k|?6qun3o.+]zw}myg-K}"~tL ӏvoOlؠAc&(k%ǂX^O; }ۗFfYͳ}|e$'Y+pB&=Xkhܺg^+DPVocL#-Z٣ .bkI)b!(A#cCtHB`71a8#D63R~r3{^MX%5 S(p[h^0Vkؾ~!/XZdJ9[JJZA_[aIH{%ߪa{'*,&R'^ZE1ĞmB&3OJٿYY;>(cddk5Bjd69Ds /%UWj.`x9ziHtv2*4x"p3gd˼qhRe^Z\z^*6^<%7laAϸ=-UWZbRƾAiVϦu9R䊛] fim,QFi3e L|2#zoI4sLI`cD3.(Bzr}ruclo}뷦lua?zn}sgYu~M1Z?UGK2GkTłmkO4rrnЬU!> AQ^`>GWAjnv?~U)5?bTUjMX|ӷJ>@SP X.6f# $K BF'@WFwG_&Lr[Od A#a` I:UC$ȃF I)xkHcT4),,Q6m +t9>CB,R\4LYN$υ&1`< >Z yF'k]#+THTb#xV:zcvDJؐoˊͅBkȎE|[+bqǢҙ(ؤŦ6UO}/pTIg*:DڍQ:TPqimi\Qvo:^`̿1uu*!@WٹIPe5C:%G3N;Hvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006074327515147550022017714 0ustar rootrootFeb 25 06:45:05 crc systemd[1]: Starting Kubernetes Kubelet... Feb 25 06:45:05 crc restorecon[4805]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:05 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:06 crc restorecon[4805]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 25 06:45:06 crc restorecon[4805]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 25 06:45:07 crc kubenswrapper[4978]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 25 06:45:07 crc kubenswrapper[4978]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 25 06:45:07 crc kubenswrapper[4978]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 25 06:45:07 crc kubenswrapper[4978]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 25 06:45:07 crc kubenswrapper[4978]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 25 06:45:07 crc kubenswrapper[4978]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.054482 4978 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065150 4978 feature_gate.go:330] unrecognized feature gate: Example Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065175 4978 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065180 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065186 4978 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065190 4978 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065195 4978 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065199 4978 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065204 4978 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065209 4978 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065213 4978 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065216 4978 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065220 4978 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065224 4978 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065228 4978 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065232 4978 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065235 4978 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065239 4978 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065244 4978 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065249 4978 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065254 4978 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065259 4978 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065264 4978 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065268 4978 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065273 4978 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065278 4978 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065284 4978 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065290 4978 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065294 4978 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065299 4978 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065311 4978 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065315 4978 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065320 4978 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065324 4978 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065327 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065331 4978 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065335 4978 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065339 4978 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065343 4978 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065348 4978 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065352 4978 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065357 4978 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065378 4978 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065382 4978 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065386 4978 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065391 4978 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065396 4978 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065400 4978 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065405 4978 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065409 4978 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065416 4978 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065420 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065424 4978 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065427 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065432 4978 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065438 4978 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065442 4978 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065446 4978 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065450 4978 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065454 4978 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065458 4978 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065462 4978 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065466 4978 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065470 4978 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065474 4978 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065478 4978 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065482 4978 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065486 4978 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065490 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065494 4978 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065498 4978 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.065501 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066466 4978 flags.go:64] FLAG: --address="0.0.0.0" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066492 4978 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066507 4978 flags.go:64] FLAG: --anonymous-auth="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066518 4978 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066528 4978 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066536 4978 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066546 4978 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066557 4978 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066566 4978 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066575 4978 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066584 4978 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066596 4978 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066604 4978 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066614 4978 flags.go:64] FLAG: --cgroup-root="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066622 4978 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066631 4978 flags.go:64] FLAG: --client-ca-file="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066639 4978 flags.go:64] FLAG: --cloud-config="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066646 4978 flags.go:64] FLAG: --cloud-provider="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066654 4978 flags.go:64] FLAG: --cluster-dns="[]" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066665 4978 flags.go:64] FLAG: --cluster-domain="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066673 4978 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066681 4978 flags.go:64] FLAG: --config-dir="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066688 4978 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066697 4978 flags.go:64] FLAG: --container-log-max-files="5" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066718 4978 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066726 4978 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066735 4978 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066743 4978 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066751 4978 flags.go:64] FLAG: --contention-profiling="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066760 4978 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066768 4978 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066777 4978 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066785 4978 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066795 4978 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066803 4978 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066811 4978 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066819 4978 flags.go:64] FLAG: --enable-load-reader="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066827 4978 flags.go:64] FLAG: --enable-server="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066835 4978 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066845 4978 flags.go:64] FLAG: --event-burst="100" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066853 4978 flags.go:64] FLAG: --event-qps="50" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066862 4978 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066870 4978 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066877 4978 flags.go:64] FLAG: --eviction-hard="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066887 4978 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066895 4978 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066902 4978 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066913 4978 flags.go:64] FLAG: --eviction-soft="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066921 4978 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066928 4978 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066936 4978 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066944 4978 flags.go:64] FLAG: --experimental-mounter-path="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066952 4978 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066960 4978 flags.go:64] FLAG: --fail-swap-on="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066968 4978 flags.go:64] FLAG: --feature-gates="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066979 4978 flags.go:64] FLAG: --file-check-frequency="20s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066987 4978 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.066995 4978 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067004 4978 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067012 4978 flags.go:64] FLAG: --healthz-port="10248" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067020 4978 flags.go:64] FLAG: --help="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067028 4978 flags.go:64] FLAG: --hostname-override="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067038 4978 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067047 4978 flags.go:64] FLAG: --http-check-frequency="20s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067055 4978 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067063 4978 flags.go:64] FLAG: --image-credential-provider-config="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067071 4978 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067079 4978 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067087 4978 flags.go:64] FLAG: --image-service-endpoint="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067094 4978 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067103 4978 flags.go:64] FLAG: --kube-api-burst="100" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067114 4978 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067124 4978 flags.go:64] FLAG: --kube-api-qps="50" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067132 4978 flags.go:64] FLAG: --kube-reserved="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067140 4978 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067147 4978 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067157 4978 flags.go:64] FLAG: --kubelet-cgroups="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067165 4978 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067173 4978 flags.go:64] FLAG: --lock-file="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067181 4978 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067188 4978 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067197 4978 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067218 4978 flags.go:64] FLAG: --log-json-split-stream="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067228 4978 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067236 4978 flags.go:64] FLAG: --log-text-split-stream="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067243 4978 flags.go:64] FLAG: --logging-format="text" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067251 4978 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067259 4978 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067267 4978 flags.go:64] FLAG: --manifest-url="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067284 4978 flags.go:64] FLAG: --manifest-url-header="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067295 4978 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067303 4978 flags.go:64] FLAG: --max-open-files="1000000" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067313 4978 flags.go:64] FLAG: --max-pods="110" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067322 4978 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067330 4978 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067338 4978 flags.go:64] FLAG: --memory-manager-policy="None" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067346 4978 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067354 4978 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067384 4978 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067394 4978 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067414 4978 flags.go:64] FLAG: --node-status-max-images="50" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067421 4978 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067431 4978 flags.go:64] FLAG: --oom-score-adj="-999" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067440 4978 flags.go:64] FLAG: --pod-cidr="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067448 4978 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067461 4978 flags.go:64] FLAG: --pod-manifest-path="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067469 4978 flags.go:64] FLAG: --pod-max-pids="-1" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067477 4978 flags.go:64] FLAG: --pods-per-core="0" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067485 4978 flags.go:64] FLAG: --port="10250" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067492 4978 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067500 4978 flags.go:64] FLAG: --provider-id="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067508 4978 flags.go:64] FLAG: --qos-reserved="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067517 4978 flags.go:64] FLAG: --read-only-port="10255" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067524 4978 flags.go:64] FLAG: --register-node="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067532 4978 flags.go:64] FLAG: --register-schedulable="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067539 4978 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067553 4978 flags.go:64] FLAG: --registry-burst="10" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067561 4978 flags.go:64] FLAG: --registry-qps="5" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067570 4978 flags.go:64] FLAG: --reserved-cpus="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067579 4978 flags.go:64] FLAG: --reserved-memory="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067589 4978 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067597 4978 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067606 4978 flags.go:64] FLAG: --rotate-certificates="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067614 4978 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067621 4978 flags.go:64] FLAG: --runonce="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067629 4978 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067637 4978 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067645 4978 flags.go:64] FLAG: --seccomp-default="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067653 4978 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067661 4978 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067687 4978 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067695 4978 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067704 4978 flags.go:64] FLAG: --storage-driver-password="root" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067712 4978 flags.go:64] FLAG: --storage-driver-secure="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067720 4978 flags.go:64] FLAG: --storage-driver-table="stats" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067728 4978 flags.go:64] FLAG: --storage-driver-user="root" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067737 4978 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067745 4978 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067754 4978 flags.go:64] FLAG: --system-cgroups="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067761 4978 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067775 4978 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067782 4978 flags.go:64] FLAG: --tls-cert-file="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067790 4978 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067801 4978 flags.go:64] FLAG: --tls-min-version="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067809 4978 flags.go:64] FLAG: --tls-private-key-file="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067824 4978 flags.go:64] FLAG: --topology-manager-policy="none" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067831 4978 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067839 4978 flags.go:64] FLAG: --topology-manager-scope="container" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067847 4978 flags.go:64] FLAG: --v="2" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067857 4978 flags.go:64] FLAG: --version="false" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067873 4978 flags.go:64] FLAG: --vmodule="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067882 4978 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.067891 4978 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068087 4978 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068098 4978 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068110 4978 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068119 4978 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068127 4978 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068134 4978 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068142 4978 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068150 4978 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068156 4978 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068163 4978 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068170 4978 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068177 4978 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068184 4978 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068191 4978 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068198 4978 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068208 4978 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068218 4978 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068226 4978 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068233 4978 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068240 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068248 4978 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068255 4978 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068262 4978 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068269 4978 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068278 4978 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068285 4978 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068293 4978 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068300 4978 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068308 4978 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068315 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068322 4978 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068328 4978 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068335 4978 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068342 4978 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068348 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068358 4978 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068390 4978 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068399 4978 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068408 4978 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068416 4978 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068424 4978 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068431 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068461 4978 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068469 4978 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068477 4978 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068484 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068492 4978 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068510 4978 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068518 4978 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068525 4978 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068532 4978 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068539 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068546 4978 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068554 4978 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068561 4978 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068568 4978 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068577 4978 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068585 4978 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068593 4978 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068600 4978 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068607 4978 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068614 4978 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068621 4978 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068628 4978 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068635 4978 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068642 4978 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068649 4978 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068655 4978 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068662 4978 feature_gate.go:330] unrecognized feature gate: Example Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068669 4978 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.068675 4978 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.068687 4978 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.079406 4978 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.079482 4978 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079640 4978 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079662 4978 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079673 4978 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079685 4978 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079694 4978 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079702 4978 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079711 4978 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079719 4978 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079730 4978 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079742 4978 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079752 4978 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079761 4978 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079770 4978 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079778 4978 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079786 4978 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079797 4978 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079806 4978 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079813 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079821 4978 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079829 4978 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079837 4978 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079844 4978 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079852 4978 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079860 4978 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079868 4978 feature_gate.go:330] unrecognized feature gate: Example Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079876 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079884 4978 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079892 4978 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079899 4978 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079907 4978 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079915 4978 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079923 4978 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079930 4978 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079938 4978 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079960 4978 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079969 4978 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079977 4978 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079984 4978 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.079992 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080000 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080008 4978 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080016 4978 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080023 4978 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080031 4978 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080041 4978 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080049 4978 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080057 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080065 4978 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080072 4978 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080080 4978 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080088 4978 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080098 4978 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080108 4978 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080116 4978 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080124 4978 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080131 4978 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080139 4978 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080147 4978 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080155 4978 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080163 4978 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080170 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080179 4978 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080186 4978 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080194 4978 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080202 4978 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080209 4978 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080220 4978 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080230 4978 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080238 4978 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080248 4978 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080268 4978 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.080282 4978 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080722 4978 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080743 4978 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080755 4978 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080766 4978 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080780 4978 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080791 4978 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080801 4978 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080811 4978 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080821 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080847 4978 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080858 4978 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080868 4978 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080878 4978 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080888 4978 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080899 4978 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080908 4978 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080918 4978 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080927 4978 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080937 4978 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080948 4978 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080957 4978 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080967 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080978 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080988 4978 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.080999 4978 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081009 4978 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081019 4978 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081029 4978 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081039 4978 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081049 4978 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081060 4978 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081070 4978 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081079 4978 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081089 4978 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081117 4978 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081127 4978 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081138 4978 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081148 4978 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081158 4978 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081172 4978 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081185 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081199 4978 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081212 4978 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081225 4978 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081238 4978 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081249 4978 feature_gate.go:330] unrecognized feature gate: Example Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081260 4978 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081270 4978 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081280 4978 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081290 4978 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081299 4978 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081310 4978 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081320 4978 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081330 4978 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081340 4978 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081350 4978 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081360 4978 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081410 4978 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081419 4978 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081429 4978 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081439 4978 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081449 4978 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081459 4978 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081469 4978 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081478 4978 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081492 4978 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081503 4978 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081513 4978 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081524 4978 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081536 4978 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.081564 4978 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.081581 4978 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.081985 4978 server.go:940] "Client rotation is on, will bootstrap in background" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.087126 4978 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.091283 4978 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.092168 4978 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.093683 4978 server.go:997] "Starting client certificate rotation" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.093713 4978 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.093916 4978 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.116780 4978 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.119351 4978 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.120194 4978 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.132930 4978 log.go:25] "Validated CRI v1 runtime API" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.181857 4978 log.go:25] "Validated CRI v1 image API" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.184609 4978 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.191331 4978 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-25-06-35-47-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.191417 4978 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.218950 4978 manager.go:217] Machine: {Timestamp:2026-02-25 06:45:07.216480068 +0000 UTC m=+0.655736607 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:a1f14dca-a47a-4420-a814-a4877946a844 BootID:1785e4a8-a914-4b77-ac0a-56ce2c18cd36 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:53:96:30 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:53:96:30 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:35:32:3e Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:c0:85:67 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:1b:d0:9d Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:b9:d1:5a Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:be:34:4c Speed:-1 Mtu:1496} {Name:ens7.44 MacAddress:52:54:00:db:33:f0 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:16:9e:16:97:95:73 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:de:1c:0a:63:bc:97 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.219450 4978 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.219675 4978 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.221469 4978 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.221774 4978 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.221836 4978 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.222263 4978 topology_manager.go:138] "Creating topology manager with none policy" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.222283 4978 container_manager_linux.go:303] "Creating device plugin manager" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.222787 4978 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.222837 4978 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.223231 4978 state_mem.go:36] "Initialized new in-memory state store" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.223889 4978 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.227423 4978 kubelet.go:418] "Attempting to sync node with API server" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.227457 4978 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.227496 4978 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.227517 4978 kubelet.go:324] "Adding apiserver pod source" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.227608 4978 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.233699 4978 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.235923 4978 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.237258 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.237422 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.237335 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.237590 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.237838 4978 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240686 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240766 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240784 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240797 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240820 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240834 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240848 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240871 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240886 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240899 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240936 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240949 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.240989 4978 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.241721 4978 server.go:1280] "Started kubelet" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.242790 4978 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.243243 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.243105 4978 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 25 06:45:07 crc systemd[1]: Started Kubernetes Kubelet. Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.244656 4978 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.249473 4978 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.249617 4978 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.251049 4978 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.251072 4978 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.251463 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.253874 4978 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.253961 4978 factory.go:55] Registering systemd factory Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.255980 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="200ms" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.256576 4978 factory.go:221] Registration of the systemd container factory successfully Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.257490 4978 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18976a589e74661d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.241674269 +0000 UTC m=+0.680930768,LastTimestamp:2026-02-25 06:45:07.241674269 +0000 UTC m=+0.680930768,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.260628 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.260758 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.261202 4978 server.go:460] "Adding debug handlers to kubelet server" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.268040 4978 factory.go:153] Registering CRI-O factory Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.268251 4978 factory.go:221] Registration of the crio container factory successfully Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.268503 4978 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.268677 4978 factory.go:103] Registering Raw factory Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.268848 4978 manager.go:1196] Started watching for new ooms in manager Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.269901 4978 manager.go:319] Starting recovery of all containers Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272088 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272183 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272201 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272216 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272231 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272244 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272256 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272270 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272287 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272301 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272315 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272330 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272341 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272396 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272407 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272418 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272428 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272465 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272478 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272492 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272503 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272515 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272530 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272545 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272558 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272575 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272590 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272607 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272620 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272655 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272667 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272680 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272694 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272707 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272720 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272732 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272744 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272755 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272767 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272778 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272791 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272806 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272820 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272833 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272850 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272860 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272875 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272888 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272900 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272911 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272924 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272937 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272959 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272974 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.272989 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273004 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273024 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273037 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273048 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273058 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273069 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273082 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273095 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273106 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273118 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273130 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273143 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273158 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273172 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273183 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273195 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273207 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273219 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273231 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273243 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273256 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273267 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273279 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273292 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273304 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273317 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273332 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273344 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273355 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273383 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273396 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273407 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273418 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273431 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273442 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273452 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273465 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273477 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273489 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273501 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273514 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273525 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273537 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273547 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273559 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273570 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273581 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273593 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273605 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273624 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273635 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.273650 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275671 4978 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275702 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275719 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275734 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275749 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275764 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275781 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275800 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275815 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275830 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275844 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275855 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.275912 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276227 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276244 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276258 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276273 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276285 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276298 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276314 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276327 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276339 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276401 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276414 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276427 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276438 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276454 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276468 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276484 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276500 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276517 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276532 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276547 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276562 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276580 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276594 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276608 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276621 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276635 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276653 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276667 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276681 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276696 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276710 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276766 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276784 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276801 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276816 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276835 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276852 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276865 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276880 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276894 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276906 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276919 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276932 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276946 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276959 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276976 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.276991 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277005 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277018 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277035 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277048 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277063 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277077 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277090 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277107 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277122 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277136 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277154 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277166 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277179 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277194 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277253 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277270 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277289 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277304 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277320 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277335 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277352 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277384 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277402 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277418 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277432 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277447 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277464 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277479 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277492 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277508 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277522 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277536 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277550 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277565 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277579 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277596 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277614 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277629 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277643 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277657 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277670 4978 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277683 4978 reconstruct.go:97] "Volume reconstruction finished" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.277692 4978 reconciler.go:26] "Reconciler: start to sync state" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.322713 4978 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.324028 4978 manager.go:324] Recovery completed Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.326253 4978 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.326312 4978 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.326341 4978 kubelet.go:2335] "Starting kubelet main sync loop" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.326576 4978 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 25 06:45:07 crc kubenswrapper[4978]: W0225 06:45:07.327749 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.327816 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.337058 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.340490 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.340532 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.340546 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.342379 4978 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.342397 4978 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.342420 4978 state_mem.go:36] "Initialized new in-memory state store" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.351620 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.362039 4978 policy_none.go:49] "None policy: Start" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.363206 4978 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.363259 4978 state_mem.go:35] "Initializing new in-memory state store" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.427110 4978 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.428862 4978 manager.go:334] "Starting Device Plugin manager" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.429123 4978 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.429261 4978 server.go:79] "Starting device plugin registration server" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.429989 4978 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.430154 4978 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.430435 4978 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.430600 4978 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.430622 4978 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.441082 4978 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.457236 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="400ms" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.532136 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.535359 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.535505 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.535539 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.535579 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.536561 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.627258 4978 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.627408 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.628822 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.628856 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.628869 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.628988 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.629420 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.629514 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.629916 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.629960 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.629970 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.630120 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.630224 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.630258 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.631415 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.631443 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.631452 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.631555 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.631577 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.631588 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.631706 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.631838 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.631864 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.631855 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.632023 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.632043 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.632907 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.632973 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.632985 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.633196 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.633242 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.633259 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.633497 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.633680 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.633728 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.635555 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.635598 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.635624 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.635949 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.635987 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.636125 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.636225 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.636248 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.638061 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.638184 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.638278 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.682932 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683030 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683194 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683316 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683497 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683629 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683693 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683735 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683762 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683790 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683845 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683909 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683941 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.683978 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.684030 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.737412 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.740164 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.740221 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.740235 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.740268 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.740815 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.785899 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786000 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786043 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786084 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786121 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786158 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786195 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786228 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786265 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786320 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786352 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786415 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786449 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786485 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.786519 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787205 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787349 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787356 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787439 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787463 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787496 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787513 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787547 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787583 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787600 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787634 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787662 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787690 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787740 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.787743 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: E0225 06:45:07.858284 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="800ms" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.967837 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 25 06:45:07 crc kubenswrapper[4978]: I0225 06:45:07.993212 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.000431 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:08 crc kubenswrapper[4978]: W0225 06:45:08.014527 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-be0889206d21224eaefb786710061e9d6cf64cb1b04d9e4f72cab1eebbfad6d1 WatchSource:0}: Error finding container be0889206d21224eaefb786710061e9d6cf64cb1b04d9e4f72cab1eebbfad6d1: Status 404 returned error can't find the container with id be0889206d21224eaefb786710061e9d6cf64cb1b04d9e4f72cab1eebbfad6d1 Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.020588 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:08 crc kubenswrapper[4978]: W0225 06:45:08.027099 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-9a495a0678d1f2a4786014b433fc3dda374ee7ee4d0c211bbbed0a88ac70d277 WatchSource:0}: Error finding container 9a495a0678d1f2a4786014b433fc3dda374ee7ee4d0c211bbbed0a88ac70d277: Status 404 returned error can't find the container with id 9a495a0678d1f2a4786014b433fc3dda374ee7ee4d0c211bbbed0a88ac70d277 Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.028481 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 06:45:08 crc kubenswrapper[4978]: W0225 06:45:08.033046 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-de568d3ab2743cc0cbad6bcb7f9a1b04c5eaa7fc433bf0cc6e09a2b805996489 WatchSource:0}: Error finding container de568d3ab2743cc0cbad6bcb7f9a1b04c5eaa7fc433bf0cc6e09a2b805996489: Status 404 returned error can't find the container with id de568d3ab2743cc0cbad6bcb7f9a1b04c5eaa7fc433bf0cc6e09a2b805996489 Feb 25 06:45:08 crc kubenswrapper[4978]: W0225 06:45:08.043263 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-d95964f50986c406874c2f431ef48598dcd87cf0684dbff6ca4005c82317d471 WatchSource:0}: Error finding container d95964f50986c406874c2f431ef48598dcd87cf0684dbff6ca4005c82317d471: Status 404 returned error can't find the container with id d95964f50986c406874c2f431ef48598dcd87cf0684dbff6ca4005c82317d471 Feb 25 06:45:08 crc kubenswrapper[4978]: W0225 06:45:08.054019 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ebaf0a0d3f37992e630c235e101092887ea85bf99d82c941d5cf00ad76866f00 WatchSource:0}: Error finding container ebaf0a0d3f37992e630c235e101092887ea85bf99d82c941d5cf00ad76866f00: Status 404 returned error can't find the container with id ebaf0a0d3f37992e630c235e101092887ea85bf99d82c941d5cf00ad76866f00 Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.140981 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.143015 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.143069 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.143085 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.143122 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:08 crc kubenswrapper[4978]: E0225 06:45:08.143806 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.244310 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.335643 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ebaf0a0d3f37992e630c235e101092887ea85bf99d82c941d5cf00ad76866f00"} Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.336714 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d95964f50986c406874c2f431ef48598dcd87cf0684dbff6ca4005c82317d471"} Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.337793 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"de568d3ab2743cc0cbad6bcb7f9a1b04c5eaa7fc433bf0cc6e09a2b805996489"} Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.339026 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9a495a0678d1f2a4786014b433fc3dda374ee7ee4d0c211bbbed0a88ac70d277"} Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.340021 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"be0889206d21224eaefb786710061e9d6cf64cb1b04d9e4f72cab1eebbfad6d1"} Feb 25 06:45:08 crc kubenswrapper[4978]: W0225 06:45:08.477266 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:08 crc kubenswrapper[4978]: E0225 06:45:08.477397 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:08 crc kubenswrapper[4978]: W0225 06:45:08.584792 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:08 crc kubenswrapper[4978]: E0225 06:45:08.584902 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:08 crc kubenswrapper[4978]: W0225 06:45:08.610510 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:08 crc kubenswrapper[4978]: E0225 06:45:08.610629 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:08 crc kubenswrapper[4978]: E0225 06:45:08.660122 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="1.6s" Feb 25 06:45:08 crc kubenswrapper[4978]: W0225 06:45:08.744524 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:08 crc kubenswrapper[4978]: E0225 06:45:08.744655 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.944853 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.947114 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.947164 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.947181 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:08 crc kubenswrapper[4978]: I0225 06:45:08.947216 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:08 crc kubenswrapper[4978]: E0225 06:45:08.947874 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.244637 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.303549 4978 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 25 06:45:09 crc kubenswrapper[4978]: E0225 06:45:09.304955 4978 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.344337 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289" exitCode=0 Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.344449 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289"} Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.344538 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.345736 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.345809 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.345827 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.346223 4978 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6f31926eb93c19d10f15275e100531fc0701b878d8088f88c065d6cd9147b2b5" exitCode=0 Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.346328 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6f31926eb93c19d10f15275e100531fc0701b878d8088f88c065d6cd9147b2b5"} Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.346500 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.347437 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.347542 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.347571 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.347583 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.348741 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.348783 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.348795 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.349335 4978 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="20f5188ccf2a3c8efd8cf05ffb527db4c6262aadcf44ee131df32d3d83236c88" exitCode=0 Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.349417 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"20f5188ccf2a3c8efd8cf05ffb527db4c6262aadcf44ee131df32d3d83236c88"} Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.349485 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.350485 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.350519 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.350537 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.352517 4978 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="a6778fbe21add32d0c23b276645776c3dc2d35772dd3ad4e1245e735003125c4" exitCode=0 Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.352605 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"a6778fbe21add32d0c23b276645776c3dc2d35772dd3ad4e1245e735003125c4"} Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.352674 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.353603 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.353634 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.353691 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.356100 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6bc957fccf17e3aff1ae0caa6046a3f6b4f600333d44ba1bf6d09ec7f08ee2a7"} Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.356143 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ffde46ebd74c2bbdec70b9e9d45e595c544bc25ebae623124219ef5abdbe20b2"} Feb 25 06:45:09 crc kubenswrapper[4978]: I0225 06:45:09.356160 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f069fc26e6703ac36de7f4179eb3ccbb60416ff19e187774c1c3ef0906c16c47"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.244861 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:10 crc kubenswrapper[4978]: E0225 06:45:10.260976 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="3.2s" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.363999 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"251cffe6262dcd87128cb9d7bf8db5a2fc28d9fcac614ba26a2117421302890a"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.364119 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.366690 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.366725 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.366738 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.373622 4978 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="20c99bdf727bfd2ae7454e4fbbf4d206d393fa2c9b7b31d1480116d3b9c65424" exitCode=0 Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.373702 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"20c99bdf727bfd2ae7454e4fbbf4d206d393fa2c9b7b31d1480116d3b9c65424"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.373841 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.375170 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.375203 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.375216 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.378902 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.379337 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"21e30c00d0b9da6b7caaabf5c8c439551cff74fd04aabb0547d9592354e6a587"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.379388 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0abcb04dc31ba2a385cf1195c9321cc09d197c76001549e9cc48c80343fe0b07"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.379399 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a2a8e6f68ede7c2e510c97798e23312539474dc06d309ac4b30d2d8b8ec5b9e6"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.380363 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.380408 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.380421 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.383115 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2e5b9bed96b9f12254cf973f3574d78276e79713302f651888cd459bb8fd5efe"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.383219 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.384095 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.384117 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.384125 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.392712 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.392756 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.392772 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.392788 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41"} Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.548808 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.550890 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.550934 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.550947 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:10 crc kubenswrapper[4978]: I0225 06:45:10.550984 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:10 crc kubenswrapper[4978]: E0225 06:45:10.551508 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.146:6443: connect: connection refused" node="crc" Feb 25 06:45:10 crc kubenswrapper[4978]: W0225 06:45:10.564772 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.146:6443: connect: connection refused Feb 25 06:45:10 crc kubenswrapper[4978]: E0225 06:45:10.564862 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.146:6443: connect: connection refused" logger="UnhandledError" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.401444 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.401406 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a1292302124561bad22248d950d72a1fb7cd432dd29ec79db03eaf31555b89f5"} Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.404315 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.404460 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.404493 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.407748 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"5ac3723b525394c938d9b36f056be8ea94be6a3a31613ac4b161e54cbdeb0219"} Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.407770 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.407618 4978 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="5ac3723b525394c938d9b36f056be8ea94be6a3a31613ac4b161e54cbdeb0219" exitCode=0 Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.407971 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.408027 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.408523 4978 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.408597 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.409219 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.409266 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.409287 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.409340 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.409439 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.409460 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.409997 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.410049 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.410070 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.410008 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.410161 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:11 crc kubenswrapper[4978]: I0225 06:45:11.410184 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.306052 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.416822 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"468c099602c2a03a762b63f6fed593f2352f47fffa67202e38f47c96432be04d"} Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.416899 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c6f76c0863ad36abe987c10726290501065e4f4cd2061e1524da8646e98d86c0"} Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.416927 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"28f8908cdfee165bb17636bbdc25b384d42642ca6893028f2468ece57e30b536"} Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.416940 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.416958 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.416987 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.418718 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.418734 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.418840 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.418793 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.418870 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.418891 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.567422 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.567577 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.569028 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.569112 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.569160 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:12 crc kubenswrapper[4978]: I0225 06:45:12.577129 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.129242 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.426400 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1c676f8b3e428e2f1d8cc9455df7fef897b2857099112b3387bd69993272b375"} Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.426486 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2112a17ceb434838113192df91661110698d269478c734003aef851a57a67395"} Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.426491 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.426624 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.426632 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.428525 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.428570 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.428616 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.428637 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.428694 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.428724 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.428747 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.428576 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.428839 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.465485 4978 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.661591 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.752467 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.754223 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.754278 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.754296 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:13 crc kubenswrapper[4978]: I0225 06:45:13.754330 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.151128 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.430052 4978 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.430099 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.430133 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.430142 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.432111 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.432162 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.432181 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.432112 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.432230 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.432243 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.432292 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.432309 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:14 crc kubenswrapper[4978]: I0225 06:45:14.432256 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:15 crc kubenswrapper[4978]: I0225 06:45:15.432184 4978 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 06:45:15 crc kubenswrapper[4978]: I0225 06:45:15.432241 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:15 crc kubenswrapper[4978]: I0225 06:45:15.433292 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:15 crc kubenswrapper[4978]: I0225 06:45:15.433360 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:15 crc kubenswrapper[4978]: I0225 06:45:15.433411 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:15 crc kubenswrapper[4978]: I0225 06:45:15.526893 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:15 crc kubenswrapper[4978]: I0225 06:45:15.527106 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:15 crc kubenswrapper[4978]: I0225 06:45:15.528624 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:15 crc kubenswrapper[4978]: I0225 06:45:15.528690 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:15 crc kubenswrapper[4978]: I0225 06:45:15.528704 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.129340 4978 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.129519 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.663762 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.664069 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.665909 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.665979 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.666004 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.927467 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.927716 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.929180 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.929244 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:16 crc kubenswrapper[4978]: I0225 06:45:16.929267 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:17 crc kubenswrapper[4978]: E0225 06:45:17.441301 4978 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 06:45:19 crc kubenswrapper[4978]: I0225 06:45:19.012147 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 25 06:45:19 crc kubenswrapper[4978]: I0225 06:45:19.012531 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:19 crc kubenswrapper[4978]: I0225 06:45:19.020043 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:19 crc kubenswrapper[4978]: I0225 06:45:19.020103 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:19 crc kubenswrapper[4978]: I0225 06:45:19.020114 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:21 crc kubenswrapper[4978]: W0225 06:45:21.145046 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 25 06:45:21 crc kubenswrapper[4978]: I0225 06:45:21.145170 4978 trace.go:236] Trace[1879605261]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Feb-2026 06:45:11.143) (total time: 10001ms): Feb 25 06:45:21 crc kubenswrapper[4978]: Trace[1879605261]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:45:21.145) Feb 25 06:45:21 crc kubenswrapper[4978]: Trace[1879605261]: [10.001801839s] [10.001801839s] END Feb 25 06:45:21 crc kubenswrapper[4978]: E0225 06:45:21.145198 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 25 06:45:21 crc kubenswrapper[4978]: I0225 06:45:21.245585 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 25 06:45:21 crc kubenswrapper[4978]: W0225 06:45:21.361593 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 25 06:45:21 crc kubenswrapper[4978]: I0225 06:45:21.361738 4978 trace.go:236] Trace[1124817597]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Feb-2026 06:45:11.359) (total time: 10002ms): Feb 25 06:45:21 crc kubenswrapper[4978]: Trace[1124817597]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:45:21.361) Feb 25 06:45:21 crc kubenswrapper[4978]: Trace[1124817597]: [10.002093159s] [10.002093159s] END Feb 25 06:45:21 crc kubenswrapper[4978]: E0225 06:45:21.361774 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 25 06:45:21 crc kubenswrapper[4978]: W0225 06:45:21.643051 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 25 06:45:21 crc kubenswrapper[4978]: I0225 06:45:21.643194 4978 trace.go:236] Trace[908652755]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Feb-2026 06:45:11.641) (total time: 10001ms): Feb 25 06:45:21 crc kubenswrapper[4978]: Trace[908652755]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:45:21.643) Feb 25 06:45:21 crc kubenswrapper[4978]: Trace[908652755]: [10.00181227s] [10.00181227s] END Feb 25 06:45:21 crc kubenswrapper[4978]: E0225 06:45:21.643231 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 25 06:45:22 crc kubenswrapper[4978]: E0225 06:45:22.151983 4978 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:22Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.18976a589e74661d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.241674269 +0000 UTC m=+0.680930768,LastTimestamp:2026-02-25 06:45:07.241674269 +0000 UTC m=+0.680930768,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:22 crc kubenswrapper[4978]: E0225 06:45:22.156565 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:22Z is after 2026-02-23T05:33:13Z" interval="6.4s" Feb 25 06:45:22 crc kubenswrapper[4978]: E0225 06:45:22.160710 4978 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 06:45:22 crc kubenswrapper[4978]: E0225 06:45:22.170542 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:22Z is after 2026-02-23T05:33:13Z" node="crc" Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.181612 4978 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.181695 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 25 06:45:22 crc kubenswrapper[4978]: W0225 06:45:22.182312 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:22Z is after 2026-02-23T05:33:13Z Feb 25 06:45:22 crc kubenswrapper[4978]: E0225 06:45:22.182417 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.186692 4978 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.186827 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.190010 4978 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:33338->192.168.126.11:17697: read: connection reset by peer" start-of-body= Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.190066 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:33338->192.168.126.11:17697: read: connection reset by peer" Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.248960 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:22Z is after 2026-02-23T05:33:13Z Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.452506 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.454463 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a1292302124561bad22248d950d72a1fb7cd432dd29ec79db03eaf31555b89f5" exitCode=255 Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.454500 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a1292302124561bad22248d950d72a1fb7cd432dd29ec79db03eaf31555b89f5"} Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.454828 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.455759 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.455804 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.455817 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:22 crc kubenswrapper[4978]: I0225 06:45:22.456479 4978 scope.go:117] "RemoveContainer" containerID="a1292302124561bad22248d950d72a1fb7cd432dd29ec79db03eaf31555b89f5" Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.246855 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:23Z is after 2026-02-23T05:33:13Z Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.459494 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.460255 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.462898 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="25a0b74a5708b0c1b59e89a50a55830d25d6a667da868fd1bd1d14a9aef75816" exitCode=255 Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.462956 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"25a0b74a5708b0c1b59e89a50a55830d25d6a667da868fd1bd1d14a9aef75816"} Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.463021 4978 scope.go:117] "RemoveContainer" containerID="a1292302124561bad22248d950d72a1fb7cd432dd29ec79db03eaf31555b89f5" Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.463139 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.464835 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.464872 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.464885 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:23 crc kubenswrapper[4978]: I0225 06:45:23.465452 4978 scope.go:117] "RemoveContainer" containerID="25a0b74a5708b0c1b59e89a50a55830d25d6a667da868fd1bd1d14a9aef75816" Feb 25 06:45:23 crc kubenswrapper[4978]: E0225 06:45:23.465769 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:45:24 crc kubenswrapper[4978]: I0225 06:45:24.250027 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:24Z is after 2026-02-23T05:33:13Z Feb 25 06:45:24 crc kubenswrapper[4978]: I0225 06:45:24.467318 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 25 06:45:25 crc kubenswrapper[4978]: I0225 06:45:25.249272 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:25Z is after 2026-02-23T05:33:13Z Feb 25 06:45:25 crc kubenswrapper[4978]: I0225 06:45:25.534239 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:25 crc kubenswrapper[4978]: I0225 06:45:25.534526 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:25 crc kubenswrapper[4978]: I0225 06:45:25.536677 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:25 crc kubenswrapper[4978]: I0225 06:45:25.536728 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:25 crc kubenswrapper[4978]: I0225 06:45:25.536745 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:25 crc kubenswrapper[4978]: I0225 06:45:25.537632 4978 scope.go:117] "RemoveContainer" containerID="25a0b74a5708b0c1b59e89a50a55830d25d6a667da868fd1bd1d14a9aef75816" Feb 25 06:45:25 crc kubenswrapper[4978]: E0225 06:45:25.537916 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:45:25 crc kubenswrapper[4978]: I0225 06:45:25.539103 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:26 crc kubenswrapper[4978]: W0225 06:45:26.075591 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:26Z is after 2026-02-23T05:33:13Z Feb 25 06:45:26 crc kubenswrapper[4978]: E0225 06:45:26.075684 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.130154 4978 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.130234 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 06:45:26 crc kubenswrapper[4978]: W0225 06:45:26.171537 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:26Z is after 2026-02-23T05:33:13Z Feb 25 06:45:26 crc kubenswrapper[4978]: E0225 06:45:26.171627 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.248998 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:26Z is after 2026-02-23T05:33:13Z Feb 25 06:45:26 crc kubenswrapper[4978]: W0225 06:45:26.311891 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:26Z is after 2026-02-23T05:33:13Z Feb 25 06:45:26 crc kubenswrapper[4978]: E0225 06:45:26.311981 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.366312 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.474263 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.475594 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.475631 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.475641 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.476234 4978 scope.go:117] "RemoveContainer" containerID="25a0b74a5708b0c1b59e89a50a55830d25d6a667da868fd1bd1d14a9aef75816" Feb 25 06:45:26 crc kubenswrapper[4978]: E0225 06:45:26.476426 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.935138 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.935303 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.937074 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.937104 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:26 crc kubenswrapper[4978]: I0225 06:45:26.937118 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:27 crc kubenswrapper[4978]: I0225 06:45:27.248354 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:27Z is after 2026-02-23T05:33:13Z Feb 25 06:45:27 crc kubenswrapper[4978]: E0225 06:45:27.441420 4978 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 06:45:27 crc kubenswrapper[4978]: I0225 06:45:27.477250 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:27 crc kubenswrapper[4978]: I0225 06:45:27.478439 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:27 crc kubenswrapper[4978]: I0225 06:45:27.478508 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:27 crc kubenswrapper[4978]: I0225 06:45:27.478527 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:27 crc kubenswrapper[4978]: I0225 06:45:27.479480 4978 scope.go:117] "RemoveContainer" containerID="25a0b74a5708b0c1b59e89a50a55830d25d6a667da868fd1bd1d14a9aef75816" Feb 25 06:45:27 crc kubenswrapper[4978]: E0225 06:45:27.479813 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:45:28 crc kubenswrapper[4978]: I0225 06:45:28.247351 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:28Z is after 2026-02-23T05:33:13Z Feb 25 06:45:28 crc kubenswrapper[4978]: E0225 06:45:28.563279 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:28Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 25 06:45:28 crc kubenswrapper[4978]: I0225 06:45:28.571461 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:28 crc kubenswrapper[4978]: I0225 06:45:28.572912 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:28 crc kubenswrapper[4978]: I0225 06:45:28.572957 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:28 crc kubenswrapper[4978]: I0225 06:45:28.572967 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:28 crc kubenswrapper[4978]: I0225 06:45:28.572996 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:28 crc kubenswrapper[4978]: E0225 06:45:28.578364 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:28Z is after 2026-02-23T05:33:13Z" node="crc" Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.044806 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.045089 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.046806 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.046875 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.046896 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.066206 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.249831 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:29Z is after 2026-02-23T05:33:13Z Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.482737 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.484283 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.484404 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:29 crc kubenswrapper[4978]: I0225 06:45:29.484444 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:30 crc kubenswrapper[4978]: I0225 06:45:30.249148 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:30Z is after 2026-02-23T05:33:13Z Feb 25 06:45:30 crc kubenswrapper[4978]: I0225 06:45:30.545168 4978 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 25 06:45:30 crc kubenswrapper[4978]: E0225 06:45:30.550326 4978 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:30Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 06:45:31 crc kubenswrapper[4978]: I0225 06:45:31.249573 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:31Z is after 2026-02-23T05:33:13Z Feb 25 06:45:31 crc kubenswrapper[4978]: I0225 06:45:31.289741 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:31 crc kubenswrapper[4978]: I0225 06:45:31.290074 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:31 crc kubenswrapper[4978]: I0225 06:45:31.291722 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:31 crc kubenswrapper[4978]: I0225 06:45:31.291786 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:31 crc kubenswrapper[4978]: I0225 06:45:31.291808 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:31 crc kubenswrapper[4978]: I0225 06:45:31.292722 4978 scope.go:117] "RemoveContainer" containerID="25a0b74a5708b0c1b59e89a50a55830d25d6a667da868fd1bd1d14a9aef75816" Feb 25 06:45:31 crc kubenswrapper[4978]: E0225 06:45:31.293010 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:45:32 crc kubenswrapper[4978]: E0225 06:45:32.157837 4978 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:32Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.18976a589e74661d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.241674269 +0000 UTC m=+0.680930768,LastTimestamp:2026-02-25 06:45:07.241674269 +0000 UTC m=+0.680930768,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:32 crc kubenswrapper[4978]: I0225 06:45:32.249215 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:32Z is after 2026-02-23T05:33:13Z Feb 25 06:45:33 crc kubenswrapper[4978]: W0225 06:45:33.204679 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:33Z is after 2026-02-23T05:33:13Z Feb 25 06:45:33 crc kubenswrapper[4978]: E0225 06:45:33.204799 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:33Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 06:45:33 crc kubenswrapper[4978]: I0225 06:45:33.248274 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:33Z is after 2026-02-23T05:33:13Z Feb 25 06:45:34 crc kubenswrapper[4978]: I0225 06:45:34.249107 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:34Z is after 2026-02-23T05:33:13Z Feb 25 06:45:34 crc kubenswrapper[4978]: W0225 06:45:34.404799 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:34Z is after 2026-02-23T05:33:13Z Feb 25 06:45:34 crc kubenswrapper[4978]: E0225 06:45:34.404901 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:34Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 06:45:35 crc kubenswrapper[4978]: I0225 06:45:35.249591 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:35Z is after 2026-02-23T05:33:13Z Feb 25 06:45:35 crc kubenswrapper[4978]: E0225 06:45:35.568152 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:35Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 25 06:45:35 crc kubenswrapper[4978]: I0225 06:45:35.578934 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:35 crc kubenswrapper[4978]: I0225 06:45:35.580741 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:35 crc kubenswrapper[4978]: I0225 06:45:35.580803 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:35 crc kubenswrapper[4978]: I0225 06:45:35.580823 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:35 crc kubenswrapper[4978]: I0225 06:45:35.580866 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:35 crc kubenswrapper[4978]: E0225 06:45:35.586041 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:35Z is after 2026-02-23T05:33:13Z" node="crc" Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.130928 4978 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.131061 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.131133 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.131300 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.132843 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.132928 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.132949 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.133868 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"ffde46ebd74c2bbdec70b9e9d45e595c544bc25ebae623124219ef5abdbe20b2"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.134310 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://ffde46ebd74c2bbdec70b9e9d45e595c544bc25ebae623124219ef5abdbe20b2" gracePeriod=30 Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.249902 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:36Z is after 2026-02-23T05:33:13Z Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.506474 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.507308 4978 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ffde46ebd74c2bbdec70b9e9d45e595c544bc25ebae623124219ef5abdbe20b2" exitCode=255 Feb 25 06:45:36 crc kubenswrapper[4978]: I0225 06:45:36.507378 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ffde46ebd74c2bbdec70b9e9d45e595c544bc25ebae623124219ef5abdbe20b2"} Feb 25 06:45:36 crc kubenswrapper[4978]: W0225 06:45:36.871079 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:36Z is after 2026-02-23T05:33:13Z Feb 25 06:45:36 crc kubenswrapper[4978]: E0225 06:45:36.871150 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:36Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 06:45:37 crc kubenswrapper[4978]: I0225 06:45:37.249356 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:37Z is after 2026-02-23T05:33:13Z Feb 25 06:45:37 crc kubenswrapper[4978]: E0225 06:45:37.441679 4978 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 06:45:37 crc kubenswrapper[4978]: I0225 06:45:37.514655 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 25 06:45:37 crc kubenswrapper[4978]: I0225 06:45:37.515663 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a68e7bc93f53c30fbdb306833c2b72b5f128bdf37a9434cf9868868a2bf39454"} Feb 25 06:45:37 crc kubenswrapper[4978]: I0225 06:45:37.515834 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:37 crc kubenswrapper[4978]: I0225 06:45:37.517943 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:37 crc kubenswrapper[4978]: I0225 06:45:37.518001 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:37 crc kubenswrapper[4978]: I0225 06:45:37.518026 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:38 crc kubenswrapper[4978]: I0225 06:45:38.252820 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:38Z is after 2026-02-23T05:33:13Z Feb 25 06:45:38 crc kubenswrapper[4978]: I0225 06:45:38.518429 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:38 crc kubenswrapper[4978]: I0225 06:45:38.520040 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:38 crc kubenswrapper[4978]: I0225 06:45:38.520137 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:38 crc kubenswrapper[4978]: I0225 06:45:38.520166 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:38 crc kubenswrapper[4978]: W0225 06:45:38.754064 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:38Z is after 2026-02-23T05:33:13Z Feb 25 06:45:38 crc kubenswrapper[4978]: E0225 06:45:38.754193 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:38Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 25 06:45:39 crc kubenswrapper[4978]: I0225 06:45:39.249247 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:39Z is after 2026-02-23T05:33:13Z Feb 25 06:45:40 crc kubenswrapper[4978]: I0225 06:45:40.251786 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:40Z is after 2026-02-23T05:33:13Z Feb 25 06:45:41 crc kubenswrapper[4978]: I0225 06:45:41.249733 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:41Z is after 2026-02-23T05:33:13Z Feb 25 06:45:42 crc kubenswrapper[4978]: E0225 06:45:42.163534 4978 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:42Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.18976a589e74661d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.241674269 +0000 UTC m=+0.680930768,LastTimestamp:2026-02-25 06:45:07.241674269 +0000 UTC m=+0.680930768,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:42 crc kubenswrapper[4978]: I0225 06:45:42.251789 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:42Z is after 2026-02-23T05:33:13Z Feb 25 06:45:42 crc kubenswrapper[4978]: E0225 06:45:42.574837 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:42Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 25 06:45:42 crc kubenswrapper[4978]: I0225 06:45:42.587197 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:42 crc kubenswrapper[4978]: I0225 06:45:42.589522 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:42 crc kubenswrapper[4978]: I0225 06:45:42.589596 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:42 crc kubenswrapper[4978]: I0225 06:45:42.589623 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:42 crc kubenswrapper[4978]: I0225 06:45:42.589672 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:42 crc kubenswrapper[4978]: E0225 06:45:42.594810 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:42Z is after 2026-02-23T05:33:13Z" node="crc" Feb 25 06:45:43 crc kubenswrapper[4978]: I0225 06:45:43.130041 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:43 crc kubenswrapper[4978]: I0225 06:45:43.130314 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:43 crc kubenswrapper[4978]: I0225 06:45:43.131955 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:43 crc kubenswrapper[4978]: I0225 06:45:43.132022 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:43 crc kubenswrapper[4978]: I0225 06:45:43.132049 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:43 crc kubenswrapper[4978]: I0225 06:45:43.247999 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:43Z is after 2026-02-23T05:33:13Z Feb 25 06:45:44 crc kubenswrapper[4978]: I0225 06:45:44.151979 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:45:44 crc kubenswrapper[4978]: I0225 06:45:44.152274 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:44 crc kubenswrapper[4978]: I0225 06:45:44.153945 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:44 crc kubenswrapper[4978]: I0225 06:45:44.154030 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:44 crc kubenswrapper[4978]: I0225 06:45:44.154049 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:44 crc kubenswrapper[4978]: I0225 06:45:44.247776 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:44Z is after 2026-02-23T05:33:13Z Feb 25 06:45:45 crc kubenswrapper[4978]: I0225 06:45:45.249333 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:45Z is after 2026-02-23T05:33:13Z Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.130651 4978 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.130761 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.249763 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-25T06:45:46Z is after 2026-02-23T05:33:13Z Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.326729 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.327944 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.327998 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.328009 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.328623 4978 scope.go:117] "RemoveContainer" containerID="25a0b74a5708b0c1b59e89a50a55830d25d6a667da868fd1bd1d14a9aef75816" Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.543574 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.545884 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bdd08f3895f5ffcce127508a45c274bf2fe04457b344de1e985cb2978c68fa57"} Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.546006 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.546852 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.546878 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:46 crc kubenswrapper[4978]: I0225 06:45:46.546887 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:47 crc kubenswrapper[4978]: I0225 06:45:47.251777 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:47 crc kubenswrapper[4978]: I0225 06:45:47.398876 4978 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 25 06:45:47 crc kubenswrapper[4978]: I0225 06:45:47.416497 4978 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 25 06:45:47 crc kubenswrapper[4978]: E0225 06:45:47.441770 4978 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.250272 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.554766 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.555584 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.558178 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bdd08f3895f5ffcce127508a45c274bf2fe04457b344de1e985cb2978c68fa57" exitCode=255 Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.558234 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"bdd08f3895f5ffcce127508a45c274bf2fe04457b344de1e985cb2978c68fa57"} Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.558280 4978 scope.go:117] "RemoveContainer" containerID="25a0b74a5708b0c1b59e89a50a55830d25d6a667da868fd1bd1d14a9aef75816" Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.558545 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.560534 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.560581 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.560603 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:48 crc kubenswrapper[4978]: I0225 06:45:48.561764 4978 scope.go:117] "RemoveContainer" containerID="bdd08f3895f5ffcce127508a45c274bf2fe04457b344de1e985cb2978c68fa57" Feb 25 06:45:48 crc kubenswrapper[4978]: E0225 06:45:48.562061 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:45:49 crc kubenswrapper[4978]: I0225 06:45:49.251475 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:49 crc kubenswrapper[4978]: I0225 06:45:49.564793 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 25 06:45:49 crc kubenswrapper[4978]: E0225 06:45:49.586127 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 25 06:45:49 crc kubenswrapper[4978]: I0225 06:45:49.595161 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:49 crc kubenswrapper[4978]: I0225 06:45:49.597135 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:49 crc kubenswrapper[4978]: I0225 06:45:49.597209 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:49 crc kubenswrapper[4978]: I0225 06:45:49.597236 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:49 crc kubenswrapper[4978]: I0225 06:45:49.597278 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:49 crc kubenswrapper[4978]: E0225 06:45:49.604221 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 25 06:45:50 crc kubenswrapper[4978]: I0225 06:45:50.251808 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:51 crc kubenswrapper[4978]: I0225 06:45:51.250906 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:51 crc kubenswrapper[4978]: I0225 06:45:51.289045 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:51 crc kubenswrapper[4978]: I0225 06:45:51.289314 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:51 crc kubenswrapper[4978]: I0225 06:45:51.291298 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:51 crc kubenswrapper[4978]: I0225 06:45:51.291600 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:51 crc kubenswrapper[4978]: I0225 06:45:51.291797 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:51 crc kubenswrapper[4978]: I0225 06:45:51.292837 4978 scope.go:117] "RemoveContainer" containerID="bdd08f3895f5ffcce127508a45c274bf2fe04457b344de1e985cb2978c68fa57" Feb 25 06:45:51 crc kubenswrapper[4978]: E0225 06:45:51.293278 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.171053 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a589e74661d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.241674269 +0000 UTC m=+0.680930768,LastTimestamp:2026-02-25 06:45:07.241674269 +0000 UTC m=+0.680930768,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.178349 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458a028 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340517416 +0000 UTC m=+0.779773885,LastTimestamp:2026-02-25 06:45:07.340517416 +0000 UTC m=+0.779773885,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.185726 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458fb91 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340540817 +0000 UTC m=+0.779797286,LastTimestamp:2026-02-25 06:45:07.340540817 +0000 UTC m=+0.779797286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.193085 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a4592fad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340554157 +0000 UTC m=+0.779810626,LastTimestamp:2026-02-25 06:45:07.340554157 +0000 UTC m=+0.779810626,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.199684 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a9d461e9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.432514025 +0000 UTC m=+0.871770524,LastTimestamp:2026-02-25 06:45:07.432514025 +0000 UTC m=+0.871770524,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.207029 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458a028\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458a028 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340517416 +0000 UTC m=+0.779773885,LastTimestamp:2026-02-25 06:45:07.535471077 +0000 UTC m=+0.974727606,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.213909 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458fb91\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458fb91 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340540817 +0000 UTC m=+0.779797286,LastTimestamp:2026-02-25 06:45:07.535527057 +0000 UTC m=+0.974783566,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.219608 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a4592fad\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a4592fad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340554157 +0000 UTC m=+0.779810626,LastTimestamp:2026-02-25 06:45:07.535552837 +0000 UTC m=+0.974809336,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.227629 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458a028\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458a028 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340517416 +0000 UTC m=+0.779773885,LastTimestamp:2026-02-25 06:45:07.628847357 +0000 UTC m=+1.068103826,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.235085 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458fb91\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458fb91 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340540817 +0000 UTC m=+0.779797286,LastTimestamp:2026-02-25 06:45:07.628864647 +0000 UTC m=+1.068121116,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.242690 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a4592fad\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a4592fad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340554157 +0000 UTC m=+0.779810626,LastTimestamp:2026-02-25 06:45:07.628876287 +0000 UTC m=+1.068132756,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: I0225 06:45:52.250029 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.250458 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458a028\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458a028 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340517416 +0000 UTC m=+0.779773885,LastTimestamp:2026-02-25 06:45:07.629947332 +0000 UTC m=+1.069203791,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.257299 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458fb91\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458fb91 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340540817 +0000 UTC m=+0.779797286,LastTimestamp:2026-02-25 06:45:07.629967232 +0000 UTC m=+1.069223691,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.264217 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a4592fad\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a4592fad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340554157 +0000 UTC m=+0.779810626,LastTimestamp:2026-02-25 06:45:07.629975132 +0000 UTC m=+1.069231591,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.271064 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458a028\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458a028 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340517416 +0000 UTC m=+0.779773885,LastTimestamp:2026-02-25 06:45:07.631432825 +0000 UTC m=+1.070689284,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.278119 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458fb91\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458fb91 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340540817 +0000 UTC m=+0.779797286,LastTimestamp:2026-02-25 06:45:07.631448805 +0000 UTC m=+1.070705264,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.286834 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a4592fad\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a4592fad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340554157 +0000 UTC m=+0.779810626,LastTimestamp:2026-02-25 06:45:07.631457265 +0000 UTC m=+1.070713724,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.293516 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458a028\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458a028 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340517416 +0000 UTC m=+0.779773885,LastTimestamp:2026-02-25 06:45:07.631570064 +0000 UTC m=+1.070826533,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.300217 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458fb91\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458fb91 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340540817 +0000 UTC m=+0.779797286,LastTimestamp:2026-02-25 06:45:07.631584564 +0000 UTC m=+1.070841043,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.307289 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a4592fad\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a4592fad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340554157 +0000 UTC m=+0.779810626,LastTimestamp:2026-02-25 06:45:07.631595164 +0000 UTC m=+1.070851633,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.314095 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458a028\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458a028 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340517416 +0000 UTC m=+0.779773885,LastTimestamp:2026-02-25 06:45:07.632000092 +0000 UTC m=+1.071256581,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.320989 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458fb91\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458fb91 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340540817 +0000 UTC m=+0.779797286,LastTimestamp:2026-02-25 06:45:07.632036942 +0000 UTC m=+1.071293441,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.327929 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a4592fad\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a4592fad default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340554157 +0000 UTC m=+0.779810626,LastTimestamp:2026-02-25 06:45:07.632051812 +0000 UTC m=+1.071308301,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.335128 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458a028\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458a028 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340517416 +0000 UTC m=+0.779773885,LastTimestamp:2026-02-25 06:45:07.632952467 +0000 UTC m=+1.072208926,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.340657 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18976a58a458fb91\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18976a58a458fb91 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:07.340540817 +0000 UTC m=+0.779797286,LastTimestamp:2026-02-25 06:45:07.632980797 +0000 UTC m=+1.072237256,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.349012 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a58cd179ff1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.024123377 +0000 UTC m=+1.463379836,LastTimestamp:2026-02-25 06:45:08.024123377 +0000 UTC m=+1.463379836,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.356549 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18976a58cd8dc9b7 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.031867319 +0000 UTC m=+1.471123778,LastTimestamp:2026-02-25 06:45:08.031867319 +0000 UTC m=+1.471123778,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.364179 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a58ce4435ba openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.043822522 +0000 UTC m=+1.483079021,LastTimestamp:2026-02-25 06:45:08.043822522 +0000 UTC m=+1.483079021,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.371491 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a58ceb5bcde openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.051262686 +0000 UTC m=+1.490519145,LastTimestamp:2026-02-25 06:45:08.051262686 +0000 UTC m=+1.490519145,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.379237 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a58cf7d4926 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.064340262 +0000 UTC m=+1.503596751,LastTimestamp:2026-02-25 06:45:08.064340262 +0000 UTC m=+1.503596751,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.387214 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a58f11237d8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.627748824 +0000 UTC m=+2.067005283,LastTimestamp:2026-02-25 06:45:08.627748824 +0000 UTC m=+2.067005283,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.394616 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a58f1167581 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.628026753 +0000 UTC m=+2.067283232,LastTimestamp:2026-02-25 06:45:08.628026753 +0000 UTC m=+2.067283232,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.401016 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a58f11b906c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.628361324 +0000 UTC m=+2.067617803,LastTimestamp:2026-02-25 06:45:08.628361324 +0000 UTC m=+2.067617803,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.407467 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a58f11d5f71 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.628479857 +0000 UTC m=+2.067736316,LastTimestamp:2026-02-25 06:45:08.628479857 +0000 UTC m=+2.067736316,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.414836 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18976a58f16eb2ba openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.633809594 +0000 UTC m=+2.073066073,LastTimestamp:2026-02-25 06:45:08.633809594 +0000 UTC m=+2.073066073,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.422146 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a58f1cbac7d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.639902845 +0000 UTC m=+2.079159314,LastTimestamp:2026-02-25 06:45:08.639902845 +0000 UTC m=+2.079159314,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.431807 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a58f1e85623 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.641781283 +0000 UTC m=+2.081037752,LastTimestamp:2026-02-25 06:45:08.641781283 +0000 UTC m=+2.081037752,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.439068 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a58f1ef8f24 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.642254628 +0000 UTC m=+2.081511087,LastTimestamp:2026-02-25 06:45:08.642254628 +0000 UTC m=+2.081511087,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.445883 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a58f1efd01b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.642271259 +0000 UTC m=+2.081527738,LastTimestamp:2026-02-25 06:45:08.642271259 +0000 UTC m=+2.081527738,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.452859 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a58f24b3e62 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.648263266 +0000 UTC m=+2.087519725,LastTimestamp:2026-02-25 06:45:08.648263266 +0000 UTC m=+2.087519725,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.459727 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18976a58f2f29e8f openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.659232399 +0000 UTC m=+2.098488858,LastTimestamp:2026-02-25 06:45:08.659232399 +0000 UTC m=+2.098488858,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.466178 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a590383cc4a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.937182282 +0000 UTC m=+2.376438781,LastTimestamp:2026-02-25 06:45:08.937182282 +0000 UTC m=+2.376438781,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.473053 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a590472974d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.952831821 +0000 UTC m=+2.392088320,LastTimestamp:2026-02-25 06:45:08.952831821 +0000 UTC m=+2.392088320,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.479571 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a59048abcf0 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.95441432 +0000 UTC m=+2.393670809,LastTimestamp:2026-02-25 06:45:08.95441432 +0000 UTC m=+2.393670809,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.486613 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a5912b0a94c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.191780684 +0000 UTC m=+2.631037153,LastTimestamp:2026-02-25 06:45:09.191780684 +0000 UTC m=+2.631037153,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.493327 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a59134b0cea openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.20189873 +0000 UTC m=+2.641155199,LastTimestamp:2026-02-25 06:45:09.20189873 +0000 UTC m=+2.641155199,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.500241 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a59135d8dfa openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.203111418 +0000 UTC m=+2.642367877,LastTimestamp:2026-02-25 06:45:09.203111418 +0000 UTC m=+2.642367877,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.507602 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a591bf35c3c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.347146812 +0000 UTC m=+2.786403301,LastTimestamp:2026-02-25 06:45:09.347146812 +0000 UTC m=+2.786403301,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.515858 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18976a591c2e63f5 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.351015413 +0000 UTC m=+2.790271882,LastTimestamp:2026-02-25 06:45:09.351015413 +0000 UTC m=+2.790271882,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.522653 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a591c4092ca openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.35220705 +0000 UTC m=+2.791463509,LastTimestamp:2026-02-25 06:45:09.35220705 +0000 UTC m=+2.791463509,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.530478 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a591c9c97e5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.358237669 +0000 UTC m=+2.797494148,LastTimestamp:2026-02-25 06:45:09.358237669 +0000 UTC m=+2.797494148,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.537751 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a591fa3005c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.408989276 +0000 UTC m=+2.848245745,LastTimestamp:2026-02-25 06:45:09.408989276 +0000 UTC m=+2.848245745,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.542486 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a5922baa121 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.460869409 +0000 UTC m=+2.900125878,LastTimestamp:2026-02-25 06:45:09.460869409 +0000 UTC m=+2.900125878,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.544891 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a592993398e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.575727502 +0000 UTC m=+3.014983961,LastTimestamp:2026-02-25 06:45:09.575727502 +0000 UTC m=+3.014983961,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.549337 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a59299cd552 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.576357202 +0000 UTC m=+3.015613671,LastTimestamp:2026-02-25 06:45:09.576357202 +0000 UTC m=+3.015613671,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.551710 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18976a5929ad02e5 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.577417445 +0000 UTC m=+3.016673914,LastTimestamp:2026-02-25 06:45:09.577417445 +0000 UTC m=+3.016673914,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.555938 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a5929b1a6b5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.577721525 +0000 UTC m=+3.016977984,LastTimestamp:2026-02-25 06:45:09.577721525 +0000 UTC m=+3.016977984,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.561418 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a592a54a1c4 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.588402628 +0000 UTC m=+3.027659087,LastTimestamp:2026-02-25 06:45:09.588402628 +0000 UTC m=+3.027659087,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.563728 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a592a6c62da openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.589959386 +0000 UTC m=+3.029215855,LastTimestamp:2026-02-25 06:45:09.589959386 +0000 UTC m=+3.029215855,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.568476 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a592a7da76b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.591091051 +0000 UTC m=+3.030347510,LastTimestamp:2026-02-25 06:45:09.591091051 +0000 UTC m=+3.030347510,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.573932 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a592a8cc99a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.592082842 +0000 UTC m=+3.031339301,LastTimestamp:2026-02-25 06:45:09.592082842 +0000 UTC m=+3.031339301,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.579104 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18976a592b8edf47 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.608996679 +0000 UTC m=+3.048253138,LastTimestamp:2026-02-25 06:45:09.608996679 +0000 UTC m=+3.048253138,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.585961 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a592c2a4876 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.619181686 +0000 UTC m=+3.058438135,LastTimestamp:2026-02-25 06:45:09.619181686 +0000 UTC m=+3.058438135,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.592999 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a5936605409 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.790495753 +0000 UTC m=+3.229752212,LastTimestamp:2026-02-25 06:45:09.790495753 +0000 UTC m=+3.229752212,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.599481 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5936b240ea openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.79586481 +0000 UTC m=+3.235121269,LastTimestamp:2026-02-25 06:45:09.79586481 +0000 UTC m=+3.235121269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.606435 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a59376b6bd6 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.807999958 +0000 UTC m=+3.247256417,LastTimestamp:2026-02-25 06:45:09.807999958 +0000 UTC m=+3.247256417,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.613881 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a59377e7289 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.809246857 +0000 UTC m=+3.248503306,LastTimestamp:2026-02-25 06:45:09.809246857 +0000 UTC m=+3.248503306,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.619736 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5937f8cbfa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.817265146 +0000 UTC m=+3.256521605,LastTimestamp:2026-02-25 06:45:09.817265146 +0000 UTC m=+3.256521605,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.625746 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5938413e57 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:09.822013015 +0000 UTC m=+3.261269494,LastTimestamp:2026-02-25 06:45:09.822013015 +0000 UTC m=+3.261269494,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.632785 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a59436aa118 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.009274648 +0000 UTC m=+3.448531107,LastTimestamp:2026-02-25 06:45:10.009274648 +0000 UTC m=+3.448531107,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.639056 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5943a32bea openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.012980202 +0000 UTC m=+3.452236661,LastTimestamp:2026-02-25 06:45:10.012980202 +0000 UTC m=+3.452236661,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.645092 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18976a5944e3e89c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.034000028 +0000 UTC m=+3.473256487,LastTimestamp:2026-02-25 06:45:10.034000028 +0000 UTC m=+3.473256487,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.652419 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a594521b5b9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.038050233 +0000 UTC m=+3.477306692,LastTimestamp:2026-02-25 06:45:10.038050233 +0000 UTC m=+3.477306692,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.658868 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a594530aaac openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.039030444 +0000 UTC m=+3.478286903,LastTimestamp:2026-02-25 06:45:10.039030444 +0000 UTC m=+3.478286903,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.665362 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a595313503b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.271987771 +0000 UTC m=+3.711244230,LastTimestamp:2026-02-25 06:45:10.271987771 +0000 UTC m=+3.711244230,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.671624 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5953d76313 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.284837651 +0000 UTC m=+3.724094110,LastTimestamp:2026-02-25 06:45:10.284837651 +0000 UTC m=+3.724094110,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.677821 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5953eda7d7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.286297047 +0000 UTC m=+3.725553506,LastTimestamp:2026-02-25 06:45:10.286297047 +0000 UTC m=+3.725553506,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.685252 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a595954422c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.376907308 +0000 UTC m=+3.816163767,LastTimestamp:2026-02-25 06:45:10.376907308 +0000 UTC m=+3.816163767,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.692457 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5960f67aa4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.504979108 +0000 UTC m=+3.944235567,LastTimestamp:2026-02-25 06:45:10.504979108 +0000 UTC m=+3.944235567,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.699142 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5961e32d2f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.520491311 +0000 UTC m=+3.959747770,LastTimestamp:2026-02-25 06:45:10.520491311 +0000 UTC m=+3.959747770,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.706892 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59641883d8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.557541336 +0000 UTC m=+3.996797815,LastTimestamp:2026-02-25 06:45:10.557541336 +0000 UTC m=+3.996797815,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.713520 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a5965146d00 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.57405056 +0000 UTC m=+4.013307049,LastTimestamp:2026-02-25 06:45:10.57405056 +0000 UTC m=+4.013307049,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.721423 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a5996fac39d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:11.411229597 +0000 UTC m=+4.850486046,LastTimestamp:2026-02-25 06:45:11.411229597 +0000 UTC m=+4.850486046,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.727946 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59a4dd83f8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:11.644193784 +0000 UTC m=+5.083450243,LastTimestamp:2026-02-25 06:45:11.644193784 +0000 UTC m=+5.083450243,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.733106 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59b1481dd6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:11.852506582 +0000 UTC m=+5.291763061,LastTimestamp:2026-02-25 06:45:11.852506582 +0000 UTC m=+5.291763061,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.739572 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59b1685f3e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:11.854620478 +0000 UTC m=+5.293876937,LastTimestamp:2026-02-25 06:45:11.854620478 +0000 UTC m=+5.293876937,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.745597 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59c07d397e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.10764531 +0000 UTC m=+5.546901809,LastTimestamp:2026-02-25 06:45:12.10764531 +0000 UTC m=+5.546901809,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.751972 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59c185dfba openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.12498937 +0000 UTC m=+5.564245869,LastTimestamp:2026-02-25 06:45:12.12498937 +0000 UTC m=+5.564245869,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.757223 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59c1a343a2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.12691549 +0000 UTC m=+5.566171979,LastTimestamp:2026-02-25 06:45:12.12691549 +0000 UTC m=+5.566171979,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.763570 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59cf8398fb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.359721211 +0000 UTC m=+5.798977670,LastTimestamp:2026-02-25 06:45:12.359721211 +0000 UTC m=+5.798977670,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.770541 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59d06352ea openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.374383338 +0000 UTC m=+5.813639817,LastTimestamp:2026-02-25 06:45:12.374383338 +0000 UTC m=+5.813639817,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.777551 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59d074f852 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.375539794 +0000 UTC m=+5.814796253,LastTimestamp:2026-02-25 06:45:12.375539794 +0000 UTC m=+5.814796253,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.786223 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59df6ca399 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.626652057 +0000 UTC m=+6.065908556,LastTimestamp:2026-02-25 06:45:12.626652057 +0000 UTC m=+6.065908556,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.793844 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59e07ecc3e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.644619326 +0000 UTC m=+6.083875825,LastTimestamp:2026-02-25 06:45:12.644619326 +0000 UTC m=+6.083875825,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.801274 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59e0916b1c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.645839644 +0000 UTC m=+6.085096133,LastTimestamp:2026-02-25 06:45:12.645839644 +0000 UTC m=+6.085096133,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.808069 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59ef274af4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.890542836 +0000 UTC m=+6.329799305,LastTimestamp:2026-02-25 06:45:12.890542836 +0000 UTC m=+6.329799305,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.814987 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18976a59efe74e33 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:12.903126579 +0000 UTC m=+6.342383048,LastTimestamp:2026-02-25 06:45:12.903126579 +0000 UTC m=+6.342383048,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.823606 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 25 06:45:52 crc kubenswrapper[4978]: &Event{ObjectMeta:{kube-controller-manager-crc.18976a5ab0359c34 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Feb 25 06:45:52 crc kubenswrapper[4978]: body: Feb 25 06:45:52 crc kubenswrapper[4978]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:16.129483828 +0000 UTC m=+9.568740327,LastTimestamp:2026-02-25 06:45:16.129483828 +0000 UTC m=+9.568740327,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 06:45:52 crc kubenswrapper[4978]: > Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.830765 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a5ab036fc98 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:16.12957404 +0000 UTC m=+9.568830539,LastTimestamp:2026-02-25 06:45:16.12957404 +0000 UTC m=+9.568830539,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.838110 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 25 06:45:52 crc kubenswrapper[4978]: &Event{ObjectMeta:{kube-apiserver-crc.18976a5c18f2b15d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 25 06:45:52 crc kubenswrapper[4978]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 25 06:45:52 crc kubenswrapper[4978]: Feb 25 06:45:52 crc kubenswrapper[4978]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:22.181673309 +0000 UTC m=+15.620929788,LastTimestamp:2026-02-25 06:45:22.181673309 +0000 UTC m=+15.620929788,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 06:45:52 crc kubenswrapper[4978]: > Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.844893 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5c18f3862f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:22.181727791 +0000 UTC m=+15.620984280,LastTimestamp:2026-02-25 06:45:22.181727791 +0000 UTC m=+15.620984280,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.851742 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 25 06:45:52 crc kubenswrapper[4978]: &Event{ObjectMeta:{kube-apiserver-crc.18976a5c1940aa06 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 25 06:45:52 crc kubenswrapper[4978]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 25 06:45:52 crc kubenswrapper[4978]: Feb 25 06:45:52 crc kubenswrapper[4978]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:22.186783238 +0000 UTC m=+15.626039717,LastTimestamp:2026-02-25 06:45:22.186783238 +0000 UTC m=+15.626039717,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 06:45:52 crc kubenswrapper[4978]: > Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.857842 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18976a5c18f3862f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5c18f3862f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:22.181727791 +0000 UTC m=+15.620984280,LastTimestamp:2026-02-25 06:45:22.186893172 +0000 UTC m=+15.626149641,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.865012 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 25 06:45:52 crc kubenswrapper[4978]: &Event{ObjectMeta:{kube-apiserver-crc.18976a5c197289ce openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:17697/healthz": read tcp 192.168.126.11:33338->192.168.126.11:17697: read: connection reset by peer Feb 25 06:45:52 crc kubenswrapper[4978]: body: Feb 25 06:45:52 crc kubenswrapper[4978]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:22.19005179 +0000 UTC m=+15.629308259,LastTimestamp:2026-02-25 06:45:22.19005179 +0000 UTC m=+15.629308259,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 06:45:52 crc kubenswrapper[4978]: > Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.871707 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5c1973285e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:33338->192.168.126.11:17697: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:22.190092382 +0000 UTC m=+15.629348861,LastTimestamp:2026-02-25 06:45:22.190092382 +0000 UTC m=+15.629348861,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.879091 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18976a5953eda7d7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18976a5953eda7d7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:10.286297047 +0000 UTC m=+3.725553506,LastTimestamp:2026-02-25 06:45:22.457928014 +0000 UTC m=+15.897184513,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.886874 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 25 06:45:52 crc kubenswrapper[4978]: &Event{ObjectMeta:{kube-controller-manager-crc.18976a5d044ca49a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 25 06:45:52 crc kubenswrapper[4978]: body: Feb 25 06:45:52 crc kubenswrapper[4978]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:26.130214042 +0000 UTC m=+19.569470501,LastTimestamp:2026-02-25 06:45:26.130214042 +0000 UTC m=+19.569470501,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 06:45:52 crc kubenswrapper[4978]: > Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.893467 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a5d044d8fba openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:26.130274234 +0000 UTC m=+19.569530703,LastTimestamp:2026-02-25 06:45:26.130274234 +0000 UTC m=+19.569530703,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.901600 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18976a5d044ca49a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 25 06:45:52 crc kubenswrapper[4978]: &Event{ObjectMeta:{kube-controller-manager-crc.18976a5d044ca49a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 25 06:45:52 crc kubenswrapper[4978]: body: Feb 25 06:45:52 crc kubenswrapper[4978]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:26.130214042 +0000 UTC m=+19.569470501,LastTimestamp:2026-02-25 06:45:36.131024652 +0000 UTC m=+29.570281111,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 06:45:52 crc kubenswrapper[4978]: > Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.906819 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18976a5d044d8fba\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a5d044d8fba openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:26.130274234 +0000 UTC m=+19.569530703,LastTimestamp:2026-02-25 06:45:36.131094164 +0000 UTC m=+29.570350623,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.914844 4978 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a5f58967632 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:36.134272562 +0000 UTC m=+29.573529081,LastTimestamp:2026-02-25 06:45:36.134272562 +0000 UTC m=+29.573529081,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.923112 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18976a58f1e85623\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a58f1e85623 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.641781283 +0000 UTC m=+2.081037752,LastTimestamp:2026-02-25 06:45:36.264237641 +0000 UTC m=+29.703494130,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.928867 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18976a590383cc4a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a590383cc4a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.937182282 +0000 UTC m=+2.376438781,LastTimestamp:2026-02-25 06:45:36.491466377 +0000 UTC m=+29.930722826,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.934433 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18976a590472974d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a590472974d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:08.952831821 +0000 UTC m=+2.392088320,LastTimestamp:2026-02-25 06:45:36.505282794 +0000 UTC m=+29.944539263,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.943202 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18976a5d044ca49a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 25 06:45:52 crc kubenswrapper[4978]: &Event{ObjectMeta:{kube-controller-manager-crc.18976a5d044ca49a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 25 06:45:52 crc kubenswrapper[4978]: body: Feb 25 06:45:52 crc kubenswrapper[4978]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:26.130214042 +0000 UTC m=+19.569470501,LastTimestamp:2026-02-25 06:45:46.130737246 +0000 UTC m=+39.569993755,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 06:45:52 crc kubenswrapper[4978]: > Feb 25 06:45:52 crc kubenswrapper[4978]: E0225 06:45:52.950192 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18976a5d044d8fba\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18976a5d044d8fba openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:26.130274234 +0000 UTC m=+19.569530703,LastTimestamp:2026-02-25 06:45:46.130798097 +0000 UTC m=+39.570054596,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:45:53 crc kubenswrapper[4978]: I0225 06:45:53.253617 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:54 crc kubenswrapper[4978]: I0225 06:45:54.250681 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:54 crc kubenswrapper[4978]: W0225 06:45:54.674551 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 25 06:45:54 crc kubenswrapper[4978]: E0225 06:45:54.674637 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 25 06:45:54 crc kubenswrapper[4978]: W0225 06:45:54.775873 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Feb 25 06:45:54 crc kubenswrapper[4978]: E0225 06:45:54.775928 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 25 06:45:55 crc kubenswrapper[4978]: I0225 06:45:55.251324 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:55 crc kubenswrapper[4978]: W0225 06:45:55.371059 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:55 crc kubenswrapper[4978]: E0225 06:45:55.371137 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.130292 4978 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.130430 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 06:45:56 crc kubenswrapper[4978]: E0225 06:45:56.138121 4978 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18976a5d044ca49a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 25 06:45:56 crc kubenswrapper[4978]: &Event{ObjectMeta:{kube-controller-manager-crc.18976a5d044ca49a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 25 06:45:56 crc kubenswrapper[4978]: body: Feb 25 06:45:56 crc kubenswrapper[4978]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:45:26.130214042 +0000 UTC m=+19.569470501,LastTimestamp:2026-02-25 06:45:56.130401845 +0000 UTC m=+49.569658304,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 25 06:45:56 crc kubenswrapper[4978]: > Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.251574 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.366841 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.367183 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.368878 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.368937 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.368962 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.369831 4978 scope.go:117] "RemoveContainer" containerID="bdd08f3895f5ffcce127508a45c274bf2fe04457b344de1e985cb2978c68fa57" Feb 25 06:45:56 crc kubenswrapper[4978]: E0225 06:45:56.370116 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:45:56 crc kubenswrapper[4978]: E0225 06:45:56.595091 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.605220 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.606939 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.607019 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.607044 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:45:56 crc kubenswrapper[4978]: I0225 06:45:56.607078 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:45:56 crc kubenswrapper[4978]: E0225 06:45:56.614878 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 25 06:45:57 crc kubenswrapper[4978]: I0225 06:45:57.250313 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:57 crc kubenswrapper[4978]: E0225 06:45:57.441971 4978 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 06:45:58 crc kubenswrapper[4978]: I0225 06:45:58.253799 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:45:59 crc kubenswrapper[4978]: I0225 06:45:59.248353 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:00 crc kubenswrapper[4978]: I0225 06:46:00.251016 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:00 crc kubenswrapper[4978]: W0225 06:46:00.708590 4978 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 25 06:46:00 crc kubenswrapper[4978]: E0225 06:46:00.708677 4978 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 25 06:46:01 crc kubenswrapper[4978]: I0225 06:46:01.247976 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:02 crc kubenswrapper[4978]: I0225 06:46:02.251054 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:02 crc kubenswrapper[4978]: I0225 06:46:02.313249 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 25 06:46:02 crc kubenswrapper[4978]: I0225 06:46:02.313462 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:02 crc kubenswrapper[4978]: I0225 06:46:02.314809 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:02 crc kubenswrapper[4978]: I0225 06:46:02.314856 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:02 crc kubenswrapper[4978]: I0225 06:46:02.314872 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.136502 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.136742 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.138532 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.138639 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.138666 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.143282 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.251661 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:03 crc kubenswrapper[4978]: E0225 06:46:03.600488 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.607657 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.608734 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.608806 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.608828 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.615571 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.616632 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.616714 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.616743 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:03 crc kubenswrapper[4978]: I0225 06:46:03.616787 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:46:03 crc kubenswrapper[4978]: E0225 06:46:03.621462 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 25 06:46:04 crc kubenswrapper[4978]: I0225 06:46:04.251104 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:05 crc kubenswrapper[4978]: I0225 06:46:05.250965 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:06 crc kubenswrapper[4978]: I0225 06:46:06.252021 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:07 crc kubenswrapper[4978]: I0225 06:46:07.249139 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:07 crc kubenswrapper[4978]: E0225 06:46:07.442814 4978 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 06:46:08 crc kubenswrapper[4978]: I0225 06:46:08.249653 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.251588 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.327354 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.329036 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.329134 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.329163 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.330264 4978 scope.go:117] "RemoveContainer" containerID="bdd08f3895f5ffcce127508a45c274bf2fe04457b344de1e985cb2978c68fa57" Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.626353 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.628579 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5"} Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.628913 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.630010 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.630145 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:09 crc kubenswrapper[4978]: I0225 06:46:09.630234 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.250023 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:10 crc kubenswrapper[4978]: E0225 06:46:10.606280 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.622615 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.623879 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.623913 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.623926 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.623951 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:46:10 crc kubenswrapper[4978]: E0225 06:46:10.627751 4978 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.632523 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.633009 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.634572 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5" exitCode=255 Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.634615 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5"} Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.634658 4978 scope.go:117] "RemoveContainer" containerID="bdd08f3895f5ffcce127508a45c274bf2fe04457b344de1e985cb2978c68fa57" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.634845 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.636085 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.636139 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.636154 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:10 crc kubenswrapper[4978]: I0225 06:46:10.636940 4978 scope.go:117] "RemoveContainer" containerID="4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5" Feb 25 06:46:10 crc kubenswrapper[4978]: E0225 06:46:10.637146 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:46:11 crc kubenswrapper[4978]: I0225 06:46:11.250643 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:11 crc kubenswrapper[4978]: I0225 06:46:11.289640 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:46:11 crc kubenswrapper[4978]: I0225 06:46:11.639992 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 25 06:46:11 crc kubenswrapper[4978]: I0225 06:46:11.643756 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:11 crc kubenswrapper[4978]: I0225 06:46:11.645065 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:11 crc kubenswrapper[4978]: I0225 06:46:11.645140 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:11 crc kubenswrapper[4978]: I0225 06:46:11.645160 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:11 crc kubenswrapper[4978]: I0225 06:46:11.646236 4978 scope.go:117] "RemoveContainer" containerID="4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5" Feb 25 06:46:11 crc kubenswrapper[4978]: E0225 06:46:11.646633 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:46:12 crc kubenswrapper[4978]: I0225 06:46:12.249983 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:13 crc kubenswrapper[4978]: I0225 06:46:13.275727 4978 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 25 06:46:13 crc kubenswrapper[4978]: I0225 06:46:13.656811 4978 csr.go:261] certificate signing request csr-r5m6g is approved, waiting to be issued Feb 25 06:46:13 crc kubenswrapper[4978]: I0225 06:46:13.666318 4978 csr.go:257] certificate signing request csr-r5m6g is issued Feb 25 06:46:13 crc kubenswrapper[4978]: I0225 06:46:13.706133 4978 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 25 06:46:14 crc kubenswrapper[4978]: I0225 06:46:14.094483 4978 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 25 06:46:14 crc kubenswrapper[4978]: I0225 06:46:14.667491 4978 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-16 13:40:10.966515767 +0000 UTC Feb 25 06:46:14 crc kubenswrapper[4978]: I0225 06:46:14.667534 4978 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6342h53m56.298985551s for next certificate rotation Feb 25 06:46:16 crc kubenswrapper[4978]: I0225 06:46:16.367126 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:46:16 crc kubenswrapper[4978]: I0225 06:46:16.368309 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:16 crc kubenswrapper[4978]: I0225 06:46:16.369988 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:16 crc kubenswrapper[4978]: I0225 06:46:16.370032 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:16 crc kubenswrapper[4978]: I0225 06:46:16.370042 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:16 crc kubenswrapper[4978]: I0225 06:46:16.370594 4978 scope.go:117] "RemoveContainer" containerID="4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5" Feb 25 06:46:16 crc kubenswrapper[4978]: E0225 06:46:16.370830 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:46:17 crc kubenswrapper[4978]: E0225 06:46:17.443858 4978 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.628478 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.630027 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.630074 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.630085 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.630190 4978 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.639936 4978 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.640311 4978 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 25 06:46:17 crc kubenswrapper[4978]: E0225 06:46:17.640357 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.645949 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.646049 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.646069 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.646216 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.646238 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:17Z","lastTransitionTime":"2026-02-25T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:17 crc kubenswrapper[4978]: E0225 06:46:17.662222 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.672725 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.672756 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.672766 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.672780 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.672789 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:17Z","lastTransitionTime":"2026-02-25T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:17 crc kubenswrapper[4978]: E0225 06:46:17.686412 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.693114 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.693185 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.693197 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.693225 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.693239 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:17Z","lastTransitionTime":"2026-02-25T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:17 crc kubenswrapper[4978]: E0225 06:46:17.704578 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.712633 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.712673 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.712681 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.712700 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:17 crc kubenswrapper[4978]: I0225 06:46:17.712713 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:17Z","lastTransitionTime":"2026-02-25T06:46:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:17 crc kubenswrapper[4978]: E0225 06:46:17.723727 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:17 crc kubenswrapper[4978]: E0225 06:46:17.723852 4978 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 25 06:46:17 crc kubenswrapper[4978]: E0225 06:46:17.723879 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:17 crc kubenswrapper[4978]: E0225 06:46:17.824982 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:17 crc kubenswrapper[4978]: E0225 06:46:17.926065 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:18 crc kubenswrapper[4978]: E0225 06:46:18.026719 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:18 crc kubenswrapper[4978]: E0225 06:46:18.127182 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:18 crc kubenswrapper[4978]: E0225 06:46:18.228257 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:18 crc kubenswrapper[4978]: E0225 06:46:18.329130 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:18 crc kubenswrapper[4978]: E0225 06:46:18.430311 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:18 crc kubenswrapper[4978]: E0225 06:46:18.530878 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:18 crc kubenswrapper[4978]: E0225 06:46:18.631056 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:18 crc kubenswrapper[4978]: E0225 06:46:18.731841 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:18 crc kubenswrapper[4978]: E0225 06:46:18.832714 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:18 crc kubenswrapper[4978]: E0225 06:46:18.933292 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:19 crc kubenswrapper[4978]: E0225 06:46:19.033613 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:19 crc kubenswrapper[4978]: E0225 06:46:19.134780 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:19 crc kubenswrapper[4978]: E0225 06:46:19.234917 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:19 crc kubenswrapper[4978]: E0225 06:46:19.335275 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:19 crc kubenswrapper[4978]: E0225 06:46:19.436120 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:19 crc kubenswrapper[4978]: E0225 06:46:19.537137 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:19 crc kubenswrapper[4978]: E0225 06:46:19.638626 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:19 crc kubenswrapper[4978]: E0225 06:46:19.739797 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:19 crc kubenswrapper[4978]: E0225 06:46:19.840549 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:19 crc kubenswrapper[4978]: E0225 06:46:19.941452 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:20 crc kubenswrapper[4978]: E0225 06:46:20.042322 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:20 crc kubenswrapper[4978]: E0225 06:46:20.142747 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:20 crc kubenswrapper[4978]: E0225 06:46:20.243836 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:20 crc kubenswrapper[4978]: E0225 06:46:20.344992 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:20 crc kubenswrapper[4978]: E0225 06:46:20.445215 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:20 crc kubenswrapper[4978]: E0225 06:46:20.545948 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:20 crc kubenswrapper[4978]: E0225 06:46:20.646580 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:20 crc kubenswrapper[4978]: E0225 06:46:20.747061 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:20 crc kubenswrapper[4978]: E0225 06:46:20.847953 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:20 crc kubenswrapper[4978]: E0225 06:46:20.948923 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:21 crc kubenswrapper[4978]: E0225 06:46:21.049730 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:21 crc kubenswrapper[4978]: E0225 06:46:21.149891 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:21 crc kubenswrapper[4978]: E0225 06:46:21.250904 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:21 crc kubenswrapper[4978]: E0225 06:46:21.351776 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:21 crc kubenswrapper[4978]: E0225 06:46:21.451941 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:21 crc kubenswrapper[4978]: E0225 06:46:21.553024 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:21 crc kubenswrapper[4978]: E0225 06:46:21.653517 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:21 crc kubenswrapper[4978]: E0225 06:46:21.753640 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:21 crc kubenswrapper[4978]: E0225 06:46:21.853877 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:21 crc kubenswrapper[4978]: E0225 06:46:21.954835 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:22 crc kubenswrapper[4978]: E0225 06:46:22.054912 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:22 crc kubenswrapper[4978]: E0225 06:46:22.155097 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:22 crc kubenswrapper[4978]: E0225 06:46:22.265356 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:22 crc kubenswrapper[4978]: E0225 06:46:22.366302 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:22 crc kubenswrapper[4978]: E0225 06:46:22.467218 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:22 crc kubenswrapper[4978]: E0225 06:46:22.567765 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:22 crc kubenswrapper[4978]: E0225 06:46:22.668108 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:22 crc kubenswrapper[4978]: E0225 06:46:22.768184 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:22 crc kubenswrapper[4978]: E0225 06:46:22.869282 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:22 crc kubenswrapper[4978]: E0225 06:46:22.970231 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:23 crc kubenswrapper[4978]: E0225 06:46:23.070646 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:23 crc kubenswrapper[4978]: E0225 06:46:23.171470 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:23 crc kubenswrapper[4978]: E0225 06:46:23.272362 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:23 crc kubenswrapper[4978]: E0225 06:46:23.372619 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:23 crc kubenswrapper[4978]: E0225 06:46:23.473566 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:23 crc kubenswrapper[4978]: E0225 06:46:23.574574 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:23 crc kubenswrapper[4978]: E0225 06:46:23.675643 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:23 crc kubenswrapper[4978]: E0225 06:46:23.776290 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:23 crc kubenswrapper[4978]: E0225 06:46:23.876717 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:23 crc kubenswrapper[4978]: E0225 06:46:23.977985 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:24 crc kubenswrapper[4978]: E0225 06:46:24.079086 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:24 crc kubenswrapper[4978]: E0225 06:46:24.179669 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:24 crc kubenswrapper[4978]: E0225 06:46:24.280567 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:24 crc kubenswrapper[4978]: E0225 06:46:24.381046 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:24 crc kubenswrapper[4978]: E0225 06:46:24.481273 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:24 crc kubenswrapper[4978]: E0225 06:46:24.581424 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:24 crc kubenswrapper[4978]: E0225 06:46:24.682671 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:24 crc kubenswrapper[4978]: E0225 06:46:24.811165 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:24 crc kubenswrapper[4978]: E0225 06:46:24.911308 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:25 crc kubenswrapper[4978]: E0225 06:46:25.012476 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:25 crc kubenswrapper[4978]: E0225 06:46:25.113635 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:25 crc kubenswrapper[4978]: E0225 06:46:25.214010 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:25 crc kubenswrapper[4978]: E0225 06:46:25.314636 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:25 crc kubenswrapper[4978]: E0225 06:46:25.414867 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:25 crc kubenswrapper[4978]: E0225 06:46:25.515394 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:25 crc kubenswrapper[4978]: E0225 06:46:25.615600 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:25 crc kubenswrapper[4978]: E0225 06:46:25.715942 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:25 crc kubenswrapper[4978]: E0225 06:46:25.816556 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:25 crc kubenswrapper[4978]: E0225 06:46:25.916803 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:26 crc kubenswrapper[4978]: E0225 06:46:26.018018 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:26 crc kubenswrapper[4978]: E0225 06:46:26.118294 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:26 crc kubenswrapper[4978]: E0225 06:46:26.219088 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:26 crc kubenswrapper[4978]: E0225 06:46:26.320168 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:26 crc kubenswrapper[4978]: E0225 06:46:26.420858 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:26 crc kubenswrapper[4978]: E0225 06:46:26.521854 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:26 crc kubenswrapper[4978]: E0225 06:46:26.622579 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:26 crc kubenswrapper[4978]: E0225 06:46:26.723580 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:26 crc kubenswrapper[4978]: E0225 06:46:26.824253 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:26 crc kubenswrapper[4978]: E0225 06:46:26.924448 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.025752 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.126142 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.227356 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.328612 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.428756 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.445003 4978 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.529068 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.629493 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.730599 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.831352 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.914441 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.925075 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.925142 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.925166 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.925195 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.925216 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:27Z","lastTransitionTime":"2026-02-25T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.941719 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.952327 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.952408 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.952426 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.952453 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.952475 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:27Z","lastTransitionTime":"2026-02-25T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.968777 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.980470 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.980601 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.980630 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.980655 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:27 crc kubenswrapper[4978]: I0225 06:46:27.980673 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:27Z","lastTransitionTime":"2026-02-25T06:46:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:27 crc kubenswrapper[4978]: E0225 06:46:27.998894 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:28 crc kubenswrapper[4978]: I0225 06:46:28.010290 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:28 crc kubenswrapper[4978]: I0225 06:46:28.010396 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:28 crc kubenswrapper[4978]: I0225 06:46:28.010422 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:28 crc kubenswrapper[4978]: I0225 06:46:28.010447 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:28 crc kubenswrapper[4978]: I0225 06:46:28.010467 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:28Z","lastTransitionTime":"2026-02-25T06:46:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.028890 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.029045 4978 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.029075 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.129969 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.230341 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.331357 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.432033 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.532656 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.633174 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.734040 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.834823 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:28 crc kubenswrapper[4978]: E0225 06:46:28.935054 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:29 crc kubenswrapper[4978]: E0225 06:46:29.035893 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:29 crc kubenswrapper[4978]: E0225 06:46:29.136608 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:29 crc kubenswrapper[4978]: E0225 06:46:29.237335 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.327070 4978 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.328422 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.328456 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.328465 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.329155 4978 scope.go:117] "RemoveContainer" containerID="4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5" Feb 25 06:46:29 crc kubenswrapper[4978]: E0225 06:46:29.329343 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:46:29 crc kubenswrapper[4978]: E0225 06:46:29.337932 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:29 crc kubenswrapper[4978]: E0225 06:46:29.438257 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:29 crc kubenswrapper[4978]: E0225 06:46:29.539475 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:29 crc kubenswrapper[4978]: E0225 06:46:29.640182 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:29 crc kubenswrapper[4978]: E0225 06:46:29.741240 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:29 crc kubenswrapper[4978]: E0225 06:46:29.841727 4978 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.912085 4978 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.944332 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.944410 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.944432 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.944456 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:29 crc kubenswrapper[4978]: I0225 06:46:29.944473 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:29Z","lastTransitionTime":"2026-02-25T06:46:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.046945 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.046985 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.047002 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.047024 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.047040 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:30Z","lastTransitionTime":"2026-02-25T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.151122 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.151193 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.151212 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.151238 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.151259 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:30Z","lastTransitionTime":"2026-02-25T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.254309 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.254364 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.254419 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.254444 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.254462 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:30Z","lastTransitionTime":"2026-02-25T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.272611 4978 apiserver.go:52] "Watching apiserver" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.278890 4978 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.279271 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.279892 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.279988 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.280082 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.280203 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.280297 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.280538 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.280574 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.280663 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.281194 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.283516 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.283917 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.284451 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.284689 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.284868 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.285028 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.285483 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.286307 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.286861 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.333040 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.352473 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.356053 4978 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.356871 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.356922 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.356937 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.356960 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.356983 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:30Z","lastTransitionTime":"2026-02-25T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.368628 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.383918 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.393647 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.403898 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.422267 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.434480 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.447750 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.447803 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.447844 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.447879 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.447912 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.447945 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.447975 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448000 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448032 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448060 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448089 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448117 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448145 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448184 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448216 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448248 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448282 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448314 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448343 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448402 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448439 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448470 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448500 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448531 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.448562 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.449055 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.449151 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.449523 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.449550 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.449765 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.450057 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.450087 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.450330 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.450482 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.450553 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451481 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451559 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451614 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451658 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451700 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451740 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451783 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451945 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451995 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452042 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452088 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452136 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452180 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452227 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452271 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452324 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452413 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452468 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452519 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452567 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452610 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452696 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.450856 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451335 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451467 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451514 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.451547 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452549 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452578 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452638 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.453193 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452684 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.452741 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.453606 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.453905 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.454174 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.453755 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.454798 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.454750 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.454923 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455012 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455101 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455122 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455154 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455200 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455299 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455428 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455529 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455644 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456485 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456569 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456638 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456680 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456743 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456800 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456906 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456948 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457019 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457059 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457127 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457186 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457232 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457297 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457355 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457417 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457454 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457533 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457590 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457616 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457641 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457694 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457717 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457764 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457788 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457810 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457850 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457873 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457894 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457938 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.458015 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.458039 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.458061 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.458195 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.458275 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.458359 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.458515 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.458604 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.458692 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459096 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459193 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459360 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459437 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459459 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459569 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459594 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459646 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459669 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459782 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459814 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459880 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459901 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459924 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459948 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459974 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459998 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460022 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460045 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460069 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460096 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460119 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460145 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460173 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460197 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460221 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460246 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460275 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460915 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460944 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460993 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461030 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461050 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461095 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461112 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461131 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461167 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461187 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461206 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461261 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461286 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461307 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461349 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455657 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455229 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.455996 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456219 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456319 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456342 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.456550 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457024 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457185 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457234 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461883 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457231 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.457424 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459608 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459671 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.459939 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460168 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460255 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460287 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460527 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460527 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460643 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460785 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.460773 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461159 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461504 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461547 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461552 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.462157 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461712 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.462020 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461935 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.462188 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.462301 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.462739 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.462762 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.463322 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.463496 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.463662 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.463674 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.463993 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464019 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464119 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464203 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464169 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464378 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464441 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.461512 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464513 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464543 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464570 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464596 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464610 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464638 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464725 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464753 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464779 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464839 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464848 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464868 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464896 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464920 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464947 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.464974 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465003 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465027 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465056 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465066 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465094 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465097 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465138 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465175 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465210 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465243 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465268 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465291 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465315 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465346 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465397 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465422 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465444 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465469 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465495 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465518 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465543 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465566 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465590 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465614 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465637 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466734 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466773 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466800 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466825 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465058 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466854 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465053 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465097 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465252 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465105 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466933 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466951 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:30Z","lastTransitionTime":"2026-02-25T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466884 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465437 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465526 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465563 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.465586 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466037 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466226 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466220 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466452 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466502 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466472 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466538 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466585 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466677 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466700 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.466777 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.467274 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.467321 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.467394 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.467455 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:46:30.967429904 +0000 UTC m=+84.406686373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469573 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469575 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469604 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469619 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469627 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469725 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469755 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469779 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469800 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469821 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469896 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469948 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469971 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469969 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469993 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470072 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470168 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470187 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470237 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.467513 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.467851 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.467892 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.467915 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.468115 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.468695 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.468927 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.469278 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470213 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470398 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470428 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470451 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470470 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470487 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470505 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470522 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470546 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470537 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470706 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470724 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470734 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470744 4978 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470755 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470764 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470773 4978 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470783 4978 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470791 4978 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470801 4978 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470809 4978 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470818 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470832 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470841 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470850 4978 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470858 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470867 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470877 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470885 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470894 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470904 4978 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470913 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470924 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470934 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470943 4978 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470955 4978 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470965 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470975 4978 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470985 4978 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.470994 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471004 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471013 4978 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471022 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471031 4978 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471040 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471050 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471062 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471084 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471096 4978 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471111 4978 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471128 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471142 4978 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471151 4978 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471160 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471168 4978 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471177 4978 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471186 4978 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471195 4978 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471204 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471213 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471223 4978 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471230 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471305 4978 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471324 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471337 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471349 4978 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471378 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471390 4978 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471401 4978 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471412 4978 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471424 4978 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471435 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471446 4978 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471458 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471469 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471481 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471492 4978 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471504 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471517 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471528 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471540 4978 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471552 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471563 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472537 4978 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472577 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472586 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472597 4978 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472607 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472618 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472629 4978 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472639 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472651 4978 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472660 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472670 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472681 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472690 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472700 4978 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472959 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472970 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472980 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472988 4978 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472999 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473009 4978 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473019 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473029 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473038 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473049 4978 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473059 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473070 4978 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473082 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471946 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471602 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471878 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.471954 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.471969 4978 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472010 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472240 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.472467 4978 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472519 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472619 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.472909 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473178 4978 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473334 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473471 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473499 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473497 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473531 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473546 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473619 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473778 4978 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473818 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.473863 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:30.973647003 +0000 UTC m=+84.412903492 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473910 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.473912 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.474711 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.475053 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:30.973920302 +0000 UTC m=+84.413176801 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.475132 4978 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.475565 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.475723 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.476188 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.476209 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.476246 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.476259 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.476253 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.476534 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.477248 4978 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.477274 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.477281 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.477329 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.477342 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.477353 4978 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.477704 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.479179 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.479676 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.479726 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.479406 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.479791 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.479973 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.480268 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.480304 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.480901 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.481872 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.484855 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.484996 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.489147 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.489191 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.489218 4978 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.489331 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:30.989301768 +0000 UTC m=+84.428558437 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.489929 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.490087 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.490160 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.490965 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.491302 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.491589 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.492238 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.492475 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.492542 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.492788 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.492970 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.493118 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.494003 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.494032 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.494064 4978 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.494123 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:30.994102994 +0000 UTC m=+84.433359453 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.494133 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.494248 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.494595 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.494720 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.494855 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.494880 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.494359 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.496918 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.496935 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.498042 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.498164 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.498085 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.498301 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.498483 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.498540 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.498742 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.499484 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.499514 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.499962 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.500109 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.500209 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.500307 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.500324 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.500425 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.500560 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.500712 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.500734 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.501021 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.501162 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.503461 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.503737 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.503798 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.503833 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.503894 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.503907 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.506616 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.520876 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.527442 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.533465 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.536951 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.570417 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.570455 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.570464 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.570479 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.570490 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:30Z","lastTransitionTime":"2026-02-25T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577664 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577698 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577752 4978 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577769 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577781 4978 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577794 4978 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577806 4978 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577819 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577820 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577831 4978 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577889 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577897 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577924 4978 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577939 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577952 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577966 4978 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577978 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.577990 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578002 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578014 4978 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578026 4978 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578038 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578050 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578061 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578073 4978 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578086 4978 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578097 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578109 4978 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578121 4978 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578133 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578144 4978 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578156 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578168 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578179 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578191 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578202 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578214 4978 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578225 4978 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578237 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578248 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578260 4978 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578272 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578284 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578295 4978 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578308 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578321 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578334 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578345 4978 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578358 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578389 4978 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578401 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578413 4978 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578424 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578436 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578449 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578460 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578472 4978 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578483 4978 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578494 4978 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578506 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578517 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578528 4978 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578544 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578558 4978 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578570 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578584 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578598 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578612 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578626 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578640 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578657 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578671 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578684 4978 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578698 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578710 4978 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578724 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578736 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578749 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578761 4978 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578774 4978 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578789 4978 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578802 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578815 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578827 4978 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578838 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578849 4978 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578860 4978 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578875 4978 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578889 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.578900 4978 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.605978 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.619348 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.630139 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.635230 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 06:46:30 crc kubenswrapper[4978]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Feb 25 06:46:30 crc kubenswrapper[4978]: if [[ -f "/env/_master" ]]; then Feb 25 06:46:30 crc kubenswrapper[4978]: set -o allexport Feb 25 06:46:30 crc kubenswrapper[4978]: source "/env/_master" Feb 25 06:46:30 crc kubenswrapper[4978]: set +o allexport Feb 25 06:46:30 crc kubenswrapper[4978]: fi Feb 25 06:46:30 crc kubenswrapper[4978]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Feb 25 06:46:30 crc kubenswrapper[4978]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Feb 25 06:46:30 crc kubenswrapper[4978]: ho_enable="--enable-hybrid-overlay" Feb 25 06:46:30 crc kubenswrapper[4978]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Feb 25 06:46:30 crc kubenswrapper[4978]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Feb 25 06:46:30 crc kubenswrapper[4978]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Feb 25 06:46:30 crc kubenswrapper[4978]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Feb 25 06:46:30 crc kubenswrapper[4978]: --webhook-cert-dir="/etc/webhook-cert" \ Feb 25 06:46:30 crc kubenswrapper[4978]: --webhook-host=127.0.0.1 \ Feb 25 06:46:30 crc kubenswrapper[4978]: --webhook-port=9743 \ Feb 25 06:46:30 crc kubenswrapper[4978]: ${ho_enable} \ Feb 25 06:46:30 crc kubenswrapper[4978]: --enable-interconnect \ Feb 25 06:46:30 crc kubenswrapper[4978]: --disable-approver \ Feb 25 06:46:30 crc kubenswrapper[4978]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Feb 25 06:46:30 crc kubenswrapper[4978]: --wait-for-kubernetes-api=200s \ Feb 25 06:46:30 crc kubenswrapper[4978]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Feb 25 06:46:30 crc kubenswrapper[4978]: --loglevel="${LOGLEVEL}" Feb 25 06:46:30 crc kubenswrapper[4978]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Feb 25 06:46:30 crc kubenswrapper[4978]: > logger="UnhandledError" Feb 25 06:46:30 crc kubenswrapper[4978]: W0225 06:46:30.640325 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-85450155b1fb5c9193b4189f86cb827d5c924037371aaa9d5632a9bd34184b3e WatchSource:0}: Error finding container 85450155b1fb5c9193b4189f86cb827d5c924037371aaa9d5632a9bd34184b3e: Status 404 returned error can't find the container with id 85450155b1fb5c9193b4189f86cb827d5c924037371aaa9d5632a9bd34184b3e Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.646211 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 06:46:30 crc kubenswrapper[4978]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Feb 25 06:46:30 crc kubenswrapper[4978]: set -o allexport Feb 25 06:46:30 crc kubenswrapper[4978]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Feb 25 06:46:30 crc kubenswrapper[4978]: source /etc/kubernetes/apiserver-url.env Feb 25 06:46:30 crc kubenswrapper[4978]: else Feb 25 06:46:30 crc kubenswrapper[4978]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Feb 25 06:46:30 crc kubenswrapper[4978]: exit 1 Feb 25 06:46:30 crc kubenswrapper[4978]: fi Feb 25 06:46:30 crc kubenswrapper[4978]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Feb 25 06:46:30 crc kubenswrapper[4978]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Feb 25 06:46:30 crc kubenswrapper[4978]: > logger="UnhandledError" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.647095 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 06:46:30 crc kubenswrapper[4978]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Feb 25 06:46:30 crc kubenswrapper[4978]: if [[ -f "/env/_master" ]]; then Feb 25 06:46:30 crc kubenswrapper[4978]: set -o allexport Feb 25 06:46:30 crc kubenswrapper[4978]: source "/env/_master" Feb 25 06:46:30 crc kubenswrapper[4978]: set +o allexport Feb 25 06:46:30 crc kubenswrapper[4978]: fi Feb 25 06:46:30 crc kubenswrapper[4978]: Feb 25 06:46:30 crc kubenswrapper[4978]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Feb 25 06:46:30 crc kubenswrapper[4978]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Feb 25 06:46:30 crc kubenswrapper[4978]: --disable-webhook \ Feb 25 06:46:30 crc kubenswrapper[4978]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Feb 25 06:46:30 crc kubenswrapper[4978]: --loglevel="${LOGLEVEL}" Feb 25 06:46:30 crc kubenswrapper[4978]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Feb 25 06:46:30 crc kubenswrapper[4978]: > logger="UnhandledError" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.649544 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.649811 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.660181 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.661605 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.673170 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.673221 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.673239 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.673264 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.673282 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:30Z","lastTransitionTime":"2026-02-25T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.776332 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.776435 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.776456 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.776484 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.776503 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:30Z","lastTransitionTime":"2026-02-25T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.879875 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.879947 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.879967 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.879992 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.880009 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:30Z","lastTransitionTime":"2026-02-25T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.982764 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.982919 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.982934 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:46:31.982899147 +0000 UTC m=+85.422155666 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.983001 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.983128 4978 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.983199 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:31.983180866 +0000 UTC m=+85.422437355 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.983240 4978 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: E0225 06:46:30.983431 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:31.983348241 +0000 UTC m=+85.422604750 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.983633 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.983676 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.983699 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.983728 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:30 crc kubenswrapper[4978]: I0225 06:46:30.983750 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:30Z","lastTransitionTime":"2026-02-25T06:46:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.084403 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.084796 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.084567 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.085075 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.085174 4978 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.084944 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.085322 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.085352 4978 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.085485 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:32.085453569 +0000 UTC m=+85.524710058 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.085540 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:32.085521701 +0000 UTC m=+85.524778270 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.086617 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.086759 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.086783 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.086811 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.086835 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:31Z","lastTransitionTime":"2026-02-25T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.189618 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.189700 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.189724 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.189754 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.189777 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:31Z","lastTransitionTime":"2026-02-25T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.293466 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.293532 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.293557 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.293591 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.293613 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:31Z","lastTransitionTime":"2026-02-25T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.319411 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8365a2e30f2268b7c8ea68feca6daee4c8f8301ec05d98b35d0cabef2303d5c4"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.321152 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"85450155b1fb5c9193b4189f86cb827d5c924037371aaa9d5632a9bd34184b3e"} Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.321557 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.322745 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.323236 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3ad4a8c8d981ad950b63a8374ec62a6f28b6cd836a068ce5ab7af694684f1b36"} Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.325396 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 06:46:31 crc kubenswrapper[4978]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Feb 25 06:46:31 crc kubenswrapper[4978]: set -o allexport Feb 25 06:46:31 crc kubenswrapper[4978]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Feb 25 06:46:31 crc kubenswrapper[4978]: source /etc/kubernetes/apiserver-url.env Feb 25 06:46:31 crc kubenswrapper[4978]: else Feb 25 06:46:31 crc kubenswrapper[4978]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Feb 25 06:46:31 crc kubenswrapper[4978]: exit 1 Feb 25 06:46:31 crc kubenswrapper[4978]: fi Feb 25 06:46:31 crc kubenswrapper[4978]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Feb 25 06:46:31 crc kubenswrapper[4978]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Feb 25 06:46:31 crc kubenswrapper[4978]: > logger="UnhandledError" Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.325527 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 06:46:31 crc kubenswrapper[4978]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Feb 25 06:46:31 crc kubenswrapper[4978]: if [[ -f "/env/_master" ]]; then Feb 25 06:46:31 crc kubenswrapper[4978]: set -o allexport Feb 25 06:46:31 crc kubenswrapper[4978]: source "/env/_master" Feb 25 06:46:31 crc kubenswrapper[4978]: set +o allexport Feb 25 06:46:31 crc kubenswrapper[4978]: fi Feb 25 06:46:31 crc kubenswrapper[4978]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Feb 25 06:46:31 crc kubenswrapper[4978]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Feb 25 06:46:31 crc kubenswrapper[4978]: ho_enable="--enable-hybrid-overlay" Feb 25 06:46:31 crc kubenswrapper[4978]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Feb 25 06:46:31 crc kubenswrapper[4978]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Feb 25 06:46:31 crc kubenswrapper[4978]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Feb 25 06:46:31 crc kubenswrapper[4978]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Feb 25 06:46:31 crc kubenswrapper[4978]: --webhook-cert-dir="/etc/webhook-cert" \ Feb 25 06:46:31 crc kubenswrapper[4978]: --webhook-host=127.0.0.1 \ Feb 25 06:46:31 crc kubenswrapper[4978]: --webhook-port=9743 \ Feb 25 06:46:31 crc kubenswrapper[4978]: ${ho_enable} \ Feb 25 06:46:31 crc kubenswrapper[4978]: --enable-interconnect \ Feb 25 06:46:31 crc kubenswrapper[4978]: --disable-approver \ Feb 25 06:46:31 crc kubenswrapper[4978]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Feb 25 06:46:31 crc kubenswrapper[4978]: --wait-for-kubernetes-api=200s \ Feb 25 06:46:31 crc kubenswrapper[4978]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Feb 25 06:46:31 crc kubenswrapper[4978]: --loglevel="${LOGLEVEL}" Feb 25 06:46:31 crc kubenswrapper[4978]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Feb 25 06:46:31 crc kubenswrapper[4978]: > logger="UnhandledError" Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.328672 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.329603 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 06:46:31 crc kubenswrapper[4978]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Feb 25 06:46:31 crc kubenswrapper[4978]: if [[ -f "/env/_master" ]]; then Feb 25 06:46:31 crc kubenswrapper[4978]: set -o allexport Feb 25 06:46:31 crc kubenswrapper[4978]: source "/env/_master" Feb 25 06:46:31 crc kubenswrapper[4978]: set +o allexport Feb 25 06:46:31 crc kubenswrapper[4978]: fi Feb 25 06:46:31 crc kubenswrapper[4978]: Feb 25 06:46:31 crc kubenswrapper[4978]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Feb 25 06:46:31 crc kubenswrapper[4978]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Feb 25 06:46:31 crc kubenswrapper[4978]: --disable-webhook \ Feb 25 06:46:31 crc kubenswrapper[4978]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Feb 25 06:46:31 crc kubenswrapper[4978]: --loglevel="${LOGLEVEL}" Feb 25 06:46:31 crc kubenswrapper[4978]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Feb 25 06:46:31 crc kubenswrapper[4978]: > logger="UnhandledError" Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.330985 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.333929 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.334789 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.335834 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.338073 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.339437 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.341544 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.343196 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.345023 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.348039 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.348298 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.349472 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.351644 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.352286 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.353325 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.354123 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.355821 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.356952 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.358037 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.359484 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.360273 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.363087 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.364766 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.366547 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.367943 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.368687 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.369002 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.369628 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.370185 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.372268 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.373204 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.374440 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.375272 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.376574 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.377274 4978 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.377435 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.380532 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.381314 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.381985 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.385340 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.386982 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.388348 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.389521 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.390818 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.391316 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.392218 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.393728 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.395204 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.396106 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.396137 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.396149 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.396169 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.396181 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:31Z","lastTransitionTime":"2026-02-25T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.396468 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.397723 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.398575 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.399969 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.401188 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.402491 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.403264 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.404098 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.405472 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.406430 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.407755 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.408515 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.424608 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.437075 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.448048 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.466630 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.477854 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.490781 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.498299 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.498491 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.498575 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.498651 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.498707 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:31Z","lastTransitionTime":"2026-02-25T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.504263 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.601819 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.602063 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.602259 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.602338 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.602448 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:31Z","lastTransitionTime":"2026-02-25T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.705745 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.705789 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.705799 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.705818 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.705832 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:31Z","lastTransitionTime":"2026-02-25T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.808935 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.808994 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.809015 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.809045 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.809067 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:31Z","lastTransitionTime":"2026-02-25T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.912927 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.912988 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.913010 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.913040 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.913060 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:31Z","lastTransitionTime":"2026-02-25T06:46:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.993823 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.993988 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:31 crc kubenswrapper[4978]: I0225 06:46:31.994054 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.994191 4978 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.994243 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:46:33.994164626 +0000 UTC m=+87.433421115 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.994328 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:33.99429634 +0000 UTC m=+87.433552849 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.994330 4978 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:31 crc kubenswrapper[4978]: E0225 06:46:31.994466 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:33.994441644 +0000 UTC m=+87.433698173 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.017675 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.017775 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.017800 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.017873 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.017899 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:32Z","lastTransitionTime":"2026-02-25T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.094392 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.094441 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.094588 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.094606 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.094619 4978 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.094619 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.094656 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.094675 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:34.094659685 +0000 UTC m=+87.533916154 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.094675 4978 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.094741 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:34.094721057 +0000 UTC m=+87.533977556 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.121192 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.121259 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.121278 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.121303 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.121322 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:32Z","lastTransitionTime":"2026-02-25T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.224924 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.225042 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.225061 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.225086 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.225104 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:32Z","lastTransitionTime":"2026-02-25T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.326559 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.326633 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.326559 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.326734 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.326839 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:32 crc kubenswrapper[4978]: E0225 06:46:32.326971 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.327972 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.328028 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.328051 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.328079 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.328100 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:32Z","lastTransitionTime":"2026-02-25T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.431103 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.431175 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.431192 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.431219 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.431243 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:32Z","lastTransitionTime":"2026-02-25T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.533861 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.533915 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.533932 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.533954 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.533972 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:32Z","lastTransitionTime":"2026-02-25T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.637324 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.637420 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.637449 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.637480 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.637505 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:32Z","lastTransitionTime":"2026-02-25T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.740914 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.741011 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.741040 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.741108 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.741141 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:32Z","lastTransitionTime":"2026-02-25T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.844514 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.844579 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.844602 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.844633 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.844659 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:32Z","lastTransitionTime":"2026-02-25T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.947397 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.947451 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.947467 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.947490 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:32 crc kubenswrapper[4978]: I0225 06:46:32.947508 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:32Z","lastTransitionTime":"2026-02-25T06:46:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.050606 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.050873 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.050994 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.051089 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.051157 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:33Z","lastTransitionTime":"2026-02-25T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.153987 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.154061 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.154083 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.154117 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.154142 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:33Z","lastTransitionTime":"2026-02-25T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.257263 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.257324 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.257348 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.257406 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.257426 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:33Z","lastTransitionTime":"2026-02-25T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.359782 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.359824 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.359835 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.359854 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.359865 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:33Z","lastTransitionTime":"2026-02-25T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.462628 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.462705 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.462726 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.462753 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.462772 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:33Z","lastTransitionTime":"2026-02-25T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.566757 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.566835 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.566853 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.566882 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.566901 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:33Z","lastTransitionTime":"2026-02-25T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.669923 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.670001 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.670024 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.670050 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.670067 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:33Z","lastTransitionTime":"2026-02-25T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.772736 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.772773 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.772785 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.772802 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.772813 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:33Z","lastTransitionTime":"2026-02-25T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.876082 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.876180 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.876216 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.876249 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.876272 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:33Z","lastTransitionTime":"2026-02-25T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.878548 4978 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.979739 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.979812 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.979835 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.979865 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:33 crc kubenswrapper[4978]: I0225 06:46:33.979883 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:33Z","lastTransitionTime":"2026-02-25T06:46:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.013706 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.013797 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.013845 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.013978 4978 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.014046 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:38.014025121 +0000 UTC m=+91.453281620 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.014252 4978 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.014346 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:46:38.01430689 +0000 UTC m=+91.453563389 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.014418 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:38.014402593 +0000 UTC m=+91.453659092 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.082066 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.082130 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.082155 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.082184 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.082206 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:34Z","lastTransitionTime":"2026-02-25T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.114935 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.115008 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.115216 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.115280 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.115302 4978 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.115311 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.115351 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.115421 4978 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.115442 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:38.115413558 +0000 UTC m=+91.554670057 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.115512 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:38.11548266 +0000 UTC m=+91.554739199 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.185462 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.185552 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.185577 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.185611 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.185634 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:34Z","lastTransitionTime":"2026-02-25T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.288871 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.288930 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.288948 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.288972 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.288991 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:34Z","lastTransitionTime":"2026-02-25T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.327189 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.327279 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.327421 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.327189 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.327548 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:34 crc kubenswrapper[4978]: E0225 06:46:34.327777 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.391811 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.391884 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.391902 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.391929 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.391950 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:34Z","lastTransitionTime":"2026-02-25T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.494794 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.494995 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.495039 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.495071 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.495095 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:34Z","lastTransitionTime":"2026-02-25T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.598860 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.598932 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.598957 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.598993 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.599020 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:34Z","lastTransitionTime":"2026-02-25T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.702496 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.702593 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.702621 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.702655 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.702682 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:34Z","lastTransitionTime":"2026-02-25T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.805565 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.805656 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.805681 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.805712 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.805735 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:34Z","lastTransitionTime":"2026-02-25T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.908740 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.908931 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.908971 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.908998 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:34 crc kubenswrapper[4978]: I0225 06:46:34.909016 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:34Z","lastTransitionTime":"2026-02-25T06:46:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.012504 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.012591 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.012615 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.012651 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.012673 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:35Z","lastTransitionTime":"2026-02-25T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.115916 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.116017 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.116033 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.116060 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.116079 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:35Z","lastTransitionTime":"2026-02-25T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.218592 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.218692 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.218723 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.218753 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.218774 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:35Z","lastTransitionTime":"2026-02-25T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.321959 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.322043 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.322065 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.322093 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.322114 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:35Z","lastTransitionTime":"2026-02-25T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.425202 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.425275 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.425297 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.425324 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.425342 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:35Z","lastTransitionTime":"2026-02-25T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.528359 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.528466 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.528484 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.528507 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.528525 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:35Z","lastTransitionTime":"2026-02-25T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.630962 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.631019 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.631041 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.631067 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.631085 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:35Z","lastTransitionTime":"2026-02-25T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.734831 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.734888 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.734905 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.734930 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.734947 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:35Z","lastTransitionTime":"2026-02-25T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.837709 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.837783 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.837801 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.837828 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.837847 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:35Z","lastTransitionTime":"2026-02-25T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.940765 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.940829 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.940846 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.940870 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:35 crc kubenswrapper[4978]: I0225 06:46:35.940889 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:35Z","lastTransitionTime":"2026-02-25T06:46:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.044290 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.044356 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.044391 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.044416 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.044433 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:36Z","lastTransitionTime":"2026-02-25T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.147213 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.147292 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.147314 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.147352 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.147409 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:36Z","lastTransitionTime":"2026-02-25T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.259603 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.259698 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.259713 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.259749 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.259764 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:36Z","lastTransitionTime":"2026-02-25T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.327282 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.327455 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.327503 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:36 crc kubenswrapper[4978]: E0225 06:46:36.327632 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:36 crc kubenswrapper[4978]: E0225 06:46:36.327784 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:36 crc kubenswrapper[4978]: E0225 06:46:36.327942 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.363146 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.363207 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.363219 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.363240 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.363251 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:36Z","lastTransitionTime":"2026-02-25T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.467474 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.467565 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.467595 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.467633 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.467660 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:36Z","lastTransitionTime":"2026-02-25T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.571505 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.571586 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.571607 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.571640 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.571663 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:36Z","lastTransitionTime":"2026-02-25T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.676005 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.676102 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.676125 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.676200 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.676230 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:36Z","lastTransitionTime":"2026-02-25T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.780254 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.780340 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.780362 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.780445 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.780468 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:36Z","lastTransitionTime":"2026-02-25T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.884917 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.885014 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.885055 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.885091 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.885115 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:36Z","lastTransitionTime":"2026-02-25T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.988828 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.988884 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.988894 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.988917 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:36 crc kubenswrapper[4978]: I0225 06:46:36.988930 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:36Z","lastTransitionTime":"2026-02-25T06:46:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.092205 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.092307 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.092342 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.092523 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.092543 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:37Z","lastTransitionTime":"2026-02-25T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.196434 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.196513 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.196538 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.196568 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.196593 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:37Z","lastTransitionTime":"2026-02-25T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.300471 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.300541 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.300558 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.300586 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.300603 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:37Z","lastTransitionTime":"2026-02-25T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.346638 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.360997 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.374241 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.384885 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.397033 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.403531 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.403763 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.403781 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.403802 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.403849 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:37Z","lastTransitionTime":"2026-02-25T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.409083 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.506597 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.506665 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.506683 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.506709 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.506728 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:37Z","lastTransitionTime":"2026-02-25T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.609781 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.609838 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.609855 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.609882 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.609900 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:37Z","lastTransitionTime":"2026-02-25T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.712546 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.712610 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.712647 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.712686 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.712709 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:37Z","lastTransitionTime":"2026-02-25T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.816111 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.816189 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.816211 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.816246 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.816276 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:37Z","lastTransitionTime":"2026-02-25T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.920294 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.920360 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.920414 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.920450 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:37 crc kubenswrapper[4978]: I0225 06:46:37.920474 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:37Z","lastTransitionTime":"2026-02-25T06:46:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.024598 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.024672 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.024690 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.024715 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.024734 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.054148 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.054336 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:46:46.054304649 +0000 UTC m=+99.493561149 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.054559 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.054637 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.054752 4978 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.054854 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:46.054837456 +0000 UTC m=+99.494093945 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.054870 4978 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.054961 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:46.054934499 +0000 UTC m=+99.494190998 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.128163 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.128225 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.128242 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.128269 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.128286 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.156002 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.156081 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.156241 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.156278 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.156286 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.156297 4978 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.156316 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.156333 4978 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.156427 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:46.156357857 +0000 UTC m=+99.595614356 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.156476 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:46.15645548 +0000 UTC m=+99.595711979 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.231486 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.231589 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.231613 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.231642 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.231670 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.288493 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.288562 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.288586 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.288614 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.288639 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.304038 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.308810 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.308855 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.308873 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.308897 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.308916 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.325447 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.327535 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.327739 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.327544 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.327941 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.327780 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.328109 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.330677 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.330738 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.330758 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.330780 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.330797 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.351657 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.356192 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.356221 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.356231 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.356249 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.356261 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.367118 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.370871 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.370933 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.370957 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.370990 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.371013 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.389472 4978 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"1785e4a8-a914-4b77-ac0a-56ce2c18cd36\\\",\\\"systemUUID\\\":\\\"a1f14dca-a47a-4420-a814-a4877946a844\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:38 crc kubenswrapper[4978]: E0225 06:46:38.389626 4978 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.391410 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.391479 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.391503 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.391535 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.391558 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.495032 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.495116 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.495134 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.495159 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.495178 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.597514 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.597567 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.597585 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.597610 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.597630 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.700481 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.700539 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.700557 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.700580 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.700598 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.803473 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.803536 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.803558 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.803587 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.803607 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.906622 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.906686 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.906705 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.906730 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:38 crc kubenswrapper[4978]: I0225 06:46:38.906747 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:38Z","lastTransitionTime":"2026-02-25T06:46:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.009528 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.009607 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.009630 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.009658 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.009679 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:39Z","lastTransitionTime":"2026-02-25T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.113013 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.113078 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.113095 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.113118 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.113135 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:39Z","lastTransitionTime":"2026-02-25T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.215965 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.216023 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.216044 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.216071 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.216093 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:39Z","lastTransitionTime":"2026-02-25T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.319684 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.319738 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.319756 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.319780 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.319798 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:39Z","lastTransitionTime":"2026-02-25T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.344053 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.422325 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.422441 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.422467 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.422498 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.422519 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:39Z","lastTransitionTime":"2026-02-25T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.524801 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.524919 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.524941 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.524964 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.524981 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:39Z","lastTransitionTime":"2026-02-25T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.627333 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.627402 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.627416 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.627435 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.627447 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:39Z","lastTransitionTime":"2026-02-25T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.730414 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.730486 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.730506 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.730531 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.730551 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:39Z","lastTransitionTime":"2026-02-25T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.834696 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.834770 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.834792 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.834822 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.834844 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:39Z","lastTransitionTime":"2026-02-25T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.938199 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.938301 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.938320 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.938343 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:39 crc kubenswrapper[4978]: I0225 06:46:39.938357 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:39Z","lastTransitionTime":"2026-02-25T06:46:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.041741 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.041873 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.041894 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.041918 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.041947 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:40Z","lastTransitionTime":"2026-02-25T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.144535 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.144592 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.144610 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.144634 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.144653 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:40Z","lastTransitionTime":"2026-02-25T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.205192 4978 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.247786 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.247865 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.247896 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.247925 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.247945 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:40Z","lastTransitionTime":"2026-02-25T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.327554 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.327668 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.327763 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:40 crc kubenswrapper[4978]: E0225 06:46:40.327960 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:40 crc kubenswrapper[4978]: E0225 06:46:40.328088 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:40 crc kubenswrapper[4978]: E0225 06:46:40.328258 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.349717 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.349775 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.349793 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.350005 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.350026 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:40Z","lastTransitionTime":"2026-02-25T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.453120 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.453160 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.453168 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.453185 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.453196 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:40Z","lastTransitionTime":"2026-02-25T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.556895 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.556967 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.556987 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.557009 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.557025 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:40Z","lastTransitionTime":"2026-02-25T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.660774 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.660859 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.660883 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.660916 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.660955 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:40Z","lastTransitionTime":"2026-02-25T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.764746 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.764816 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.764829 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.764847 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.764861 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:40Z","lastTransitionTime":"2026-02-25T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.868724 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.868804 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.868826 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.868860 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.868882 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:40Z","lastTransitionTime":"2026-02-25T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.971806 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.971875 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.971895 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.971936 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:40 crc kubenswrapper[4978]: I0225 06:46:40.971957 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:40Z","lastTransitionTime":"2026-02-25T06:46:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.074447 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.074525 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.074542 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.074567 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.074586 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:41Z","lastTransitionTime":"2026-02-25T06:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.177677 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.177764 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.177784 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.177818 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.177836 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:41Z","lastTransitionTime":"2026-02-25T06:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.281117 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.281185 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.281206 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.281234 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.281254 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:41Z","lastTransitionTime":"2026-02-25T06:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.385471 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.385595 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.385617 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.385657 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.385679 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:41Z","lastTransitionTime":"2026-02-25T06:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.489063 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.489160 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.489187 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.489233 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.489270 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:41Z","lastTransitionTime":"2026-02-25T06:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.593122 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.593238 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.593263 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.593291 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.593310 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:41Z","lastTransitionTime":"2026-02-25T06:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.696260 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.696341 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.696397 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.696434 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.696457 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:41Z","lastTransitionTime":"2026-02-25T06:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.799633 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.799704 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.799717 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.799755 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.799768 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:41Z","lastTransitionTime":"2026-02-25T06:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.902343 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.902422 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.902436 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.902458 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:41 crc kubenswrapper[4978]: I0225 06:46:41.902473 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:41Z","lastTransitionTime":"2026-02-25T06:46:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.005098 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.005149 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.005166 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.005189 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.005206 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:42Z","lastTransitionTime":"2026-02-25T06:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.108814 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.108892 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.108911 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.108943 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.108963 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:42Z","lastTransitionTime":"2026-02-25T06:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.212974 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.213053 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.213072 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.213102 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.213124 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:42Z","lastTransitionTime":"2026-02-25T06:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.316501 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.316582 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.316606 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.316640 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.316668 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:42Z","lastTransitionTime":"2026-02-25T06:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.326929 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.327014 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.326947 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:42 crc kubenswrapper[4978]: E0225 06:46:42.327240 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:42 crc kubenswrapper[4978]: E0225 06:46:42.327461 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:42 crc kubenswrapper[4978]: E0225 06:46:42.327618 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.430086 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.430148 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.430161 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.430185 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.430201 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:42Z","lastTransitionTime":"2026-02-25T06:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.533439 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.533513 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.533537 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.533568 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.533594 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:42Z","lastTransitionTime":"2026-02-25T06:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.637428 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.637527 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.637547 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.637617 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.637637 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:42Z","lastTransitionTime":"2026-02-25T06:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.740663 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.740726 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.740748 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.740772 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.740790 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:42Z","lastTransitionTime":"2026-02-25T06:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.843818 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.843879 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.843896 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.843921 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.843940 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:42Z","lastTransitionTime":"2026-02-25T06:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.947685 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.947754 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.947774 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.947799 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:42 crc kubenswrapper[4978]: I0225 06:46:42.947818 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:42Z","lastTransitionTime":"2026-02-25T06:46:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.050790 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.050885 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.050904 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.050930 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.050948 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:43Z","lastTransitionTime":"2026-02-25T06:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.155034 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.155110 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.155129 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.155162 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.155184 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:43Z","lastTransitionTime":"2026-02-25T06:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.257651 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.257708 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.257724 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.257748 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.257764 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:43Z","lastTransitionTime":"2026-02-25T06:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.346411 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.346814 4978 scope.go:117] "RemoveContainer" containerID="4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5" Feb 25 06:46:43 crc kubenswrapper[4978]: E0225 06:46:43.347948 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.358061 4978 scope.go:117] "RemoveContainer" containerID="4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5" Feb 25 06:46:43 crc kubenswrapper[4978]: E0225 06:46:43.359144 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.360751 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.360813 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.360831 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.360858 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.360885 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:43Z","lastTransitionTime":"2026-02-25T06:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.463345 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.463410 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.463422 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.463439 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.463455 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:43Z","lastTransitionTime":"2026-02-25T06:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.566408 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.566464 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.566481 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.566506 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.566525 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:43Z","lastTransitionTime":"2026-02-25T06:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.672318 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.672446 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.672481 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.672521 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.672559 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:43Z","lastTransitionTime":"2026-02-25T06:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.775841 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.775903 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.775922 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.775950 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.775970 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:43Z","lastTransitionTime":"2026-02-25T06:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.879087 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.879140 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.879152 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.879167 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.879176 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:43Z","lastTransitionTime":"2026-02-25T06:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.982887 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.982951 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.982972 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.983001 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:43 crc kubenswrapper[4978]: I0225 06:46:43.983024 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:43Z","lastTransitionTime":"2026-02-25T06:46:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.085764 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.085871 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.085903 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.085920 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.085934 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:44Z","lastTransitionTime":"2026-02-25T06:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.188905 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.189010 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.189024 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.189051 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.189065 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:44Z","lastTransitionTime":"2026-02-25T06:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.294940 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.295000 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.295018 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.295045 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.295064 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:44Z","lastTransitionTime":"2026-02-25T06:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.326939 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.327017 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:44 crc kubenswrapper[4978]: E0225 06:46:44.327066 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:44 crc kubenswrapper[4978]: E0225 06:46:44.327154 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.327018 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:44 crc kubenswrapper[4978]: E0225 06:46:44.327297 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.364534 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7c3e25e041917aa7bab39311882cc58d65a55d47d73d2b6999de191de3c9b7b0"} Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.377505 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81de5935-01e0-420d-a973-d0f6a3e40222\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T06:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T06:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T06:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T06:45:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T06:45:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://251cffe6262dcd87128cb9d7bf8db5a2fc28d9fcac614ba26a2117421302890a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-25T06:45:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f31926eb93c19d10f15275e100531fc0701b878d8088f88c065d6cd9147b2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f31926eb93c19d10f15275e100531fc0701b878d8088f88c065d6cd9147b2b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-25T06:45:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-25T06:45:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-25T06:45:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.392933 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.397963 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.398012 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.398027 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.398050 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.398066 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:44Z","lastTransitionTime":"2026-02-25T06:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.408421 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.421550 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:44Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c3e25e041917aa7bab39311882cc58d65a55d47d73d2b6999de191de3c9b7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-25T06:46:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.435230 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33715162-a697-4e39-ba6d-468d3c5d1d74\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T06:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T06:45:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T06:45:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T06:45:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T06:45:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-25T06:45:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-25T06:45:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-25T06:45:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-25T06:46:10Z\\\",\\\"message\\\":\\\"le observer\\\\nW0225 06:46:09.858875 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0225 06:46:09.859033 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0225 06:46:09.859972 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1075389772/tls.crt::/tmp/serving-cert-1075389772/tls.key\\\\\\\"\\\\nI0225 06:46:10.217857 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0225 06:46:10.220397 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0225 06:46:10.220447 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0225 06:46:10.220492 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0225 06:46:10.220517 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0225 06:46:10.228454 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0225 06:46:10.228507 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0225 06:46:10.228522 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0225 06:46:10.228552 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0225 06:46:10.228561 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0225 06:46:10.228571 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0225 06:46:10.228580 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0225 06:46:10.228463 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0225 06:46:10.230573 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-25T06:46:09Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-25T06:45:10Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-25T06:45:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-25T06:45:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-25T06:45:07Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.449545 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.464334 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.495574 4978 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-25T06:46:30Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.500205 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.500237 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.500247 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.500262 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.500271 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:44Z","lastTransitionTime":"2026-02-25T06:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.602757 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.602787 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.602796 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.602811 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.602821 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:44Z","lastTransitionTime":"2026-02-25T06:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.705144 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.705223 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.705238 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.705255 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.705269 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:44Z","lastTransitionTime":"2026-02-25T06:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.808031 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.808121 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.808148 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.808184 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.808208 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:44Z","lastTransitionTime":"2026-02-25T06:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.884774 4978 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.911158 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.911230 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.911253 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.911284 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:44 crc kubenswrapper[4978]: I0225 06:46:44.911301 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:44Z","lastTransitionTime":"2026-02-25T06:46:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.015767 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.015841 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.015861 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.015890 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.015910 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:45Z","lastTransitionTime":"2026-02-25T06:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.118950 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.119036 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.119061 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.119097 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.119120 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:45Z","lastTransitionTime":"2026-02-25T06:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.222757 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.222837 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.222858 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.222884 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.222957 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:45Z","lastTransitionTime":"2026-02-25T06:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.326447 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.326501 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.326523 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.326547 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.326569 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:45Z","lastTransitionTime":"2026-02-25T06:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.429380 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.429481 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.429501 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.429530 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.429551 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:45Z","lastTransitionTime":"2026-02-25T06:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.533982 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.534018 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.534029 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.534049 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.534062 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:45Z","lastTransitionTime":"2026-02-25T06:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.637154 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.637253 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.637283 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.637321 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.637348 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:45Z","lastTransitionTime":"2026-02-25T06:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.740554 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.740624 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.740643 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.740670 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.740689 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:45Z","lastTransitionTime":"2026-02-25T06:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.843818 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.843891 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.843912 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.843937 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.843954 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:45Z","lastTransitionTime":"2026-02-25T06:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.947136 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.947220 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.947238 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.947268 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:45 crc kubenswrapper[4978]: I0225 06:46:45.947284 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:45Z","lastTransitionTime":"2026-02-25T06:46:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.050680 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.050764 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.050916 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.051454 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.051551 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:46Z","lastTransitionTime":"2026-02-25T06:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.082300 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.082503 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.082527 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.082501718 +0000 UTC m=+115.521758177 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.082610 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.082664 4978 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.082709 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.082698414 +0000 UTC m=+115.521954873 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.082786 4978 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.082901 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.082871329 +0000 UTC m=+115.522127828 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.154467 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.154525 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.154546 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.154574 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.154594 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:46Z","lastTransitionTime":"2026-02-25T06:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.183108 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.183187 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.183408 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.183461 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.183464 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.183485 4978 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.183505 4978 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.183524 4978 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.183577 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.183552874 +0000 UTC m=+115.622809373 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.183630 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.183618586 +0000 UTC m=+115.622875075 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.190250 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-tck4b"] Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.190725 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tck4b" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.192740 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.194336 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.197051 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.213962 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-j496h"] Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.214517 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.217006 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-lrt6k"] Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.217026 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.217064 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.217076 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.217511 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.217713 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-rjvfr"] Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.218768 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.218842 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.219796 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.221740 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.221918 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.223041 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.223057 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.224105 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.224572 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.224712 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.251714 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2gb8f"] Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.253461 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.256090 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.256665 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.257216 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.257274 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.257742 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.258863 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.258896 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.261404 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.261449 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.261468 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.261488 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.261502 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:46Z","lastTransitionTime":"2026-02-25T06:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.281879 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=7.281835557 podStartE2EDuration="7.281835557s" podCreationTimestamp="2026-02-25 06:46:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:46:46.280197068 +0000 UTC m=+99.719453537" watchObservedRunningTime="2026-02-25 06:46:46.281835557 +0000 UTC m=+99.721092026" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.286099 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a5f01015-5dea-4e59-a9fc-326c84b85aed-mcd-auth-proxy-config\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.286344 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85zs5\" (UniqueName: \"kubernetes.io/projected/a5f01015-5dea-4e59-a9fc-326c84b85aed-kube-api-access-85zs5\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.286598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-system-cni-dir\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.286767 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-os-release\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.286968 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a5f01015-5dea-4e59-a9fc-326c84b85aed-rootfs\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.287147 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a5f01015-5dea-4e59-a9fc-326c84b85aed-proxy-tls\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.287295 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-cnibin\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.326598 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.326995 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.327500 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.327677 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.328866 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.329048 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.364229 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.364270 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.364283 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.364304 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.364317 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:46Z","lastTransitionTime":"2026-02-25T06:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.373174 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a13693c580c6920f1c356127165b4b51c686be05682620f559ab59661aa2a996"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.373216 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6724966944e17620979100d747a9649818b173c930aec4659edb515123cf639a"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.388400 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb827\" (UniqueName: \"kubernetes.io/projected/594f940e-2757-4577-94b4-301e27ad4fab-kube-api-access-vb827\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.388747 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-openvswitch\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.388865 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.388966 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-script-lib\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.389131 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-var-lib-cni-bin\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.389249 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.389367 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-slash\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.389494 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-bin\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.389636 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b3724722-3928-4d0b-8ef3-f81f91753a30-cni-binary-copy\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.389757 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-kubelet\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.389871 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-netns\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.389999 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-conf-dir\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.390098 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-daemon-config\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.390439 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-etc-kubernetes\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.390572 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-var-lib-openvswitch\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.390745 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-netd\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.390856 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-config\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.390965 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-os-release\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391095 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-system-cni-dir\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391202 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-etc-openvswitch\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391270 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-system-cni-dir\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391311 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx4vx\" (UniqueName: \"kubernetes.io/projected/b3724722-3928-4d0b-8ef3-f81f91753a30-kube-api-access-xx4vx\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391444 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a5f01015-5dea-4e59-a9fc-326c84b85aed-rootfs\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391484 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-cnibin\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391527 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-systemd-units\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391573 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-log-socket\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391612 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-env-overrides\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391650 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-var-lib-cni-multus\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391686 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovn-node-metrics-cert\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391718 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdcj7\" (UniqueName: \"kubernetes.io/projected/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-kube-api-access-zdcj7\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391828 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-run-multus-certs\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391875 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a5f01015-5dea-4e59-a9fc-326c84b85aed-rootfs\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391916 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-cnibin\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391920 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/594f940e-2757-4577-94b4-301e27ad4fab-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.391985 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bzz2\" (UniqueName: \"kubernetes.io/projected/209c158f-67a9-4f3b-8c4a-80f1f41c9fe1-kube-api-access-9bzz2\") pod \"node-resolver-tck4b\" (UID: \"209c158f-67a9-4f3b-8c4a-80f1f41c9fe1\") " pod="openshift-dns/node-resolver-tck4b" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392029 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-system-cni-dir\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392119 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-var-lib-kubelet\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392156 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-hostroot\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392229 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-cni-dir\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392301 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a5f01015-5dea-4e59-a9fc-326c84b85aed-mcd-auth-proxy-config\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392408 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85zs5\" (UniqueName: \"kubernetes.io/projected/a5f01015-5dea-4e59-a9fc-326c84b85aed-kube-api-access-85zs5\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392452 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-os-release\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392491 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-node-log\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392524 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/209c158f-67a9-4f3b-8c4a-80f1f41c9fe1-hosts-file\") pod \"node-resolver-tck4b\" (UID: \"209c158f-67a9-4f3b-8c4a-80f1f41c9fe1\") " pod="openshift-dns/node-resolver-tck4b" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-run-k8s-cni-cncf-io\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392634 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a5f01015-5dea-4e59-a9fc-326c84b85aed-proxy-tls\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392668 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/594f940e-2757-4577-94b4-301e27ad4fab-cni-binary-copy\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392699 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-systemd\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392730 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-ovn\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392763 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-ovn-kubernetes\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392795 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-cnibin\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392867 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-socket-dir-parent\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.392899 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-run-netns\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.393655 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-os-release\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.394109 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a5f01015-5dea-4e59-a9fc-326c84b85aed-mcd-auth-proxy-config\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.409850 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a5f01015-5dea-4e59-a9fc-326c84b85aed-proxy-tls\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.420428 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85zs5\" (UniqueName: \"kubernetes.io/projected/a5f01015-5dea-4e59-a9fc-326c84b85aed-kube-api-access-85zs5\") pod \"machine-config-daemon-j496h\" (UID: \"a5f01015-5dea-4e59-a9fc-326c84b85aed\") " pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.460945 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zgr79"] Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.461575 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.464107 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.464493 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.465139 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.465816 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.467604 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.467660 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.467680 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.467709 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.467730 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:46Z","lastTransitionTime":"2026-02-25T06:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493297 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493344 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-slash\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493366 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-bin\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493398 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b3724722-3928-4d0b-8ef3-f81f91753a30-cni-binary-copy\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493419 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-kubelet\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493434 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-netns\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493451 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-conf-dir\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493465 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-daemon-config\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493479 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-etc-kubernetes\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493495 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-var-lib-openvswitch\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493509 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-netd\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493523 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-config\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493540 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-os-release\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493556 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx4vx\" (UniqueName: \"kubernetes.io/projected/b3724722-3928-4d0b-8ef3-f81f91753a30-kube-api-access-xx4vx\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493573 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4c2c748e-5801-4563-94f7-1bd7baab6572-serviceca\") pod \"node-ca-zgr79\" (UID: \"4c2c748e-5801-4563-94f7-1bd7baab6572\") " pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493602 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-etc-openvswitch\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493626 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-systemd-units\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493646 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-log-socket\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493642 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-netns\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493704 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-slash\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493664 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-env-overrides\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493762 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-netd\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493825 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-var-lib-cni-multus\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493787 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-var-lib-cni-multus\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493662 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-bin\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493891 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovn-node-metrics-cert\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493938 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdcj7\" (UniqueName: \"kubernetes.io/projected/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-kube-api-access-zdcj7\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493962 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-run-multus-certs\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493938 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-etc-openvswitch\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.493991 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/594f940e-2757-4577-94b4-301e27ad4fab-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494054 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-557fl\" (UniqueName: \"kubernetes.io/projected/4c2c748e-5801-4563-94f7-1bd7baab6572-kube-api-access-557fl\") pod \"node-ca-zgr79\" (UID: \"4c2c748e-5801-4563-94f7-1bd7baab6572\") " pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494098 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bzz2\" (UniqueName: \"kubernetes.io/projected/209c158f-67a9-4f3b-8c4a-80f1f41c9fe1-kube-api-access-9bzz2\") pod \"node-resolver-tck4b\" (UID: \"209c158f-67a9-4f3b-8c4a-80f1f41c9fe1\") " pod="openshift-dns/node-resolver-tck4b" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494132 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-system-cni-dir\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494152 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-os-release\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494201 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-var-lib-kubelet\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494237 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-hostroot\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494250 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-kubelet\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494270 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-cni-dir\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494308 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-node-log\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494339 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/209c158f-67a9-4f3b-8c4a-80f1f41c9fe1-hosts-file\") pod \"node-resolver-tck4b\" (UID: \"209c158f-67a9-4f3b-8c4a-80f1f41c9fe1\") " pod="openshift-dns/node-resolver-tck4b" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494406 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-env-overrides\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494442 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-run-k8s-cni-cncf-io\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494405 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-run-k8s-cni-cncf-io\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494453 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-systemd-units\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494474 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-ovn-kubernetes\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494492 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-cnibin\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494501 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-log-socket\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494513 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-socket-dir-parent\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494532 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-run-netns\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494549 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/594f940e-2757-4577-94b4-301e27ad4fab-cni-binary-copy\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494563 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-systemd\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494581 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-ovn\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494599 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494623 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-script-lib\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494650 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb827\" (UniqueName: \"kubernetes.io/projected/594f940e-2757-4577-94b4-301e27ad4fab-kube-api-access-vb827\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494669 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-openvswitch\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494690 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-var-lib-cni-bin\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494707 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c2c748e-5801-4563-94f7-1bd7baab6572-host\") pod \"node-ca-zgr79\" (UID: \"4c2c748e-5801-4563-94f7-1bd7baab6572\") " pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494700 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-ovn-kubernetes\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494772 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-cnibin\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494921 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-config\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494948 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-system-cni-dir\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.494994 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-socket-dir-parent\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495017 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-var-lib-kubelet\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495032 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b3724722-3928-4d0b-8ef3-f81f91753a30-cni-binary-copy\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495060 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-run-netns\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495072 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-hostroot\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495107 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/594f940e-2757-4577-94b4-301e27ad4fab-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495122 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-ovn\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495179 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-var-lib-openvswitch\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495181 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-conf-dir\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495261 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-systemd\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495268 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/594f940e-2757-4577-94b4-301e27ad4fab-cni-binary-copy\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495123 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495324 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-openvswitch\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495336 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/209c158f-67a9-4f3b-8c4a-80f1f41c9fe1-hosts-file\") pod \"node-resolver-tck4b\" (UID: \"209c158f-67a9-4f3b-8c4a-80f1f41c9fe1\") " pod="openshift-dns/node-resolver-tck4b" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495372 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-var-lib-cni-bin\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495325 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-node-log\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495422 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-etc-kubernetes\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495274 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-cni-dir\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495488 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b3724722-3928-4d0b-8ef3-f81f91753a30-host-run-multus-certs\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.495645 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/594f940e-2757-4577-94b4-301e27ad4fab-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.496218 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-script-lib\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.496280 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b3724722-3928-4d0b-8ef3-f81f91753a30-multus-daemon-config\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.499007 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovn-node-metrics-cert\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.522127 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdcj7\" (UniqueName: \"kubernetes.io/projected/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-kube-api-access-zdcj7\") pod \"ovnkube-node-2gb8f\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.535149 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bzz2\" (UniqueName: \"kubernetes.io/projected/209c158f-67a9-4f3b-8c4a-80f1f41c9fe1-kube-api-access-9bzz2\") pod \"node-resolver-tck4b\" (UID: \"209c158f-67a9-4f3b-8c4a-80f1f41c9fe1\") " pod="openshift-dns/node-resolver-tck4b" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.536288 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb827\" (UniqueName: \"kubernetes.io/projected/594f940e-2757-4577-94b4-301e27ad4fab-kube-api-access-vb827\") pod \"multus-additional-cni-plugins-rjvfr\" (UID: \"594f940e-2757-4577-94b4-301e27ad4fab\") " pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.536455 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx4vx\" (UniqueName: \"kubernetes.io/projected/b3724722-3928-4d0b-8ef3-f81f91753a30-kube-api-access-xx4vx\") pod \"multus-lrt6k\" (UID: \"b3724722-3928-4d0b-8ef3-f81f91753a30\") " pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.539457 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.552543 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lrt6k" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.561541 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.569335 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.573692 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.573757 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.573771 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.573791 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.573804 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:46Z","lastTransitionTime":"2026-02-25T06:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.595069 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-557fl\" (UniqueName: \"kubernetes.io/projected/4c2c748e-5801-4563-94f7-1bd7baab6572-kube-api-access-557fl\") pod \"node-ca-zgr79\" (UID: \"4c2c748e-5801-4563-94f7-1bd7baab6572\") " pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.595422 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c2c748e-5801-4563-94f7-1bd7baab6572-host\") pod \"node-ca-zgr79\" (UID: \"4c2c748e-5801-4563-94f7-1bd7baab6572\") " pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.595446 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4c2c748e-5801-4563-94f7-1bd7baab6572-serviceca\") pod \"node-ca-zgr79\" (UID: \"4c2c748e-5801-4563-94f7-1bd7baab6572\") " pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: W0225 06:46:46.595518 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9a7a2cb_b9af_48db_aa1b_18beab1fe0b5.slice/crio-6ccf95dfe30f0cf5ba6d44f75b9b4ee67f798431bc25d006a16c8aca57789666 WatchSource:0}: Error finding container 6ccf95dfe30f0cf5ba6d44f75b9b4ee67f798431bc25d006a16c8aca57789666: Status 404 returned error can't find the container with id 6ccf95dfe30f0cf5ba6d44f75b9b4ee67f798431bc25d006a16c8aca57789666 Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.595602 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c2c748e-5801-4563-94f7-1bd7baab6572-host\") pod \"node-ca-zgr79\" (UID: \"4c2c748e-5801-4563-94f7-1bd7baab6572\") " pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.596330 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4c2c748e-5801-4563-94f7-1bd7baab6572-serviceca\") pod \"node-ca-zgr79\" (UID: \"4c2c748e-5801-4563-94f7-1bd7baab6572\") " pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.615194 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-557fl\" (UniqueName: \"kubernetes.io/projected/4c2c748e-5801-4563-94f7-1bd7baab6572-kube-api-access-557fl\") pod \"node-ca-zgr79\" (UID: \"4c2c748e-5801-4563-94f7-1bd7baab6572\") " pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.677290 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.677333 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.677344 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.677367 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.677398 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:46Z","lastTransitionTime":"2026-02-25T06:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.721423 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv"] Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.722352 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.725484 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.725642 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.746409 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-6gpsj"] Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.746828 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.746891 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6gpsj" podUID="6b8950bf-8499-44e3-88e1-6ec33500c0c3" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.775260 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zgr79" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.784087 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.784114 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.784123 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.784140 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.784150 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:46Z","lastTransitionTime":"2026-02-25T06:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.796616 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.796657 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6grs\" (UniqueName: \"kubernetes.io/projected/6b8950bf-8499-44e3-88e1-6ec33500c0c3-kube-api-access-q6grs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.796683 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a645cb5-204b-4d9a-8f91-bd34d81708d6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.796711 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a645cb5-204b-4d9a-8f91-bd34d81708d6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.796727 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a645cb5-204b-4d9a-8f91-bd34d81708d6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.796745 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2cnr\" (UniqueName: \"kubernetes.io/projected/5a645cb5-204b-4d9a-8f91-bd34d81708d6-kube-api-access-k2cnr\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.808813 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tck4b" Feb 25 06:46:46 crc kubenswrapper[4978]: W0225 06:46:46.820966 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod209c158f_67a9_4f3b_8c4a_80f1f41c9fe1.slice/crio-db2eae421959b315be2ba89b03c4dd971293e4bde2461368fde85290ac2e51a7 WatchSource:0}: Error finding container db2eae421959b315be2ba89b03c4dd971293e4bde2461368fde85290ac2e51a7: Status 404 returned error can't find the container with id db2eae421959b315be2ba89b03c4dd971293e4bde2461368fde85290ac2e51a7 Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.886556 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.886609 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.886621 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.886645 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.886692 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:46Z","lastTransitionTime":"2026-02-25T06:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.897333 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.897411 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6grs\" (UniqueName: \"kubernetes.io/projected/6b8950bf-8499-44e3-88e1-6ec33500c0c3-kube-api-access-q6grs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.897438 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a645cb5-204b-4d9a-8f91-bd34d81708d6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.897459 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a645cb5-204b-4d9a-8f91-bd34d81708d6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.897478 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a645cb5-204b-4d9a-8f91-bd34d81708d6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.897509 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2cnr\" (UniqueName: \"kubernetes.io/projected/5a645cb5-204b-4d9a-8f91-bd34d81708d6-kube-api-access-k2cnr\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.897595 4978 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 06:46:46 crc kubenswrapper[4978]: E0225 06:46:46.897687 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs podName:6b8950bf-8499-44e3-88e1-6ec33500c0c3 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:47.397661615 +0000 UTC m=+100.836918094 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs") pod "network-metrics-daemon-6gpsj" (UID: "6b8950bf-8499-44e3-88e1-6ec33500c0c3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.898520 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5a645cb5-204b-4d9a-8f91-bd34d81708d6-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.899521 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5a645cb5-204b-4d9a-8f91-bd34d81708d6-env-overrides\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.911213 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5a645cb5-204b-4d9a-8f91-bd34d81708d6-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.915511 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2cnr\" (UniqueName: \"kubernetes.io/projected/5a645cb5-204b-4d9a-8f91-bd34d81708d6-kube-api-access-k2cnr\") pod \"ovnkube-control-plane-749d76644c-t59sv\" (UID: \"5a645cb5-204b-4d9a-8f91-bd34d81708d6\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.922028 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6grs\" (UniqueName: \"kubernetes.io/projected/6b8950bf-8499-44e3-88e1-6ec33500c0c3-kube-api-access-q6grs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.989866 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.989905 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.989922 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.989944 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:46 crc kubenswrapper[4978]: I0225 06:46:46.989959 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:46Z","lastTransitionTime":"2026-02-25T06:46:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.077168 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.093626 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.093674 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.093683 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.093713 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.093724 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:47Z","lastTransitionTime":"2026-02-25T06:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:47 crc kubenswrapper[4978]: W0225 06:46:47.142568 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a645cb5_204b_4d9a_8f91_bd34d81708d6.slice/crio-f196f9271fcbb9335d787763438e9ca297ab2b0a332b8679be4c041e81093489 WatchSource:0}: Error finding container f196f9271fcbb9335d787763438e9ca297ab2b0a332b8679be4c041e81093489: Status 404 returned error can't find the container with id f196f9271fcbb9335d787763438e9ca297ab2b0a332b8679be4c041e81093489 Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.196186 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.196216 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.196226 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.196243 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.196255 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:47Z","lastTransitionTime":"2026-02-25T06:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.302573 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.302607 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.302624 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.302643 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.302655 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:47Z","lastTransitionTime":"2026-02-25T06:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.378763 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zgr79" event={"ID":"4c2c748e-5801-4563-94f7-1bd7baab6572","Type":"ContainerStarted","Data":"ec9be1b7cfc1f1bb1d7f43bfd5ea0014cf249c36f8488f27d4b42442f6bca1f0"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.378813 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zgr79" event={"ID":"4c2c748e-5801-4563-94f7-1bd7baab6572","Type":"ContainerStarted","Data":"9dd47b553eac32bf0f680162f1d3d75e0c7f8425a4ef362f548e6bfef86daea9"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.500912 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:47 crc kubenswrapper[4978]: E0225 06:46:47.502252 4978 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 06:46:47 crc kubenswrapper[4978]: E0225 06:46:47.502321 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs podName:6b8950bf-8499-44e3-88e1-6ec33500c0c3 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:48.502301933 +0000 UTC m=+101.941558412 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs") pod "network-metrics-daemon-6gpsj" (UID: "6b8950bf-8499-44e3-88e1-6ec33500c0c3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.510069 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.510098 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.510107 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.510125 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.510134 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:47Z","lastTransitionTime":"2026-02-25T06:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.512054 4978 generic.go:334] "Generic (PLEG): container finished" podID="594f940e-2757-4577-94b4-301e27ad4fab" containerID="aa4c27e5cba84f56ead51bd76a0a1d51dc13240f7c4d41c245587dbda91f5010" exitCode=0 Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.512123 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" event={"ID":"594f940e-2757-4577-94b4-301e27ad4fab","Type":"ContainerDied","Data":"aa4c27e5cba84f56ead51bd76a0a1d51dc13240f7c4d41c245587dbda91f5010"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.512145 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" event={"ID":"594f940e-2757-4577-94b4-301e27ad4fab","Type":"ContainerStarted","Data":"bedf85b2956ca0f55afc949762c969ed6303ec8616a8c03baaa5eb3a271647f1"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.517053 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"3e071fb8a341e0189d65f3458e9858375ba51d00c8948e55de1fc8e6af556d5b"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.517097 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"1681565645f2fc8aba05522208118ff62097a02395870d10eb9361b066bdb939"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.517108 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"ad28cb941c7dac324a198b853457f301ed01bf976d72e7734a253513744943b4"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.518666 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zgr79" podStartSLOduration=49.518650849 podStartE2EDuration="49.518650849s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:46:47.518292279 +0000 UTC m=+100.957548748" watchObservedRunningTime="2026-02-25 06:46:47.518650849 +0000 UTC m=+100.957907308" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.520229 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" event={"ID":"5a645cb5-204b-4d9a-8f91-bd34d81708d6","Type":"ContainerStarted","Data":"e81a72dee91c3382c47b26c9a246e2b67d9c85965d1d1e92ba299eebaf155c3f"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.520279 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" event={"ID":"5a645cb5-204b-4d9a-8f91-bd34d81708d6","Type":"ContainerStarted","Data":"f196f9271fcbb9335d787763438e9ca297ab2b0a332b8679be4c041e81093489"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.525617 4978 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerID="e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574" exitCode=0 Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.525686 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerDied","Data":"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.525733 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerStarted","Data":"6ccf95dfe30f0cf5ba6d44f75b9b4ee67f798431bc25d006a16c8aca57789666"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.527019 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lrt6k" event={"ID":"b3724722-3928-4d0b-8ef3-f81f91753a30","Type":"ContainerStarted","Data":"091bbfdc48a42fe53c5a7cb75c83171476fe2a6a650e2eaa350688796dac097b"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.527059 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lrt6k" event={"ID":"b3724722-3928-4d0b-8ef3-f81f91753a30","Type":"ContainerStarted","Data":"9e1fbbbc46f1992d219588c5840c6846ae0af04bc253ac83eae2d5335e1f88f3"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.537415 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podStartSLOduration=49.537369928 podStartE2EDuration="49.537369928s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:46:47.533212801 +0000 UTC m=+100.972469280" watchObservedRunningTime="2026-02-25 06:46:47.537369928 +0000 UTC m=+100.976626407" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.544410 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tck4b" event={"ID":"209c158f-67a9-4f3b-8c4a-80f1f41c9fe1","Type":"ContainerStarted","Data":"007acc97e0c9540e87510dc4cb98140efc7ce649978e7fd6678616b35f07b375"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.544445 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tck4b" event={"ID":"209c158f-67a9-4f3b-8c4a-80f1f41c9fe1","Type":"ContainerStarted","Data":"db2eae421959b315be2ba89b03c4dd971293e4bde2461368fde85290ac2e51a7"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.572807 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-tck4b" podStartSLOduration=49.572774372 podStartE2EDuration="49.572774372s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:46:47.571725481 +0000 UTC m=+101.010981940" watchObservedRunningTime="2026-02-25 06:46:47.572774372 +0000 UTC m=+101.012030831" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.610993 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-lrt6k" podStartSLOduration=49.610975162 podStartE2EDuration="49.610975162s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:46:47.588026495 +0000 UTC m=+101.027282954" watchObservedRunningTime="2026-02-25 06:46:47.610975162 +0000 UTC m=+101.050231621" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.613608 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.613639 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.613648 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.613663 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.613673 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:47Z","lastTransitionTime":"2026-02-25T06:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.716979 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.717312 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.717324 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.717345 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.717360 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:47Z","lastTransitionTime":"2026-02-25T06:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.821642 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.821704 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.821723 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.821749 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.821768 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:47Z","lastTransitionTime":"2026-02-25T06:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.924557 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.924608 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.924621 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.924638 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:47 crc kubenswrapper[4978]: I0225 06:46:47.924648 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:47Z","lastTransitionTime":"2026-02-25T06:46:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.028435 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.028484 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.028500 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.028519 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.028533 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:48Z","lastTransitionTime":"2026-02-25T06:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.131674 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.131711 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.131721 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.131737 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.131749 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:48Z","lastTransitionTime":"2026-02-25T06:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.234264 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.234701 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.234720 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.234748 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.234767 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:48Z","lastTransitionTime":"2026-02-25T06:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.327337 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:48 crc kubenswrapper[4978]: E0225 06:46:48.327503 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.327738 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:48 crc kubenswrapper[4978]: E0225 06:46:48.327817 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.327859 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:48 crc kubenswrapper[4978]: E0225 06:46:48.327900 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.327950 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:48 crc kubenswrapper[4978]: E0225 06:46:48.328013 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6gpsj" podUID="6b8950bf-8499-44e3-88e1-6ec33500c0c3" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.343844 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.343891 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.343905 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.343923 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.343940 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:48Z","lastTransitionTime":"2026-02-25T06:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.446413 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.446550 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.446625 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.446706 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.446790 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:48Z","lastTransitionTime":"2026-02-25T06:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.511685 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:48 crc kubenswrapper[4978]: E0225 06:46:48.511808 4978 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 06:46:48 crc kubenswrapper[4978]: E0225 06:46:48.511852 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs podName:6b8950bf-8499-44e3-88e1-6ec33500c0c3 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:50.51183657 +0000 UTC m=+103.951093029 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs") pod "network-metrics-daemon-6gpsj" (UID: "6b8950bf-8499-44e3-88e1-6ec33500c0c3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.549234 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.549562 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.549751 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.549859 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.550697 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:48Z","lastTransitionTime":"2026-02-25T06:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.550845 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c460005ddf2a8665fa2ab67e5caaa381649bd5d21de70efb4ed6bcb58a007950"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.556737 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" event={"ID":"594f940e-2757-4577-94b4-301e27ad4fab","Type":"ContainerStarted","Data":"93aecc2de10a2cc35aa26c22ed23f08f6827843eacaf19dc44e736055019a54e"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.557525 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" event={"ID":"5a645cb5-204b-4d9a-8f91-bd34d81708d6","Type":"ContainerStarted","Data":"5f02d7ccf7b1dfd5d110588c63bbc84ede4269f94400bcf3f15ff5b2a45d99c4"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.562918 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerStarted","Data":"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.562952 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerStarted","Data":"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.562964 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerStarted","Data":"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.562975 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerStarted","Data":"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.607330 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-t59sv" podStartSLOduration=49.607304347 podStartE2EDuration="49.607304347s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:46:48.606683587 +0000 UTC m=+102.045940036" watchObservedRunningTime="2026-02-25 06:46:48.607304347 +0000 UTC m=+102.046560826" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.654590 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.654629 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.654638 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.654653 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.654662 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:48Z","lastTransitionTime":"2026-02-25T06:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.730957 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.731386 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.731400 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.731418 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.731429 4978 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-25T06:46:48Z","lastTransitionTime":"2026-02-25T06:46:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.781644 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z"] Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.781983 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.783735 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.783955 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.784105 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.784124 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.814619 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bacf2e1-fc7a-4951-94d7-3225f455afa4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.814872 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5bacf2e1-fc7a-4951-94d7-3225f455afa4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.814996 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bacf2e1-fc7a-4951-94d7-3225f455afa4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.815127 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5bacf2e1-fc7a-4951-94d7-3225f455afa4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.815233 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bacf2e1-fc7a-4951-94d7-3225f455afa4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.915738 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5bacf2e1-fc7a-4951-94d7-3225f455afa4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.915801 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bacf2e1-fc7a-4951-94d7-3225f455afa4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.915851 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5bacf2e1-fc7a-4951-94d7-3225f455afa4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.915905 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bacf2e1-fc7a-4951-94d7-3225f455afa4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.915971 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bacf2e1-fc7a-4951-94d7-3225f455afa4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.916416 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5bacf2e1-fc7a-4951-94d7-3225f455afa4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.917287 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bacf2e1-fc7a-4951-94d7-3225f455afa4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.917344 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5bacf2e1-fc7a-4951-94d7-3225f455afa4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.925779 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bacf2e1-fc7a-4951-94d7-3225f455afa4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:48 crc kubenswrapper[4978]: I0225 06:46:48.931157 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bacf2e1-fc7a-4951-94d7-3225f455afa4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8bs8z\" (UID: \"5bacf2e1-fc7a-4951-94d7-3225f455afa4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:49 crc kubenswrapper[4978]: I0225 06:46:49.095507 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" Feb 25 06:46:49 crc kubenswrapper[4978]: I0225 06:46:49.310728 4978 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 25 06:46:49 crc kubenswrapper[4978]: I0225 06:46:49.319763 4978 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 25 06:46:49 crc kubenswrapper[4978]: I0225 06:46:49.577286 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerStarted","Data":"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf"} Feb 25 06:46:49 crc kubenswrapper[4978]: I0225 06:46:49.577334 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerStarted","Data":"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3"} Feb 25 06:46:49 crc kubenswrapper[4978]: I0225 06:46:49.579034 4978 generic.go:334] "Generic (PLEG): container finished" podID="594f940e-2757-4577-94b4-301e27ad4fab" containerID="93aecc2de10a2cc35aa26c22ed23f08f6827843eacaf19dc44e736055019a54e" exitCode=0 Feb 25 06:46:49 crc kubenswrapper[4978]: I0225 06:46:49.579084 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" event={"ID":"594f940e-2757-4577-94b4-301e27ad4fab","Type":"ContainerDied","Data":"93aecc2de10a2cc35aa26c22ed23f08f6827843eacaf19dc44e736055019a54e"} Feb 25 06:46:49 crc kubenswrapper[4978]: I0225 06:46:49.581276 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" event={"ID":"5bacf2e1-fc7a-4951-94d7-3225f455afa4","Type":"ContainerStarted","Data":"b8e023e5996c1fe9f82d0ca1e9055c87cf4f46ece673b58162a6d0bb853ea9de"} Feb 25 06:46:49 crc kubenswrapper[4978]: I0225 06:46:49.581356 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" event={"ID":"5bacf2e1-fc7a-4951-94d7-3225f455afa4","Type":"ContainerStarted","Data":"b5fe6f33db328df3865cad649e62b4feac6095de5adc6680815692aee5db6503"} Feb 25 06:46:49 crc kubenswrapper[4978]: I0225 06:46:49.630099 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8bs8z" podStartSLOduration=51.630079584 podStartE2EDuration="51.630079584s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:46:49.629925339 +0000 UTC m=+103.069181838" watchObservedRunningTime="2026-02-25 06:46:49.630079584 +0000 UTC m=+103.069336043" Feb 25 06:46:50 crc kubenswrapper[4978]: I0225 06:46:50.327348 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:50 crc kubenswrapper[4978]: I0225 06:46:50.327397 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:50 crc kubenswrapper[4978]: E0225 06:46:50.327810 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6gpsj" podUID="6b8950bf-8499-44e3-88e1-6ec33500c0c3" Feb 25 06:46:50 crc kubenswrapper[4978]: I0225 06:46:50.327447 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:50 crc kubenswrapper[4978]: I0225 06:46:50.327425 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:50 crc kubenswrapper[4978]: E0225 06:46:50.327903 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:50 crc kubenswrapper[4978]: E0225 06:46:50.327980 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:50 crc kubenswrapper[4978]: E0225 06:46:50.328152 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:50 crc kubenswrapper[4978]: I0225 06:46:50.534338 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:50 crc kubenswrapper[4978]: E0225 06:46:50.534576 4978 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 06:46:50 crc kubenswrapper[4978]: E0225 06:46:50.534687 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs podName:6b8950bf-8499-44e3-88e1-6ec33500c0c3 nodeName:}" failed. No retries permitted until 2026-02-25 06:46:54.534662345 +0000 UTC m=+107.973918814 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs") pod "network-metrics-daemon-6gpsj" (UID: "6b8950bf-8499-44e3-88e1-6ec33500c0c3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 06:46:50 crc kubenswrapper[4978]: I0225 06:46:50.589070 4978 generic.go:334] "Generic (PLEG): container finished" podID="594f940e-2757-4577-94b4-301e27ad4fab" containerID="70c981d15b6ad57f90ddd70ddc28bcfe8cfa4cca7749211db79875df38a1ed1b" exitCode=0 Feb 25 06:46:50 crc kubenswrapper[4978]: I0225 06:46:50.589139 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" event={"ID":"594f940e-2757-4577-94b4-301e27ad4fab","Type":"ContainerDied","Data":"70c981d15b6ad57f90ddd70ddc28bcfe8cfa4cca7749211db79875df38a1ed1b"} Feb 25 06:46:51 crc kubenswrapper[4978]: I0225 06:46:51.596189 4978 generic.go:334] "Generic (PLEG): container finished" podID="594f940e-2757-4577-94b4-301e27ad4fab" containerID="0e9af345ddc7e1b8538489ec33d156d083acc88e7586bbd489120d8707849cd1" exitCode=0 Feb 25 06:46:51 crc kubenswrapper[4978]: I0225 06:46:51.596299 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" event={"ID":"594f940e-2757-4577-94b4-301e27ad4fab","Type":"ContainerDied","Data":"0e9af345ddc7e1b8538489ec33d156d083acc88e7586bbd489120d8707849cd1"} Feb 25 06:46:51 crc kubenswrapper[4978]: I0225 06:46:51.600800 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerStarted","Data":"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22"} Feb 25 06:46:52 crc kubenswrapper[4978]: I0225 06:46:52.331094 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:52 crc kubenswrapper[4978]: I0225 06:46:52.331182 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:52 crc kubenswrapper[4978]: I0225 06:46:52.331207 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:52 crc kubenswrapper[4978]: I0225 06:46:52.331242 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:52 crc kubenswrapper[4978]: E0225 06:46:52.331290 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:52 crc kubenswrapper[4978]: E0225 06:46:52.331397 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:52 crc kubenswrapper[4978]: E0225 06:46:52.331533 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6gpsj" podUID="6b8950bf-8499-44e3-88e1-6ec33500c0c3" Feb 25 06:46:52 crc kubenswrapper[4978]: E0225 06:46:52.331627 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:52 crc kubenswrapper[4978]: I0225 06:46:52.607476 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" event={"ID":"594f940e-2757-4577-94b4-301e27ad4fab","Type":"ContainerStarted","Data":"a3b8e377755f14cfac2160ff257bfd93e37233bfe16d3ffceb770ead4ee31a39"} Feb 25 06:46:53 crc kubenswrapper[4978]: I0225 06:46:53.615080 4978 generic.go:334] "Generic (PLEG): container finished" podID="594f940e-2757-4577-94b4-301e27ad4fab" containerID="a3b8e377755f14cfac2160ff257bfd93e37233bfe16d3ffceb770ead4ee31a39" exitCode=0 Feb 25 06:46:53 crc kubenswrapper[4978]: I0225 06:46:53.615142 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" event={"ID":"594f940e-2757-4577-94b4-301e27ad4fab","Type":"ContainerDied","Data":"a3b8e377755f14cfac2160ff257bfd93e37233bfe16d3ffceb770ead4ee31a39"} Feb 25 06:46:53 crc kubenswrapper[4978]: I0225 06:46:53.619835 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerStarted","Data":"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768"} Feb 25 06:46:53 crc kubenswrapper[4978]: I0225 06:46:53.620301 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:53 crc kubenswrapper[4978]: I0225 06:46:53.620357 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:53 crc kubenswrapper[4978]: I0225 06:46:53.620419 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:53 crc kubenswrapper[4978]: I0225 06:46:53.666670 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:53 crc kubenswrapper[4978]: I0225 06:46:53.669579 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" podStartSLOduration=55.669562178 podStartE2EDuration="55.669562178s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:46:53.66929645 +0000 UTC m=+107.108552939" watchObservedRunningTime="2026-02-25 06:46:53.669562178 +0000 UTC m=+107.108818657" Feb 25 06:46:53 crc kubenswrapper[4978]: I0225 06:46:53.671427 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:46:54 crc kubenswrapper[4978]: I0225 06:46:54.327664 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:54 crc kubenswrapper[4978]: I0225 06:46:54.327703 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:54 crc kubenswrapper[4978]: I0225 06:46:54.327774 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:54 crc kubenswrapper[4978]: I0225 06:46:54.327672 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:54 crc kubenswrapper[4978]: E0225 06:46:54.327832 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:54 crc kubenswrapper[4978]: E0225 06:46:54.327903 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:54 crc kubenswrapper[4978]: E0225 06:46:54.327980 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6gpsj" podUID="6b8950bf-8499-44e3-88e1-6ec33500c0c3" Feb 25 06:46:54 crc kubenswrapper[4978]: E0225 06:46:54.328385 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:54 crc kubenswrapper[4978]: I0225 06:46:54.328819 4978 scope.go:117] "RemoveContainer" containerID="4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5" Feb 25 06:46:54 crc kubenswrapper[4978]: I0225 06:46:54.580399 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:54 crc kubenswrapper[4978]: E0225 06:46:54.580582 4978 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 06:46:54 crc kubenswrapper[4978]: E0225 06:46:54.580799 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs podName:6b8950bf-8499-44e3-88e1-6ec33500c0c3 nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.580784461 +0000 UTC m=+116.020040910 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs") pod "network-metrics-daemon-6gpsj" (UID: "6b8950bf-8499-44e3-88e1-6ec33500c0c3") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 25 06:46:54 crc kubenswrapper[4978]: I0225 06:46:54.631896 4978 generic.go:334] "Generic (PLEG): container finished" podID="594f940e-2757-4577-94b4-301e27ad4fab" containerID="00f100ecd9ca3479c3a212eddf084a55985ded53e672a6521dd30df0051377b4" exitCode=0 Feb 25 06:46:54 crc kubenswrapper[4978]: I0225 06:46:54.631939 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" event={"ID":"594f940e-2757-4577-94b4-301e27ad4fab","Type":"ContainerDied","Data":"00f100ecd9ca3479c3a212eddf084a55985ded53e672a6521dd30df0051377b4"} Feb 25 06:46:54 crc kubenswrapper[4978]: I0225 06:46:54.636565 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 25 06:46:54 crc kubenswrapper[4978]: I0225 06:46:54.641499 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19"} Feb 25 06:46:55 crc kubenswrapper[4978]: I0225 06:46:55.457651 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=12.457558838 podStartE2EDuration="12.457558838s" podCreationTimestamp="2026-02-25 06:46:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:46:54.690874302 +0000 UTC m=+108.130130801" watchObservedRunningTime="2026-02-25 06:46:55.457558838 +0000 UTC m=+108.896815287" Feb 25 06:46:55 crc kubenswrapper[4978]: I0225 06:46:55.458650 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6gpsj"] Feb 25 06:46:55 crc kubenswrapper[4978]: I0225 06:46:55.458801 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:55 crc kubenswrapper[4978]: E0225 06:46:55.458917 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6gpsj" podUID="6b8950bf-8499-44e3-88e1-6ec33500c0c3" Feb 25 06:46:55 crc kubenswrapper[4978]: I0225 06:46:55.651223 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" event={"ID":"594f940e-2757-4577-94b4-301e27ad4fab","Type":"ContainerStarted","Data":"ca678ef94d87a3462a8479095e37e8b199ec63b463bba9a0a50438e764ea436b"} Feb 25 06:46:55 crc kubenswrapper[4978]: I0225 06:46:55.680278 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rjvfr" podStartSLOduration=57.680253126 podStartE2EDuration="57.680253126s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:46:55.679080131 +0000 UTC m=+109.118336620" watchObservedRunningTime="2026-02-25 06:46:55.680253126 +0000 UTC m=+109.119509625" Feb 25 06:46:56 crc kubenswrapper[4978]: I0225 06:46:56.326966 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:56 crc kubenswrapper[4978]: E0225 06:46:56.327156 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:56 crc kubenswrapper[4978]: I0225 06:46:56.327238 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:56 crc kubenswrapper[4978]: E0225 06:46:56.327293 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:56 crc kubenswrapper[4978]: I0225 06:46:56.327352 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:56 crc kubenswrapper[4978]: E0225 06:46:56.327434 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:57 crc kubenswrapper[4978]: I0225 06:46:57.327816 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:57 crc kubenswrapper[4978]: E0225 06:46:57.329146 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-6gpsj" podUID="6b8950bf-8499-44e3-88e1-6ec33500c0c3" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.327347 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.327448 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.328154 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:46:58 crc kubenswrapper[4978]: E0225 06:46:58.328462 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 25 06:46:58 crc kubenswrapper[4978]: E0225 06:46:58.328607 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 25 06:46:58 crc kubenswrapper[4978]: E0225 06:46:58.329749 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.350605 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.580098 4978 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.580297 4978 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.627041 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.628259 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.637909 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.638336 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.639188 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.639590 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.649478 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qcr2x"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.650535 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gkwg7"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.650936 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.651598 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.652030 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.652872 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.664791 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.665264 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.665309 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.665448 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.665766 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.665809 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.666113 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.666273 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.666313 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.666659 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.666698 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.667006 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.667493 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.667773 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.667878 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-md9t9"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.669058 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.669481 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.669569 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.671604 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.672501 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-jmhxz"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.672958 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-jmhxz" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.673021 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.673504 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rrcbj"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.674100 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.674800 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-r5cx2"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.675350 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.691848 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.692066 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.692260 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.692708 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.692742 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.692793 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.692855 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.692964 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.693019 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.693146 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.693254 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.693637 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.697061 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.697198 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.697312 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.697445 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.697502 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.697462 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.697989 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.698124 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.698138 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.698229 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.698283 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.698663 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.699038 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.699298 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.699547 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.699754 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.699958 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.700154 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.700349 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.700656 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.700715 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bgbqw"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.701397 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.702101 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.702509 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.702761 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.702878 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.702961 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.703033 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.703101 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.703182 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.703295 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.703471 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.703549 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.703803 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.703874 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.703946 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.707064 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.710883 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ljw8k"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.711790 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nhkg4"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.712195 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.712443 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-grzdq"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.713110 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.722084 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.722282 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-tqzfq"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.733922 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-serving-cert\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.733969 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81f10398-4c3a-4642-9e4f-df746612a9b2-serving-cert\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.734000 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.734031 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2dt2\" (UniqueName: \"kubernetes.io/projected/81f10398-4c3a-4642-9e4f-df746612a9b2-kube-api-access-f2dt2\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.734069 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2vwr\" (UniqueName: \"kubernetes.io/projected/411cf3c7-6262-4458-a3bd-245b2c002260-kube-api-access-t2vwr\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.734094 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81f10398-4c3a-4642-9e4f-df746612a9b2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.734132 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h25wg\" (UniqueName: \"kubernetes.io/projected/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-kube-api-access-h25wg\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.734541 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81f10398-4c3a-4642-9e4f-df746612a9b2-encryption-config\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.734571 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/81f10398-4c3a-4642-9e4f-df746612a9b2-audit-policies\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.734842 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81f10398-4c3a-4642-9e4f-df746612a9b2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.734923 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-config\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.735070 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/411cf3c7-6262-4458-a3bd-245b2c002260-images\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.753624 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/411cf3c7-6262-4458-a3bd-245b2c002260-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.753659 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81f10398-4c3a-4642-9e4f-df746612a9b2-etcd-client\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.753677 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-client-ca\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.753697 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/411cf3c7-6262-4458-a3bd-245b2c002260-config\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.753748 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81f10398-4c3a-4642-9e4f-df746612a9b2-audit-dir\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.754538 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.754785 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.754924 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.755262 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.755321 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.755608 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.755940 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.755982 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.756193 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.756560 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.756640 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.756967 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.757064 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.757162 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.757258 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.757333 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.757503 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.757566 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.757617 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.757652 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.757818 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.757629 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.758058 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.758062 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.758114 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.758326 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.760094 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.760227 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.761505 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.765740 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.768457 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.768674 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.771931 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w6cv7"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.772594 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.772659 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.773028 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.773047 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.773064 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.773222 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.773242 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.773470 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.773651 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.776712 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.778237 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.779356 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.779524 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.779847 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.779985 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.780082 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.780165 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.779859 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.780316 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.787308 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.795139 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.795450 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.795721 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-crqpx"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.796402 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.797152 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.797411 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.799917 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.809926 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qf9f"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.810657 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.810671 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.812596 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.815193 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.815541 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.817137 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.817582 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.817707 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.818098 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.818213 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.818407 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.818670 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.818942 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.820404 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.821056 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-ml6mx"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.821579 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.822552 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.823304 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.823829 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.823978 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.824104 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.825447 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.825819 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.825909 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.826435 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.826702 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.826850 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-wqj9s"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.827553 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-md9t9"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.827629 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.828385 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qcr2x"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.830625 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gkwg7"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.830780 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.831465 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.833160 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.833379 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.834351 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.835640 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rrcbj"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.836715 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bgbqw"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.838113 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.838659 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-2bjgq"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.839675 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qf9f"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.839771 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2bjgq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.842417 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-r5cx2"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.843574 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.844491 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nhkg4"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.845932 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-grzdq"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.846922 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.848010 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-tqzfq"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.849317 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.850298 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.850815 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.851385 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.852340 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854129 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w6cv7"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854512 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lrvb\" (UniqueName: \"kubernetes.io/projected/3494cb07-ccde-443f-8286-f73a5d9ab58f-kube-api-access-8lrvb\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854544 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h25wg\" (UniqueName: \"kubernetes.io/projected/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-kube-api-access-h25wg\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854568 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fhjl\" (UniqueName: \"kubernetes.io/projected/b93c351c-80f1-4568-8251-14d138884828-kube-api-access-9fhjl\") pod \"openshift-apiserver-operator-796bbdcf4f-4zmq2\" (UID: \"b93c351c-80f1-4568-8251-14d138884828\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854589 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-service-ca\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854615 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5c9a6a73-8690-43ac-928f-153713a523f6-node-pullsecrets\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854635 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79t9v\" (UniqueName: \"kubernetes.io/projected/90b4f63b-4ce2-46db-8f97-ea117c763799-kube-api-access-79t9v\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854656 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854737 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5c9a6a73-8690-43ac-928f-153713a523f6-encryption-config\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854821 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/daa8133c-0938-4871-a573-f7025bee1103-metrics-tls\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854860 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-serving-cert\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854907 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-client-ca\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854950 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.854989 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-audit\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855032 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea47c114-e38b-424a-9a46-c0c67dce2947-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-svrlg\" (UID: \"ea47c114-e38b-424a-9a46-c0c67dce2947\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855057 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855171 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3414da22-9e7f-4ae1-953e-b56ec8e78a2d-metrics-tls\") pod \"dns-operator-744455d44c-grzdq\" (UID: \"3414da22-9e7f-4ae1-953e-b56ec8e78a2d\") " pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855203 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw5p5\" (UniqueName: \"kubernetes.io/projected/3414da22-9e7f-4ae1-953e-b56ec8e78a2d-kube-api-access-bw5p5\") pod \"dns-operator-744455d44c-grzdq\" (UID: \"3414da22-9e7f-4ae1-953e-b56ec8e78a2d\") " pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855245 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1fc2a49b-47ae-4601-a397-c96d56fd10cb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-md9t9\" (UID: \"1fc2a49b-47ae-4601-a397-c96d56fd10cb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855263 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855284 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855327 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90b4f63b-4ce2-46db-8f97-ea117c763799-service-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855346 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-etcd-client\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855381 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-machine-approver-tls\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855400 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855425 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81f10398-4c3a-4642-9e4f-df746612a9b2-encryption-config\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855447 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855471 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/81f10398-4c3a-4642-9e4f-df746612a9b2-audit-policies\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855497 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81f10398-4c3a-4642-9e4f-df746612a9b2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855518 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-oauth-config\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855536 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5c9a6a73-8690-43ac-928f-153713a523f6-audit-dir\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855554 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b93c351c-80f1-4568-8251-14d138884828-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4zmq2\" (UID: \"b93c351c-80f1-4568-8251-14d138884828\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855576 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-config\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855593 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/411cf3c7-6262-4458-a3bd-245b2c002260-images\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.855613 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/411cf3c7-6262-4458-a3bd-245b2c002260-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856107 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-serving-cert\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856149 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-etcd-ca\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856179 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81f10398-4c3a-4642-9e4f-df746612a9b2-etcd-client\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856197 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-client-ca\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856217 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/411cf3c7-6262-4458-a3bd-245b2c002260-config\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856238 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b4f63b-4ce2-46db-8f97-ea117c763799-config\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856256 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b93c351c-80f1-4568-8251-14d138884828-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4zmq2\" (UID: \"b93c351c-80f1-4568-8251-14d138884828\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856274 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-config\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856296 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856315 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-serving-cert\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856332 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-image-import-ca\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856351 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqltp\" (UniqueName: \"kubernetes.io/projected/5c9a6a73-8690-43ac-928f-153713a523f6-kube-api-access-gqltp\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856383 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/39e6256a-c532-4534-bcda-7531865d965c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ncnfh\" (UID: \"39e6256a-c532-4534-bcda-7531865d965c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856401 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-dir\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856418 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea47c114-e38b-424a-9a46-c0c67dce2947-config\") pod \"kube-controller-manager-operator-78b949d7b-svrlg\" (UID: \"ea47c114-e38b-424a-9a46-c0c67dce2947\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856440 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81f10398-4c3a-4642-9e4f-df746612a9b2-audit-dir\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856460 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14b0df6-0f54-47fb-8584-1dcdcd54f735-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kgnwp\" (UID: \"d14b0df6-0f54-47fb-8584-1dcdcd54f735\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856479 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856493 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/81f10398-4c3a-4642-9e4f-df746612a9b2-audit-policies\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856497 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pktd4\" (UniqueName: \"kubernetes.io/projected/028ba3b7-0bca-4541-a5be-7f63b26bdea4-kube-api-access-pktd4\") pod \"downloads-7954f5f757-jmhxz\" (UID: \"028ba3b7-0bca-4541-a5be-7f63b26bdea4\") " pod="openshift-console/downloads-7954f5f757-jmhxz" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856553 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856582 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856585 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81f10398-4c3a-4642-9e4f-df746612a9b2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856607 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6150f41-517c-4408-9a8e-a553c4f28ecb-serving-cert\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856636 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5j2n\" (UniqueName: \"kubernetes.io/projected/1fc2a49b-47ae-4601-a397-c96d56fd10cb-kube-api-access-x5j2n\") pod \"openshift-config-operator-7777fb866f-md9t9\" (UID: \"1fc2a49b-47ae-4601-a397-c96d56fd10cb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.856661 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-config\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.858267 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-client-ca\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.859298 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wmbf\" (UniqueName: \"kubernetes.io/projected/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-kube-api-access-9wmbf\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.859319 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/411cf3c7-6262-4458-a3bd-245b2c002260-config\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.859582 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/411cf3c7-6262-4458-a3bd-245b2c002260-images\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.859634 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81f10398-4c3a-4642-9e4f-df746612a9b2-audit-dir\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.859651 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-config\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.859686 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-policies\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.859813 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b4qn\" (UniqueName: \"kubernetes.io/projected/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-kube-api-access-6b4qn\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860136 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/daa8133c-0938-4871-a573-f7025bee1103-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860183 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-config\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860296 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-trusted-ca-bundle\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860349 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c9a6a73-8690-43ac-928f-153713a523f6-etcd-client\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860432 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9p9q\" (UniqueName: \"kubernetes.io/projected/39e6256a-c532-4534-bcda-7531865d965c-kube-api-access-t9p9q\") pod \"cluster-samples-operator-665b6dd947-ncnfh\" (UID: \"39e6256a-c532-4534-bcda-7531865d965c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860477 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-oauth-serving-cert\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860516 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c9a6a73-8690-43ac-928f-153713a523f6-serving-cert\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860597 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-trusted-ca\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860668 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzm7z\" (UniqueName: \"kubernetes.io/projected/e6150f41-517c-4408-9a8e-a553c4f28ecb-kube-api-access-jzm7z\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860743 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-auth-proxy-config\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860823 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.861436 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81f10398-4c3a-4642-9e4f-df746612a9b2-etcd-client\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.861635 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/411cf3c7-6262-4458-a3bd-245b2c002260-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.860936 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-serving-cert\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.861752 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.861795 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81f10398-4c3a-4642-9e4f-df746612a9b2-serving-cert\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.861868 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m67qm\" (UniqueName: \"kubernetes.io/projected/d14b0df6-0f54-47fb-8584-1dcdcd54f735-kube-api-access-m67qm\") pod \"openshift-controller-manager-operator-756b6f6bc6-kgnwp\" (UID: \"d14b0df6-0f54-47fb-8584-1dcdcd54f735\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.861905 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bb86\" (UniqueName: \"kubernetes.io/projected/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-kube-api-access-5bb86\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.861942 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.861975 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea47c114-e38b-424a-9a46-c0c67dce2947-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-svrlg\" (UID: \"ea47c114-e38b-424a-9a46-c0c67dce2947\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862010 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2dt2\" (UniqueName: \"kubernetes.io/projected/81f10398-4c3a-4642-9e4f-df746612a9b2-kube-api-access-f2dt2\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862042 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2vwr\" (UniqueName: \"kubernetes.io/projected/411cf3c7-6262-4458-a3bd-245b2c002260-kube-api-access-t2vwr\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862076 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq67m\" (UniqueName: \"kubernetes.io/projected/daa8133c-0938-4871-a573-f7025bee1103-kube-api-access-gq67m\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862113 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fc2a49b-47ae-4601-a397-c96d56fd10cb-serving-cert\") pod \"openshift-config-operator-7777fb866f-md9t9\" (UID: \"1fc2a49b-47ae-4601-a397-c96d56fd10cb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862145 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-config\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862174 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862208 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81f10398-4c3a-4642-9e4f-df746612a9b2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862241 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4m8t\" (UniqueName: \"kubernetes.io/projected/b51a2671-2dea-418e-8f99-512e4ba4e34c-kube-api-access-c4m8t\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862314 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d14b0df6-0f54-47fb-8584-1dcdcd54f735-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kgnwp\" (UID: \"d14b0df6-0f54-47fb-8584-1dcdcd54f735\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862349 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-config\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862400 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-config\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862440 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90b4f63b-4ce2-46db-8f97-ea117c763799-serving-cert\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862467 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81f10398-4c3a-4642-9e4f-df746612a9b2-encryption-config\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862510 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ljw8k"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862578 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/daa8133c-0938-4871-a573-f7025bee1103-trusted-ca\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862616 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-etcd-service-ca\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862700 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90b4f63b-4ce2-46db-8f97-ea117c763799-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.862874 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81f10398-4c3a-4642-9e4f-df746612a9b2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.863215 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.863881 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-serving-cert\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.867282 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-jmhxz"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.867820 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81f10398-4c3a-4642-9e4f-df746612a9b2-serving-cert\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.868180 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.871307 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.872433 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.874054 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-wqj9s"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.875553 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.876314 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.883218 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-crqpx"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.884514 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.885407 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.887127 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.887568 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.890893 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.891644 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.891721 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-mhnd6"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.893463 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6j76w"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.893713 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.895148 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2bjgq"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.895254 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6j76w"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.895289 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.895629 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-mxdv7"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.896197 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.896483 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dqrgj"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.897436 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dqrgj"] Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.897868 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dqrgj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.911537 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.930865 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.950960 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963328 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-serving-cert\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963363 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-etcd-ca\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963396 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-serving-cert\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963413 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-image-import-ca\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963429 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqltp\" (UniqueName: \"kubernetes.io/projected/5c9a6a73-8690-43ac-928f-153713a523f6-kube-api-access-gqltp\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963446 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b4f63b-4ce2-46db-8f97-ea117c763799-config\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963464 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b93c351c-80f1-4568-8251-14d138884828-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4zmq2\" (UID: \"b93c351c-80f1-4568-8251-14d138884828\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963485 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-config\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963506 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963527 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/39e6256a-c532-4534-bcda-7531865d965c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ncnfh\" (UID: \"39e6256a-c532-4534-bcda-7531865d965c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963545 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-dir\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963566 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea47c114-e38b-424a-9a46-c0c67dce2947-config\") pod \"kube-controller-manager-operator-78b949d7b-svrlg\" (UID: \"ea47c114-e38b-424a-9a46-c0c67dce2947\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963583 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14b0df6-0f54-47fb-8584-1dcdcd54f735-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kgnwp\" (UID: \"d14b0df6-0f54-47fb-8584-1dcdcd54f735\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963603 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963620 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pktd4\" (UniqueName: \"kubernetes.io/projected/028ba3b7-0bca-4541-a5be-7f63b26bdea4-kube-api-access-pktd4\") pod \"downloads-7954f5f757-jmhxz\" (UID: \"028ba3b7-0bca-4541-a5be-7f63b26bdea4\") " pod="openshift-console/downloads-7954f5f757-jmhxz" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963637 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5j2n\" (UniqueName: \"kubernetes.io/projected/1fc2a49b-47ae-4601-a397-c96d56fd10cb-kube-api-access-x5j2n\") pod \"openshift-config-operator-7777fb866f-md9t9\" (UID: \"1fc2a49b-47ae-4601-a397-c96d56fd10cb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963654 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-config\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963670 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963688 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963705 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6150f41-517c-4408-9a8e-a553c4f28ecb-serving-cert\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963724 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/daa8133c-0938-4871-a573-f7025bee1103-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963740 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-config\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963757 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wmbf\" (UniqueName: \"kubernetes.io/projected/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-kube-api-access-9wmbf\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963773 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-policies\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963792 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b4qn\" (UniqueName: \"kubernetes.io/projected/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-kube-api-access-6b4qn\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963808 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-trusted-ca-bundle\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963824 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c9a6a73-8690-43ac-928f-153713a523f6-etcd-client\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963849 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9p9q\" (UniqueName: \"kubernetes.io/projected/39e6256a-c532-4534-bcda-7531865d965c-kube-api-access-t9p9q\") pod \"cluster-samples-operator-665b6dd947-ncnfh\" (UID: \"39e6256a-c532-4534-bcda-7531865d965c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963867 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-oauth-serving-cert\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963882 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c9a6a73-8690-43ac-928f-153713a523f6-serving-cert\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.963898 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-trusted-ca\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964009 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-auth-proxy-config\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964032 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964054 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzm7z\" (UniqueName: \"kubernetes.io/projected/e6150f41-517c-4408-9a8e-a553c4f28ecb-kube-api-access-jzm7z\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964079 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964095 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea47c114-e38b-424a-9a46-c0c67dce2947-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-svrlg\" (UID: \"ea47c114-e38b-424a-9a46-c0c67dce2947\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964111 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m67qm\" (UniqueName: \"kubernetes.io/projected/d14b0df6-0f54-47fb-8584-1dcdcd54f735-kube-api-access-m67qm\") pod \"openshift-controller-manager-operator-756b6f6bc6-kgnwp\" (UID: \"d14b0df6-0f54-47fb-8584-1dcdcd54f735\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964128 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bb86\" (UniqueName: \"kubernetes.io/projected/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-kube-api-access-5bb86\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964156 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq67m\" (UniqueName: \"kubernetes.io/projected/daa8133c-0938-4871-a573-f7025bee1103-kube-api-access-gq67m\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964215 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-config\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964250 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964268 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fc2a49b-47ae-4601-a397-c96d56fd10cb-serving-cert\") pod \"openshift-config-operator-7777fb866f-md9t9\" (UID: \"1fc2a49b-47ae-4601-a397-c96d56fd10cb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964292 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4m8t\" (UniqueName: \"kubernetes.io/projected/b51a2671-2dea-418e-8f99-512e4ba4e34c-kube-api-access-c4m8t\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964311 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d14b0df6-0f54-47fb-8584-1dcdcd54f735-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kgnwp\" (UID: \"d14b0df6-0f54-47fb-8584-1dcdcd54f735\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964330 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-config\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964348 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90b4f63b-4ce2-46db-8f97-ea117c763799-serving-cert\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964341 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-dir\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964366 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-config\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964455 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/daa8133c-0938-4871-a573-f7025bee1103-trusted-ca\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964487 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-etcd-service-ca\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964520 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90b4f63b-4ce2-46db-8f97-ea117c763799-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964563 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lrvb\" (UniqueName: \"kubernetes.io/projected/3494cb07-ccde-443f-8286-f73a5d9ab58f-kube-api-access-8lrvb\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964591 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fhjl\" (UniqueName: \"kubernetes.io/projected/b93c351c-80f1-4568-8251-14d138884828-kube-api-access-9fhjl\") pod \"openshift-apiserver-operator-796bbdcf4f-4zmq2\" (UID: \"b93c351c-80f1-4568-8251-14d138884828\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964617 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-service-ca\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964641 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5c9a6a73-8690-43ac-928f-153713a523f6-node-pullsecrets\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964670 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79t9v\" (UniqueName: \"kubernetes.io/projected/90b4f63b-4ce2-46db-8f97-ea117c763799-kube-api-access-79t9v\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964683 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-image-import-ca\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.965263 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-config\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.966294 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-config\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.966622 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea47c114-e38b-424a-9a46-c0c67dce2947-config\") pod \"kube-controller-manager-operator-78b949d7b-svrlg\" (UID: \"ea47c114-e38b-424a-9a46-c0c67dce2947\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.967004 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/daa8133c-0938-4871-a573-f7025bee1103-trusted-ca\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.967089 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-policies\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.967589 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-service-ca\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.968240 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-trusted-ca-bundle\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.968294 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6150f41-517c-4408-9a8e-a553c4f28ecb-serving-cert\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.968359 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90b4f63b-4ce2-46db-8f97-ea117c763799-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.968501 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-serving-cert\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.968844 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/39e6256a-c532-4534-bcda-7531865d965c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ncnfh\" (UID: \"39e6256a-c532-4534-bcda-7531865d965c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.968865 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/daa8133c-0938-4871-a573-f7025bee1103-metrics-tls\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.964698 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/daa8133c-0938-4871-a573-f7025bee1103-metrics-tls\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.968929 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-serving-cert\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.968959 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.968980 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5c9a6a73-8690-43ac-928f-153713a523f6-node-pullsecrets\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.968984 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5c9a6a73-8690-43ac-928f-153713a523f6-encryption-config\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969054 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-client-ca\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969085 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea47c114-e38b-424a-9a46-c0c67dce2947-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-svrlg\" (UID: \"ea47c114-e38b-424a-9a46-c0c67dce2947\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969108 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-audit\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969137 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3414da22-9e7f-4ae1-953e-b56ec8e78a2d-metrics-tls\") pod \"dns-operator-744455d44c-grzdq\" (UID: \"3414da22-9e7f-4ae1-953e-b56ec8e78a2d\") " pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969167 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw5p5\" (UniqueName: \"kubernetes.io/projected/3414da22-9e7f-4ae1-953e-b56ec8e78a2d-kube-api-access-bw5p5\") pod \"dns-operator-744455d44c-grzdq\" (UID: \"3414da22-9e7f-4ae1-953e-b56ec8e78a2d\") " pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969195 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969218 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969237 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1fc2a49b-47ae-4601-a397-c96d56fd10cb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-md9t9\" (UID: \"1fc2a49b-47ae-4601-a397-c96d56fd10cb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969271 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969304 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969336 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90b4f63b-4ce2-46db-8f97-ea117c763799-service-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969353 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b93c351c-80f1-4568-8251-14d138884828-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-4zmq2\" (UID: \"b93c351c-80f1-4568-8251-14d138884828\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969368 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-etcd-client\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969388 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-auth-proxy-config\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969414 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-machine-approver-tls\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969445 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969479 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969509 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-oauth-config\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969535 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5c9a6a73-8690-43ac-928f-153713a523f6-audit-dir\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.969563 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b93c351c-80f1-4568-8251-14d138884828-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4zmq2\" (UID: \"b93c351c-80f1-4568-8251-14d138884828\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.970228 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90b4f63b-4ce2-46db-8f97-ea117c763799-config\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.970323 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-oauth-serving-cert\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.970462 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-config\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.970845 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-trusted-ca\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.971156 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-config\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.971319 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90b4f63b-4ce2-46db-8f97-ea117c763799-service-ca-bundle\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.971556 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/1fc2a49b-47ae-4601-a397-c96d56fd10cb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-md9t9\" (UID: \"1fc2a49b-47ae-4601-a397-c96d56fd10cb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.971749 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-audit\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.972506 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.972611 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5c9a6a73-8690-43ac-928f-153713a523f6-audit-dir\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.972669 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.972542 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.972861 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-client-ca\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.973090 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b93c351c-80f1-4568-8251-14d138884828-config\") pod \"openshift-apiserver-operator-796bbdcf4f-4zmq2\" (UID: \"b93c351c-80f1-4568-8251-14d138884828\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.973467 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-serving-cert\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.973498 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.974007 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-config\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.974174 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d14b0df6-0f54-47fb-8584-1dcdcd54f735-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kgnwp\" (UID: \"d14b0df6-0f54-47fb-8584-1dcdcd54f735\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.974403 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.974408 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c9a6a73-8690-43ac-928f-153713a523f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.974572 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.974587 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5c9a6a73-8690-43ac-928f-153713a523f6-encryption-config\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.975067 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c9a6a73-8690-43ac-928f-153713a523f6-serving-cert\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.975346 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90b4f63b-4ce2-46db-8f97-ea117c763799-serving-cert\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.975492 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea47c114-e38b-424a-9a46-c0c67dce2947-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-svrlg\" (UID: \"ea47c114-e38b-424a-9a46-c0c67dce2947\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.975537 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.975761 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.975884 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fc2a49b-47ae-4601-a397-c96d56fd10cb-serving-cert\") pod \"openshift-config-operator-7777fb866f-md9t9\" (UID: \"1fc2a49b-47ae-4601-a397-c96d56fd10cb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.976161 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-machine-approver-tls\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.976174 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c9a6a73-8690-43ac-928f-153713a523f6-etcd-client\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.976189 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.977138 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.977194 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3414da22-9e7f-4ae1-953e-b56ec8e78a2d-metrics-tls\") pod \"dns-operator-744455d44c-grzdq\" (UID: \"3414da22-9e7f-4ae1-953e-b56ec8e78a2d\") " pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.978391 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.983274 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-serving-cert\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.984726 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.991024 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-oauth-config\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:46:58 crc kubenswrapper[4978]: I0225 06:46:58.991842 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.011534 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.021276 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d14b0df6-0f54-47fb-8584-1dcdcd54f735-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kgnwp\" (UID: \"d14b0df6-0f54-47fb-8584-1dcdcd54f735\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.031321 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.038198 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-config\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.052415 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.067334 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-etcd-client\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.074694 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.084610 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-etcd-ca\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.092902 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.098561 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-etcd-service-ca\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.111524 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.151847 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.171505 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.191401 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.210726 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.232398 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.251802 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.271840 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.302222 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.311300 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.327141 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.332515 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.352708 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.372308 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.392522 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.412641 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.432629 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.451113 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.471097 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.492653 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.512415 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.531473 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.563151 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.571868 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.592019 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.613115 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.631616 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.651802 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.671089 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.691882 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.711261 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.732680 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.751527 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.770782 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.791614 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.811662 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.829460 4978 request.go:700] Waited for 1.011073392s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.831910 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.851565 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.871836 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.892009 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.912556 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.931948 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.950928 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.971788 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 25 06:46:59 crc kubenswrapper[4978]: I0225 06:46:59.992405 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.011772 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.031455 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.052112 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.072081 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.092088 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.111484 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.133707 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.152161 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.172059 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.192225 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.211190 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.233167 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.251999 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.272566 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.292238 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.311348 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.337334 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.337427 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.337432 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.340553 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.353281 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.376728 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.394773 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.410477 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.431225 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.451569 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.472093 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.514675 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h25wg\" (UniqueName: \"kubernetes.io/projected/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-kube-api-access-h25wg\") pod \"controller-manager-879f6c89f-qcr2x\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.531543 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2dt2\" (UniqueName: \"kubernetes.io/projected/81f10398-4c3a-4642-9e4f-df746612a9b2-kube-api-access-f2dt2\") pod \"apiserver-7bbb656c7d-g9xwf\" (UID: \"81f10398-4c3a-4642-9e4f-df746612a9b2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.552446 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2vwr\" (UniqueName: \"kubernetes.io/projected/411cf3c7-6262-4458-a3bd-245b2c002260-kube-api-access-t2vwr\") pod \"machine-api-operator-5694c8668f-gkwg7\" (UID: \"411cf3c7-6262-4458-a3bd-245b2c002260\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.572303 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.591136 4978 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.612668 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.632554 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.652484 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.671737 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.692437 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.711770 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.732880 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.751949 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.783693 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.796780 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.797918 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lrvb\" (UniqueName: \"kubernetes.io/projected/3494cb07-ccde-443f-8286-f73a5d9ab58f-kube-api-access-8lrvb\") pod \"oauth-openshift-558db77b4-bgbqw\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.806248 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.812603 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/daa8133c-0938-4871-a573-f7025bee1103-bound-sa-token\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.837230 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzm7z\" (UniqueName: \"kubernetes.io/projected/e6150f41-517c-4408-9a8e-a553c4f28ecb-kube-api-access-jzm7z\") pod \"route-controller-manager-6576b87f9c-cmkt5\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.849874 4978 request.go:700] Waited for 1.883052974s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/default/token Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.850936 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wmbf\" (UniqueName: \"kubernetes.io/projected/3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf-kube-api-access-9wmbf\") pod \"etcd-operator-b45778765-tqzfq\" (UID: \"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.857685 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.868894 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pktd4\" (UniqueName: \"kubernetes.io/projected/028ba3b7-0bca-4541-a5be-7f63b26bdea4-kube-api-access-pktd4\") pod \"downloads-7954f5f757-jmhxz\" (UID: \"028ba3b7-0bca-4541-a5be-7f63b26bdea4\") " pod="openshift-console/downloads-7954f5f757-jmhxz" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.896720 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5j2n\" (UniqueName: \"kubernetes.io/projected/1fc2a49b-47ae-4601-a397-c96d56fd10cb-kube-api-access-x5j2n\") pod \"openshift-config-operator-7777fb866f-md9t9\" (UID: \"1fc2a49b-47ae-4601-a397-c96d56fd10cb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.919351 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b4qn\" (UniqueName: \"kubernetes.io/projected/8ea11877-f9ce-4ea1-b885-a4da7bcb63fe-kube-api-access-6b4qn\") pod \"console-operator-58897d9998-rrcbj\" (UID: \"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe\") " pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.924434 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-jmhxz" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.931088 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fhjl\" (UniqueName: \"kubernetes.io/projected/b93c351c-80f1-4568-8251-14d138884828-kube-api-access-9fhjl\") pod \"openshift-apiserver-operator-796bbdcf4f-4zmq2\" (UID: \"b93c351c-80f1-4568-8251-14d138884828\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.943607 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.954768 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea47c114-e38b-424a-9a46-c0c67dce2947-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-svrlg\" (UID: \"ea47c114-e38b-424a-9a46-c0c67dce2947\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.959091 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.970465 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m67qm\" (UniqueName: \"kubernetes.io/projected/d14b0df6-0f54-47fb-8584-1dcdcd54f735-kube-api-access-m67qm\") pod \"openshift-controller-manager-operator-756b6f6bc6-kgnwp\" (UID: \"d14b0df6-0f54-47fb-8584-1dcdcd54f735\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:47:00 crc kubenswrapper[4978]: I0225 06:47:00.986564 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bb86\" (UniqueName: \"kubernetes.io/projected/32ebcc4f-f0f3-4c1b-9efd-510363f3138a-kube-api-access-5bb86\") pod \"machine-approver-56656f9798-2fbhg\" (UID: \"32ebcc4f-f0f3-4c1b-9efd-510363f3138a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.008649 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9p9q\" (UniqueName: \"kubernetes.io/projected/39e6256a-c532-4534-bcda-7531865d965c-kube-api-access-t9p9q\") pod \"cluster-samples-operator-665b6dd947-ncnfh\" (UID: \"39e6256a-c532-4534-bcda-7531865d965c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.009240 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.014231 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gkwg7"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.020883 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.027691 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.031653 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq67m\" (UniqueName: \"kubernetes.io/projected/daa8133c-0938-4871-a573-f7025bee1103-kube-api-access-gq67m\") pod \"ingress-operator-5b745b69d9-vc5bs\" (UID: \"daa8133c-0938-4871-a573-f7025bee1103\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.051921 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqltp\" (UniqueName: \"kubernetes.io/projected/5c9a6a73-8690-43ac-928f-153713a523f6-kube-api-access-gqltp\") pod \"apiserver-76f77b778f-ljw8k\" (UID: \"5c9a6a73-8690-43ac-928f-153713a523f6\") " pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:47:01 crc kubenswrapper[4978]: W0225 06:47:01.056125 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod411cf3c7_6262_4458_a3bd_245b2c002260.slice/crio-ccd0c6217c9af5ea827a3dc6b25915e06ae248989221b20eb9f3c831e2c1a2b8 WatchSource:0}: Error finding container ccd0c6217c9af5ea827a3dc6b25915e06ae248989221b20eb9f3c831e2c1a2b8: Status 404 returned error can't find the container with id ccd0c6217c9af5ea827a3dc6b25915e06ae248989221b20eb9f3c831e2c1a2b8 Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.066039 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qcr2x"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.069060 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79t9v\" (UniqueName: \"kubernetes.io/projected/90b4f63b-4ce2-46db-8f97-ea117c763799-kube-api-access-79t9v\") pod \"authentication-operator-69f744f599-nhkg4\" (UID: \"90b4f63b-4ce2-46db-8f97-ea117c763799\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.091838 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw5p5\" (UniqueName: \"kubernetes.io/projected/3414da22-9e7f-4ae1-953e-b56ec8e78a2d-kube-api-access-bw5p5\") pod \"dns-operator-744455d44c-grzdq\" (UID: \"3414da22-9e7f-4ae1-953e-b56ec8e78a2d\") " pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.101377 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.109357 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4m8t\" (UniqueName: \"kubernetes.io/projected/b51a2671-2dea-418e-8f99-512e4ba4e34c-kube-api-access-c4m8t\") pod \"console-f9d7485db-r5cx2\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.133140 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.140810 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.140789984 podStartE2EDuration="3.140789984s" podCreationTimestamp="2026-02-25 06:46:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:01.140213606 +0000 UTC m=+114.579470085" watchObservedRunningTime="2026-02-25 06:47:01.140789984 +0000 UTC m=+114.580046453" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.151063 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.171891 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.173118 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.189832 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.194807 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.211677 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.217643 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.233793 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.234285 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.245144 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.251676 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.290454 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.290680 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.294132 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.303210 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.367162 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-jmhxz"] Feb 25 06:47:01 crc kubenswrapper[4978]: W0225 06:47:01.403116 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod028ba3b7_0bca_4541_a5be_7f63b26bdea4.slice/crio-e6fbd45c2fe8aeef57896ca1e7c754b5a845eb590f6e8ebc160cde6691be4d31 WatchSource:0}: Error finding container e6fbd45c2fe8aeef57896ca1e7c754b5a845eb590f6e8ebc160cde6691be4d31: Status 404 returned error can't find the container with id e6fbd45c2fe8aeef57896ca1e7c754b5a845eb590f6e8ebc160cde6691be4d31 Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.419255 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.426623 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.502178 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.623229 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.634869 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-r5cx2"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.655757 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-md9t9"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.903727 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-bound-sa-token\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.903997 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" event={"ID":"e6150f41-517c-4408-9a8e-a553c4f28ecb","Type":"ContainerStarted","Data":"7df80d639d9c2311238ed4cf984ef9286efda80bf1927b95181e9d9bac88608a"} Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.904326 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df24m\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-kube-api-access-df24m\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.904777 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-tls\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.905027 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-trusted-ca\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.907149 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ljw8k"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.908201 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" event={"ID":"32ebcc4f-f0f3-4c1b-9efd-510363f3138a","Type":"ContainerStarted","Data":"197edd2805130e8d1674699bb21582b40c0b2633683d73ee8d6b3389fab56b9e"} Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.908793 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:01 crc kubenswrapper[4978]: E0225 06:47:01.909418 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.409359026 +0000 UTC m=+115.848615485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.910012 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-grzdq"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.912448 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.919193 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.921916 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" event={"ID":"411cf3c7-6262-4458-a3bd-245b2c002260","Type":"ContainerStarted","Data":"ccd0c6217c9af5ea827a3dc6b25915e06ae248989221b20eb9f3c831e2c1a2b8"} Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.923306 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-certificates\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.925041 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" event={"ID":"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f","Type":"ContainerStarted","Data":"d1782f842647f2c06ef2802e1ba932a4b61679dc816c347acf60cb3f28c7bd5c"} Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.932736 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-jmhxz" event={"ID":"028ba3b7-0bca-4541-a5be-7f63b26bdea4","Type":"ContainerStarted","Data":"e6fbd45c2fe8aeef57896ca1e7c754b5a845eb590f6e8ebc160cde6691be4d31"} Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.934293 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.938500 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-nhkg4"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.942489 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-tqzfq"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.943618 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rrcbj"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.944891 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bgbqw"] Feb 25 06:47:01 crc kubenswrapper[4978]: I0225 06:47:01.946084 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs"] Feb 25 06:47:01 crc kubenswrapper[4978]: W0225 06:47:01.954549 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90b4f63b_4ce2_46db_8f97_ea117c763799.slice/crio-6970df8793844183b14b1f3ba41484005d87158d41e2f2edd6cd8a241b583501 WatchSource:0}: Error finding container 6970df8793844183b14b1f3ba41484005d87158d41e2f2edd6cd8a241b583501: Status 404 returned error can't find the container with id 6970df8793844183b14b1f3ba41484005d87158d41e2f2edd6cd8a241b583501 Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.036508 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:02 crc kubenswrapper[4978]: E0225 06:47:02.036667 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.536644599 +0000 UTC m=+115.975901058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037160 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg4ml\" (UniqueName: \"kubernetes.io/projected/20061d67-2059-4022-97ad-607c9f85540c-kube-api-access-rg4ml\") pod \"control-plane-machine-set-operator-78cbb6b69f-rwfz5\" (UID: \"20061d67-2059-4022-97ad-607c9f85540c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037188 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/441c14a4-92eb-4853-a964-2a8e1bd67415-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tld2j\" (UID: \"441c14a4-92eb-4853-a964-2a8e1bd67415\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037219 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7cb83e77-c92b-4736-91f2-61c8a1241eb0-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037246 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9qf9f\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037263 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9bde1ae3-e13f-4020-a11d-73a1c15d1759-images\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037282 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0b5a34af-02b0-488a-8e20-2d71c48a9c5e-metrics-tls\") pod \"dns-default-dqrgj\" (UID: \"0b5a34af-02b0-488a-8e20-2d71c48a9c5e\") " pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037296 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtvtd\" (UniqueName: \"kubernetes.io/projected/0b5a34af-02b0-488a-8e20-2d71c48a9c5e-kube-api-access-xtvtd\") pod \"dns-default-dqrgj\" (UID: \"0b5a34af-02b0-488a-8e20-2d71c48a9c5e\") " pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037315 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7xkz\" (UniqueName: \"kubernetes.io/projected/7fd9f3d7-d711-43cb-a896-7c77507621bc-kube-api-access-n7xkz\") pod \"collect-profiles-29533365-krpxk\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037329 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d4b61b96-afd0-47a6-be02-ead379d948b2-srv-cert\") pod \"olm-operator-6b444d44fb-9d8lt\" (UID: \"d4b61b96-afd0-47a6-be02-ead379d948b2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037357 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6hh5\" (UniqueName: \"kubernetes.io/projected/b0726188-e635-49f0-97ce-2cad95e63c13-kube-api-access-p6hh5\") pod \"catalog-operator-68c6474976-8wxst\" (UID: \"b0726188-e635-49f0-97ce-2cad95e63c13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037396 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9bde1ae3-e13f-4020-a11d-73a1c15d1759-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037413 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/811afd11-d673-4ad8-b6ba-4d766d6affe5-service-ca-bundle\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037449 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b5a34af-02b0-488a-8e20-2d71c48a9c5e-config-volume\") pod \"dns-default-dqrgj\" (UID: \"0b5a34af-02b0-488a-8e20-2d71c48a9c5e\") " pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037468 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtlhx\" (UniqueName: \"kubernetes.io/projected/9b843a14-9586-4f29-bfba-8413f4e52104-kube-api-access-rtlhx\") pod \"marketplace-operator-79b997595-9qf9f\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037486 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9dqg\" (UniqueName: \"kubernetes.io/projected/31337b76-7c44-42ab-a947-24c2932e2110-kube-api-access-g9dqg\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037504 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzn5t\" (UniqueName: \"kubernetes.io/projected/bb0ed3e8-663a-4605-9c73-78369c878589-kube-api-access-wzn5t\") pod \"service-ca-operator-777779d784-hb5xc\" (UID: \"bb0ed3e8-663a-4605-9c73-78369c878589\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037519 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e-node-bootstrap-token\") pod \"machine-config-server-mxdv7\" (UID: \"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e\") " pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037537 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/20061d67-2059-4022-97ad-607c9f85540c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rwfz5\" (UID: \"20061d67-2059-4022-97ad-607c9f85540c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037553 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt5bx\" (UniqueName: \"kubernetes.io/projected/21e98fa4-f620-4124-952b-55ff3782a39a-kube-api-access-qt5bx\") pod \"service-ca-9c57cc56f-crqpx\" (UID: \"21e98fa4-f620-4124-952b-55ff3782a39a\") " pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037574 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f8ca91-000f-46a3-b4ac-b04b834530e4-config\") pod \"kube-apiserver-operator-766d6c64bb-8cv8g\" (UID: \"d8f8ca91-000f-46a3-b4ac-b04b834530e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037589 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjtbj\" (UniqueName: \"kubernetes.io/projected/69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb-kube-api-access-zjtbj\") pod \"ingress-canary-2bjgq\" (UID: \"69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb\") " pod="openshift-ingress-canary/ingress-canary-2bjgq" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037619 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de7a95f8-3a89-4b73-b2c3-259880310e7a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rkjpm\" (UID: \"de7a95f8-3a89-4b73-b2c3-259880310e7a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037635 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9qf9f\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037655 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29f98\" (UniqueName: \"kubernetes.io/projected/208a8072-c385-4aa0-be49-2e4f915b75fa-kube-api-access-29f98\") pod \"migrator-59844c95c7-5kcn9\" (UID: \"208a8072-c385-4aa0-be49-2e4f915b75fa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037671 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-plugins-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037687 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-csi-data-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037714 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/378633a1-a84b-4040-afb9-b3a549797a86-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-5zmxd\" (UID: \"378633a1-a84b-4040-afb9-b3a549797a86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037730 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8f8ca91-000f-46a3-b4ac-b04b834530e4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8cv8g\" (UID: \"d8f8ca91-000f-46a3-b4ac-b04b834530e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037747 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/21e98fa4-f620-4124-952b-55ff3782a39a-signing-cabundle\") pod \"service-ca-9c57cc56f-crqpx\" (UID: \"21e98fa4-f620-4124-952b-55ff3782a39a\") " pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037764 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgdz6\" (UniqueName: \"kubernetes.io/projected/05ca1d75-53d9-4c65-9a78-124816a66f51-kube-api-access-sgdz6\") pod \"package-server-manager-789f6589d5-fvmbt\" (UID: \"05ca1d75-53d9-4c65-9a78-124816a66f51\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037782 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpsd9\" (UniqueName: \"kubernetes.io/projected/9bde1ae3-e13f-4020-a11d-73a1c15d1759-kube-api-access-lpsd9\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037798 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e-certs\") pod \"machine-config-server-mxdv7\" (UID: \"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e\") " pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037817 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6779444a-7630-434c-9d59-036633f57a09-webhook-cert\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037833 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb-cert\") pod \"ingress-canary-2bjgq\" (UID: \"69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb\") " pod="openshift-ingress-canary/ingress-canary-2bjgq" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037882 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/811afd11-d673-4ad8-b6ba-4d766d6affe5-default-certificate\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037899 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d4b61b96-afd0-47a6-be02-ead379d948b2-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9d8lt\" (UID: \"d4b61b96-afd0-47a6-be02-ead379d948b2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037919 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fd9f3d7-d711-43cb-a896-7c77507621bc-secret-volume\") pod \"collect-profiles-29533365-krpxk\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037934 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/21e98fa4-f620-4124-952b-55ff3782a39a-signing-key\") pod \"service-ca-9c57cc56f-crqpx\" (UID: \"21e98fa4-f620-4124-952b-55ff3782a39a\") " pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037950 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ck7d\" (UniqueName: \"kubernetes.io/projected/07b2922a-b293-4d76-8178-e73813f70ea0-kube-api-access-5ck7d\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037967 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfddj\" (UniqueName: \"kubernetes.io/projected/de7a95f8-3a89-4b73-b2c3-259880310e7a-kube-api-access-rfddj\") pod \"kube-storage-version-migrator-operator-b67b599dd-rkjpm\" (UID: \"de7a95f8-3a89-4b73-b2c3-259880310e7a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.037986 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0726188-e635-49f0-97ce-2cad95e63c13-srv-cert\") pod \"catalog-operator-68c6474976-8wxst\" (UID: \"b0726188-e635-49f0-97ce-2cad95e63c13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038005 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8f8ca91-000f-46a3-b4ac-b04b834530e4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8cv8g\" (UID: \"d8f8ca91-000f-46a3-b4ac-b04b834530e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038020 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb0ed3e8-663a-4605-9c73-78369c878589-config\") pod \"service-ca-operator-777779d784-hb5xc\" (UID: \"bb0ed3e8-663a-4605-9c73-78369c878589\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038047 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6779444a-7630-434c-9d59-036633f57a09-tmpfs\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038092 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfkpf\" (UniqueName: \"kubernetes.io/projected/378633a1-a84b-4040-afb9-b3a549797a86-kube-api-access-kfkpf\") pod \"machine-config-controller-84d6567774-5zmxd\" (UID: \"378633a1-a84b-4040-afb9-b3a549797a86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038110 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6779444a-7630-434c-9d59-036633f57a09-apiservice-cert\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038127 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-bound-sa-token\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038153 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df24m\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-kube-api-access-df24m\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038173 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-tls\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038188 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-trusted-ca\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038205 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/441c14a4-92eb-4853-a964-2a8e1bd67415-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tld2j\" (UID: \"441c14a4-92eb-4853-a964-2a8e1bd67415\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038223 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-mountpoint-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038252 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038268 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/441c14a4-92eb-4853-a964-2a8e1bd67415-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tld2j\" (UID: \"441c14a4-92eb-4853-a964-2a8e1bd67415\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038283 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/7cb83e77-c92b-4736-91f2-61c8a1241eb0-ready\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038339 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwhtv\" (UniqueName: \"kubernetes.io/projected/7cb83e77-c92b-4736-91f2-61c8a1241eb0-kube-api-access-gwhtv\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038357 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zzss\" (UniqueName: \"kubernetes.io/projected/ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e-kube-api-access-7zzss\") pod \"machine-config-server-mxdv7\" (UID: \"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e\") " pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038401 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llwfn\" (UniqueName: \"kubernetes.io/projected/ed458a40-82d4-48ce-b81c-c5a8e09e5296-kube-api-access-llwfn\") pod \"multus-admission-controller-857f4d67dd-wqj9s\" (UID: \"ed458a40-82d4-48ce-b81c-c5a8e09e5296\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038431 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ed458a40-82d4-48ce-b81c-c5a8e09e5296-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-wqj9s\" (UID: \"ed458a40-82d4-48ce-b81c-c5a8e09e5296\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038468 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07b2922a-b293-4d76-8178-e73813f70ea0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038486 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9qm5\" (UniqueName: \"kubernetes.io/projected/d4b61b96-afd0-47a6-be02-ead379d948b2-kube-api-access-z9qm5\") pod \"olm-operator-6b444d44fb-9d8lt\" (UID: \"d4b61b96-afd0-47a6-be02-ead379d948b2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038503 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038518 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-socket-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038556 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-registration-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038581 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fd9f3d7-d711-43cb-a896-7c77507621bc-config-volume\") pod \"collect-profiles-29533365-krpxk\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038596 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9bde1ae3-e13f-4020-a11d-73a1c15d1759-proxy-tls\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038633 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/05ca1d75-53d9-4c65-9a78-124816a66f51-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fvmbt\" (UID: \"05ca1d75-53d9-4c65-9a78-124816a66f51\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038651 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07b2922a-b293-4d76-8178-e73813f70ea0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038681 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/07b2922a-b293-4d76-8178-e73813f70ea0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038719 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de7a95f8-3a89-4b73-b2c3-259880310e7a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rkjpm\" (UID: \"de7a95f8-3a89-4b73-b2c3-259880310e7a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038745 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-certificates\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038763 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/811afd11-d673-4ad8-b6ba-4d766d6affe5-stats-auth\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038810 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmsdg\" (UniqueName: \"kubernetes.io/projected/6779444a-7630-434c-9d59-036633f57a09-kube-api-access-tmsdg\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038829 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038867 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0726188-e635-49f0-97ce-2cad95e63c13-profile-collector-cert\") pod \"catalog-operator-68c6474976-8wxst\" (UID: \"b0726188-e635-49f0-97ce-2cad95e63c13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038888 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4whx8\" (UniqueName: \"kubernetes.io/projected/811afd11-d673-4ad8-b6ba-4d766d6affe5-kube-api-access-4whx8\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038904 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/378633a1-a84b-4040-afb9-b3a549797a86-proxy-tls\") pod \"machine-config-controller-84d6567774-5zmxd\" (UID: \"378633a1-a84b-4040-afb9-b3a549797a86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038920 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/811afd11-d673-4ad8-b6ba-4d766d6affe5-metrics-certs\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038958 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7cb83e77-c92b-4736-91f2-61c8a1241eb0-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.038988 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0ed3e8-663a-4605-9c73-78369c878589-serving-cert\") pod \"service-ca-operator-777779d784-hb5xc\" (UID: \"bb0ed3e8-663a-4605-9c73-78369c878589\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.042934 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-ca-trust-extracted\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.043083 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-certificates\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: E0225 06:47:02.043467 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.543411294 +0000 UTC m=+115.982667743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.045181 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-trusted-ca\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.055150 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-installation-pull-secrets\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.055413 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-tls\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.069060 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df24m\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-kube-api-access-df24m\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.070918 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-bound-sa-token\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142081 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:02 crc kubenswrapper[4978]: E0225 06:47:02.142466 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.642419019 +0000 UTC m=+116.081675528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142533 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/441c14a4-92eb-4853-a964-2a8e1bd67415-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tld2j\" (UID: \"441c14a4-92eb-4853-a964-2a8e1bd67415\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142587 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/7cb83e77-c92b-4736-91f2-61c8a1241eb0-ready\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142630 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-mountpoint-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142675 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zzss\" (UniqueName: \"kubernetes.io/projected/ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e-kube-api-access-7zzss\") pod \"machine-config-server-mxdv7\" (UID: \"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e\") " pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142732 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwhtv\" (UniqueName: \"kubernetes.io/projected/7cb83e77-c92b-4736-91f2-61c8a1241eb0-kube-api-access-gwhtv\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142768 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llwfn\" (UniqueName: \"kubernetes.io/projected/ed458a40-82d4-48ce-b81c-c5a8e09e5296-kube-api-access-llwfn\") pod \"multus-admission-controller-857f4d67dd-wqj9s\" (UID: \"ed458a40-82d4-48ce-b81c-c5a8e09e5296\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142794 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ed458a40-82d4-48ce-b81c-c5a8e09e5296-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-wqj9s\" (UID: \"ed458a40-82d4-48ce-b81c-c5a8e09e5296\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142829 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07b2922a-b293-4d76-8178-e73813f70ea0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142862 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-mountpoint-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142876 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9qm5\" (UniqueName: \"kubernetes.io/projected/d4b61b96-afd0-47a6-be02-ead379d948b2-kube-api-access-z9qm5\") pod \"olm-operator-6b444d44fb-9d8lt\" (UID: \"d4b61b96-afd0-47a6-be02-ead379d948b2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142923 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-socket-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142943 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-registration-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.142990 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fd9f3d7-d711-43cb-a896-7c77507621bc-config-volume\") pod \"collect-profiles-29533365-krpxk\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143043 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9bde1ae3-e13f-4020-a11d-73a1c15d1759-proxy-tls\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143085 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/05ca1d75-53d9-4c65-9a78-124816a66f51-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fvmbt\" (UID: \"05ca1d75-53d9-4c65-9a78-124816a66f51\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143106 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07b2922a-b293-4d76-8178-e73813f70ea0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143141 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143197 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/07b2922a-b293-4d76-8178-e73813f70ea0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143224 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de7a95f8-3a89-4b73-b2c3-259880310e7a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rkjpm\" (UID: \"de7a95f8-3a89-4b73-b2c3-259880310e7a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143245 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/7cb83e77-c92b-4736-91f2-61c8a1241eb0-ready\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143270 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmsdg\" (UniqueName: \"kubernetes.io/projected/6779444a-7630-434c-9d59-036633f57a09-kube-api-access-tmsdg\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143323 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/811afd11-d673-4ad8-b6ba-4d766d6affe5-stats-auth\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143356 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143418 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0726188-e635-49f0-97ce-2cad95e63c13-profile-collector-cert\") pod \"catalog-operator-68c6474976-8wxst\" (UID: \"b0726188-e635-49f0-97ce-2cad95e63c13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143461 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4whx8\" (UniqueName: \"kubernetes.io/projected/811afd11-d673-4ad8-b6ba-4d766d6affe5-kube-api-access-4whx8\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143489 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/378633a1-a84b-4040-afb9-b3a549797a86-proxy-tls\") pod \"machine-config-controller-84d6567774-5zmxd\" (UID: \"378633a1-a84b-4040-afb9-b3a549797a86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143509 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/811afd11-d673-4ad8-b6ba-4d766d6affe5-metrics-certs\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143530 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7cb83e77-c92b-4736-91f2-61c8a1241eb0-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143531 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-registration-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143553 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0ed3e8-663a-4605-9c73-78369c878589-serving-cert\") pod \"service-ca-operator-777779d784-hb5xc\" (UID: \"bb0ed3e8-663a-4605-9c73-78369c878589\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143620 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-socket-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143672 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/441c14a4-92eb-4853-a964-2a8e1bd67415-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tld2j\" (UID: \"441c14a4-92eb-4853-a964-2a8e1bd67415\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143725 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg4ml\" (UniqueName: \"kubernetes.io/projected/20061d67-2059-4022-97ad-607c9f85540c-kube-api-access-rg4ml\") pod \"control-plane-machine-set-operator-78cbb6b69f-rwfz5\" (UID: \"20061d67-2059-4022-97ad-607c9f85540c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143770 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7cb83e77-c92b-4736-91f2-61c8a1241eb0-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143803 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9qf9f\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143827 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9bde1ae3-e13f-4020-a11d-73a1c15d1759-images\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143845 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0b5a34af-02b0-488a-8e20-2d71c48a9c5e-metrics-tls\") pod \"dns-default-dqrgj\" (UID: \"0b5a34af-02b0-488a-8e20-2d71c48a9c5e\") " pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143877 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtvtd\" (UniqueName: \"kubernetes.io/projected/0b5a34af-02b0-488a-8e20-2d71c48a9c5e-kube-api-access-xtvtd\") pod \"dns-default-dqrgj\" (UID: \"0b5a34af-02b0-488a-8e20-2d71c48a9c5e\") " pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143898 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d4b61b96-afd0-47a6-be02-ead379d948b2-srv-cert\") pod \"olm-operator-6b444d44fb-9d8lt\" (UID: \"d4b61b96-afd0-47a6-be02-ead379d948b2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143922 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7xkz\" (UniqueName: \"kubernetes.io/projected/7fd9f3d7-d711-43cb-a896-7c77507621bc-kube-api-access-n7xkz\") pod \"collect-profiles-29533365-krpxk\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143943 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6hh5\" (UniqueName: \"kubernetes.io/projected/b0726188-e635-49f0-97ce-2cad95e63c13-kube-api-access-p6hh5\") pod \"catalog-operator-68c6474976-8wxst\" (UID: \"b0726188-e635-49f0-97ce-2cad95e63c13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143970 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9bde1ae3-e13f-4020-a11d-73a1c15d1759-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.143989 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/811afd11-d673-4ad8-b6ba-4d766d6affe5-service-ca-bundle\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144016 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b5a34af-02b0-488a-8e20-2d71c48a9c5e-config-volume\") pod \"dns-default-dqrgj\" (UID: \"0b5a34af-02b0-488a-8e20-2d71c48a9c5e\") " pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144039 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtlhx\" (UniqueName: \"kubernetes.io/projected/9b843a14-9586-4f29-bfba-8413f4e52104-kube-api-access-rtlhx\") pod \"marketplace-operator-79b997595-9qf9f\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144061 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9dqg\" (UniqueName: \"kubernetes.io/projected/31337b76-7c44-42ab-a947-24c2932e2110-kube-api-access-g9dqg\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144092 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzn5t\" (UniqueName: \"kubernetes.io/projected/bb0ed3e8-663a-4605-9c73-78369c878589-kube-api-access-wzn5t\") pod \"service-ca-operator-777779d784-hb5xc\" (UID: \"bb0ed3e8-663a-4605-9c73-78369c878589\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144110 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e-node-bootstrap-token\") pod \"machine-config-server-mxdv7\" (UID: \"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e\") " pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144157 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/20061d67-2059-4022-97ad-607c9f85540c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rwfz5\" (UID: \"20061d67-2059-4022-97ad-607c9f85540c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144184 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt5bx\" (UniqueName: \"kubernetes.io/projected/21e98fa4-f620-4124-952b-55ff3782a39a-kube-api-access-qt5bx\") pod \"service-ca-9c57cc56f-crqpx\" (UID: \"21e98fa4-f620-4124-952b-55ff3782a39a\") " pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144204 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f8ca91-000f-46a3-b4ac-b04b834530e4-config\") pod \"kube-apiserver-operator-766d6c64bb-8cv8g\" (UID: \"d8f8ca91-000f-46a3-b4ac-b04b834530e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144227 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjtbj\" (UniqueName: \"kubernetes.io/projected/69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb-kube-api-access-zjtbj\") pod \"ingress-canary-2bjgq\" (UID: \"69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb\") " pod="openshift-ingress-canary/ingress-canary-2bjgq" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144259 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de7a95f8-3a89-4b73-b2c3-259880310e7a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rkjpm\" (UID: \"de7a95f8-3a89-4b73-b2c3-259880310e7a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144277 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9qf9f\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144302 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29f98\" (UniqueName: \"kubernetes.io/projected/208a8072-c385-4aa0-be49-2e4f915b75fa-kube-api-access-29f98\") pod \"migrator-59844c95c7-5kcn9\" (UID: \"208a8072-c385-4aa0-be49-2e4f915b75fa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144327 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-plugins-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144345 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-csi-data-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144384 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/378633a1-a84b-4040-afb9-b3a549797a86-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-5zmxd\" (UID: \"378633a1-a84b-4040-afb9-b3a549797a86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144410 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8f8ca91-000f-46a3-b4ac-b04b834530e4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8cv8g\" (UID: \"d8f8ca91-000f-46a3-b4ac-b04b834530e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144430 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/21e98fa4-f620-4124-952b-55ff3782a39a-signing-cabundle\") pod \"service-ca-9c57cc56f-crqpx\" (UID: \"21e98fa4-f620-4124-952b-55ff3782a39a\") " pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144456 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e-certs\") pod \"machine-config-server-mxdv7\" (UID: \"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e\") " pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144474 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgdz6\" (UniqueName: \"kubernetes.io/projected/05ca1d75-53d9-4c65-9a78-124816a66f51-kube-api-access-sgdz6\") pod \"package-server-manager-789f6589d5-fvmbt\" (UID: \"05ca1d75-53d9-4c65-9a78-124816a66f51\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144491 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpsd9\" (UniqueName: \"kubernetes.io/projected/9bde1ae3-e13f-4020-a11d-73a1c15d1759-kube-api-access-lpsd9\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144512 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6779444a-7630-434c-9d59-036633f57a09-webhook-cert\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144532 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb-cert\") pod \"ingress-canary-2bjgq\" (UID: \"69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb\") " pod="openshift-ingress-canary/ingress-canary-2bjgq" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144543 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fd9f3d7-d711-43cb-a896-7c77507621bc-config-volume\") pod \"collect-profiles-29533365-krpxk\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144555 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/811afd11-d673-4ad8-b6ba-4d766d6affe5-default-certificate\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144581 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d4b61b96-afd0-47a6-be02-ead379d948b2-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9d8lt\" (UID: \"d4b61b96-afd0-47a6-be02-ead379d948b2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.144779 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07b2922a-b293-4d76-8178-e73813f70ea0-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.146972 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-csi-data-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.147272 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/31337b76-7c44-42ab-a947-24c2932e2110-plugins-dir\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.149854 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7cb83e77-c92b-4736-91f2-61c8a1241eb0-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.151912 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.150384 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fd9f3d7-d711-43cb-a896-7c77507621bc-secret-volume\") pod \"collect-profiles-29533365-krpxk\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.153478 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/21e98fa4-f620-4124-952b-55ff3782a39a-signing-cabundle\") pod \"service-ca-9c57cc56f-crqpx\" (UID: \"21e98fa4-f620-4124-952b-55ff3782a39a\") " pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.153972 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f8ca91-000f-46a3-b4ac-b04b834530e4-config\") pod \"kube-apiserver-operator-766d6c64bb-8cv8g\" (UID: \"d8f8ca91-000f-46a3-b4ac-b04b834530e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.155119 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/21e98fa4-f620-4124-952b-55ff3782a39a-signing-key\") pod \"service-ca-9c57cc56f-crqpx\" (UID: \"21e98fa4-f620-4124-952b-55ff3782a39a\") " pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.155249 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ck7d\" (UniqueName: \"kubernetes.io/projected/07b2922a-b293-4d76-8178-e73813f70ea0-kube-api-access-5ck7d\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.155402 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfddj\" (UniqueName: \"kubernetes.io/projected/de7a95f8-3a89-4b73-b2c3-259880310e7a-kube-api-access-rfddj\") pod \"kube-storage-version-migrator-operator-b67b599dd-rkjpm\" (UID: \"de7a95f8-3a89-4b73-b2c3-259880310e7a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.155531 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0726188-e635-49f0-97ce-2cad95e63c13-srv-cert\") pod \"catalog-operator-68c6474976-8wxst\" (UID: \"b0726188-e635-49f0-97ce-2cad95e63c13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.155721 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8f8ca91-000f-46a3-b4ac-b04b834530e4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8cv8g\" (UID: \"d8f8ca91-000f-46a3-b4ac-b04b834530e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.155830 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb0ed3e8-663a-4605-9c73-78369c878589-config\") pod \"service-ca-operator-777779d784-hb5xc\" (UID: \"bb0ed3e8-663a-4605-9c73-78369c878589\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.155926 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6779444a-7630-434c-9d59-036633f57a09-tmpfs\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.156029 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfkpf\" (UniqueName: \"kubernetes.io/projected/378633a1-a84b-4040-afb9-b3a549797a86-kube-api-access-kfkpf\") pod \"machine-config-controller-84d6567774-5zmxd\" (UID: \"378633a1-a84b-4040-afb9-b3a549797a86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.156151 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6779444a-7630-434c-9d59-036633f57a09-apiservice-cert\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.156323 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/441c14a4-92eb-4853-a964-2a8e1bd67415-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tld2j\" (UID: \"441c14a4-92eb-4853-a964-2a8e1bd67415\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.156967 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d4b61b96-afd0-47a6-be02-ead379d948b2-profile-collector-cert\") pod \"olm-operator-6b444d44fb-9d8lt\" (UID: \"d4b61b96-afd0-47a6-be02-ead379d948b2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.156994 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6779444a-7630-434c-9d59-036633f57a09-tmpfs\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.157725 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb0ed3e8-663a-4605-9c73-78369c878589-serving-cert\") pod \"service-ca-operator-777779d784-hb5xc\" (UID: \"bb0ed3e8-663a-4605-9c73-78369c878589\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.157772 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ed458a40-82d4-48ce-b81c-c5a8e09e5296-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-wqj9s\" (UID: \"ed458a40-82d4-48ce-b81c-c5a8e09e5296\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.158237 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7cb83e77-c92b-4736-91f2-61c8a1241eb0-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.158426 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9bde1ae3-e13f-4020-a11d-73a1c15d1759-auth-proxy-config\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.158505 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d4b61b96-afd0-47a6-be02-ead379d948b2-srv-cert\") pod \"olm-operator-6b444d44fb-9d8lt\" (UID: \"d4b61b96-afd0-47a6-be02-ead379d948b2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.156159 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de7a95f8-3a89-4b73-b2c3-259880310e7a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rkjpm\" (UID: \"de7a95f8-3a89-4b73-b2c3-259880310e7a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.158727 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9bde1ae3-e13f-4020-a11d-73a1c15d1759-images\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.159065 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb0ed3e8-663a-4605-9c73-78369c878589-config\") pod \"service-ca-operator-777779d784-hb5xc\" (UID: \"bb0ed3e8-663a-4605-9c73-78369c878589\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.159458 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9qf9f\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.159606 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9qf9f\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.159739 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8f8ca91-000f-46a3-b4ac-b04b834530e4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8cv8g\" (UID: \"d8f8ca91-000f-46a3-b4ac-b04b834530e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.160417 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/20061d67-2059-4022-97ad-607c9f85540c-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rwfz5\" (UID: \"20061d67-2059-4022-97ad-607c9f85540c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.161656 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/441c14a4-92eb-4853-a964-2a8e1bd67415-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tld2j\" (UID: \"441c14a4-92eb-4853-a964-2a8e1bd67415\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.161784 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/811afd11-d673-4ad8-b6ba-4d766d6affe5-stats-auth\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.162355 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6779444a-7630-434c-9d59-036633f57a09-apiservice-cert\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.162620 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/811afd11-d673-4ad8-b6ba-4d766d6affe5-service-ca-bundle\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.162972 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de7a95f8-3a89-4b73-b2c3-259880310e7a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rkjpm\" (UID: \"de7a95f8-3a89-4b73-b2c3-259880310e7a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.163166 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/07b2922a-b293-4d76-8178-e73813f70ea0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.163711 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b5a34af-02b0-488a-8e20-2d71c48a9c5e-config-volume\") pod \"dns-default-dqrgj\" (UID: \"0b5a34af-02b0-488a-8e20-2d71c48a9c5e\") " pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.164130 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/441c14a4-92eb-4853-a964-2a8e1bd67415-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tld2j\" (UID: \"441c14a4-92eb-4853-a964-2a8e1bd67415\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.164963 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fd9f3d7-d711-43cb-a896-7c77507621bc-secret-volume\") pod \"collect-profiles-29533365-krpxk\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.165168 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/21e98fa4-f620-4124-952b-55ff3782a39a-signing-key\") pod \"service-ca-9c57cc56f-crqpx\" (UID: \"21e98fa4-f620-4124-952b-55ff3782a39a\") " pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.165428 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6779444a-7630-434c-9d59-036633f57a09-webhook-cert\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.165615 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/378633a1-a84b-4040-afb9-b3a549797a86-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-5zmxd\" (UID: \"378633a1-a84b-4040-afb9-b3a549797a86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.166027 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e-certs\") pod \"machine-config-server-mxdv7\" (UID: \"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e\") " pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.166031 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb-cert\") pod \"ingress-canary-2bjgq\" (UID: \"69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb\") " pod="openshift-ingress-canary/ingress-canary-2bjgq" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.166281 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.167159 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/811afd11-d673-4ad8-b6ba-4d766d6affe5-default-certificate\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.167685 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/441c14a4-92eb-4853-a964-2a8e1bd67415-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tld2j\" (UID: \"441c14a4-92eb-4853-a964-2a8e1bd67415\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.167754 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/811afd11-d673-4ad8-b6ba-4d766d6affe5-metrics-certs\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.173867 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9bde1ae3-e13f-4020-a11d-73a1c15d1759-proxy-tls\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.174159 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0b5a34af-02b0-488a-8e20-2d71c48a9c5e-metrics-tls\") pod \"dns-default-dqrgj\" (UID: \"0b5a34af-02b0-488a-8e20-2d71c48a9c5e\") " pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.174521 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e-node-bootstrap-token\") pod \"machine-config-server-mxdv7\" (UID: \"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e\") " pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.175038 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b0726188-e635-49f0-97ce-2cad95e63c13-profile-collector-cert\") pod \"catalog-operator-68c6474976-8wxst\" (UID: \"b0726188-e635-49f0-97ce-2cad95e63c13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.175168 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/05ca1d75-53d9-4c65-9a78-124816a66f51-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fvmbt\" (UID: \"05ca1d75-53d9-4c65-9a78-124816a66f51\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.175209 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b0726188-e635-49f0-97ce-2cad95e63c13-srv-cert\") pod \"catalog-operator-68c6474976-8wxst\" (UID: \"b0726188-e635-49f0-97ce-2cad95e63c13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.175527 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/378633a1-a84b-4040-afb9-b3a549797a86-proxy-tls\") pod \"machine-config-controller-84d6567774-5zmxd\" (UID: \"378633a1-a84b-4040-afb9-b3a549797a86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.183883 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.188161 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zzss\" (UniqueName: \"kubernetes.io/projected/ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e-kube-api-access-7zzss\") pod \"machine-config-server-mxdv7\" (UID: \"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e\") " pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.209258 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwhtv\" (UniqueName: \"kubernetes.io/projected/7cb83e77-c92b-4736-91f2-61c8a1241eb0-kube-api-access-gwhtv\") pod \"cni-sysctl-allowlist-ds-mhnd6\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.227329 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/07b2922a-b293-4d76-8178-e73813f70ea0-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.245548 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llwfn\" (UniqueName: \"kubernetes.io/projected/ed458a40-82d4-48ce-b81c-c5a8e09e5296-kube-api-access-llwfn\") pod \"multus-admission-controller-857f4d67dd-wqj9s\" (UID: \"ed458a40-82d4-48ce-b81c-c5a8e09e5296\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.257458 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.257512 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.257548 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:47:02 crc kubenswrapper[4978]: E0225 06:47:02.259797 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.75977355 +0000 UTC m=+116.199030189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.263406 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.266435 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.267223 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmsdg\" (UniqueName: \"kubernetes.io/projected/6779444a-7630-434c-9d59-036633f57a09-kube-api-access-tmsdg\") pod \"packageserver-d55dfcdfc-pzb7w\" (UID: \"6779444a-7630-434c-9d59-036633f57a09\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.286342 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9qm5\" (UniqueName: \"kubernetes.io/projected/d4b61b96-afd0-47a6-be02-ead379d948b2-kube-api-access-z9qm5\") pod \"olm-operator-6b444d44fb-9d8lt\" (UID: \"d4b61b96-afd0-47a6-be02-ead379d948b2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.310852 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpsd9\" (UniqueName: \"kubernetes.io/projected/9bde1ae3-e13f-4020-a11d-73a1c15d1759-kube-api-access-lpsd9\") pod \"machine-config-operator-74547568cd-kdrt4\" (UID: \"9bde1ae3-e13f-4020-a11d-73a1c15d1759\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.327407 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8f8ca91-000f-46a3-b4ac-b04b834530e4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8cv8g\" (UID: \"d8f8ca91-000f-46a3-b4ac-b04b834530e4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.353738 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtvtd\" (UniqueName: \"kubernetes.io/projected/0b5a34af-02b0-488a-8e20-2d71c48a9c5e-kube-api-access-xtvtd\") pod \"dns-default-dqrgj\" (UID: \"0b5a34af-02b0-488a-8e20-2d71c48a9c5e\") " pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.357059 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.358477 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:02 crc kubenswrapper[4978]: E0225 06:47:02.359113 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.859068994 +0000 UTC m=+116.298325453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.359224 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: E0225 06:47:02.359641 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.85963194 +0000 UTC m=+116.298888399 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.370201 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjtbj\" (UniqueName: \"kubernetes.io/projected/69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb-kube-api-access-zjtbj\") pod \"ingress-canary-2bjgq\" (UID: \"69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb\") " pod="openshift-ingress-canary/ingress-canary-2bjgq" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.380089 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.393430 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29f98\" (UniqueName: \"kubernetes.io/projected/208a8072-c385-4aa0-be49-2e4f915b75fa-kube-api-access-29f98\") pod \"migrator-59844c95c7-5kcn9\" (UID: \"208a8072-c385-4aa0-be49-2e4f915b75fa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.407042 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7xkz\" (UniqueName: \"kubernetes.io/projected/7fd9f3d7-d711-43cb-a896-7c77507621bc-kube-api-access-n7xkz\") pod \"collect-profiles-29533365-krpxk\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.409324 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.412034 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.425229 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4whx8\" (UniqueName: \"kubernetes.io/projected/811afd11-d673-4ad8-b6ba-4d766d6affe5-kube-api-access-4whx8\") pod \"router-default-5444994796-ml6mx\" (UID: \"811afd11-d673-4ad8-b6ba-4d766d6affe5\") " pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.425313 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-2bjgq" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.438509 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.454691 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt5bx\" (UniqueName: \"kubernetes.io/projected/21e98fa4-f620-4124-952b-55ff3782a39a-kube-api-access-qt5bx\") pod \"service-ca-9c57cc56f-crqpx\" (UID: \"21e98fa4-f620-4124-952b-55ff3782a39a\") " pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.460624 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:02 crc kubenswrapper[4978]: E0225 06:47:02.460779 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.960751489 +0000 UTC m=+116.400007948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.461191 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:02 crc kubenswrapper[4978]: E0225 06:47:02.461535 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:02.961527393 +0000 UTC m=+116.400783852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.466731 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.470151 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9dqg\" (UniqueName: \"kubernetes.io/projected/31337b76-7c44-42ab-a947-24c2932e2110-kube-api-access-g9dqg\") pod \"csi-hostpathplugin-6j76w\" (UID: \"31337b76-7c44-42ab-a947-24c2932e2110\") " pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.478879 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-mxdv7" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.486114 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.488268 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzn5t\" (UniqueName: \"kubernetes.io/projected/bb0ed3e8-663a-4605-9c73-78369c878589-kube-api-access-wzn5t\") pod \"service-ca-operator-777779d784-hb5xc\" (UID: \"bb0ed3e8-663a-4605-9c73-78369c878589\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.499868 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.505969 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6hh5\" (UniqueName: \"kubernetes.io/projected/b0726188-e635-49f0-97ce-2cad95e63c13-kube-api-access-p6hh5\") pod \"catalog-operator-68c6474976-8wxst\" (UID: \"b0726188-e635-49f0-97ce-2cad95e63c13\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.538001 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgdz6\" (UniqueName: \"kubernetes.io/projected/05ca1d75-53d9-4c65-9a78-124816a66f51-kube-api-access-sgdz6\") pod \"package-server-manager-789f6589d5-fvmbt\" (UID: \"05ca1d75-53d9-4c65-9a78-124816a66f51\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.539424 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" Feb 25 06:47:02 crc kubenswrapper[4978]: I0225 06:47:02.548064 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg4ml\" (UniqueName: \"kubernetes.io/projected/20061d67-2059-4022-97ad-607c9f85540c-kube-api-access-rg4ml\") pod \"control-plane-machine-set-operator-78cbb6b69f-rwfz5\" (UID: \"20061d67-2059-4022-97ad-607c9f85540c\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.135312 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.135877 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6j76w" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.136190 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.136456 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.136663 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.137445 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.138094 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.138657 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.13860973 +0000 UTC m=+117.577866359 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.139213 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.139359 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.139549 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.139832 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.145128 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.158467 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6b8950bf-8499-44e3-88e1-6ec33500c0c3-metrics-certs\") pod \"network-metrics-daemon-6gpsj\" (UID: \"6b8950bf-8499-44e3-88e1-6ec33500c0c3\") " pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.165440 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfddj\" (UniqueName: \"kubernetes.io/projected/de7a95f8-3a89-4b73-b2c3-259880310e7a-kube-api-access-rfddj\") pod \"kube-storage-version-migrator-operator-b67b599dd-rkjpm\" (UID: \"de7a95f8-3a89-4b73-b2c3-259880310e7a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.168148 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" event={"ID":"ea47c114-e38b-424a-9a46-c0c67dce2947","Type":"ContainerStarted","Data":"4d89f5ae8edbb08c46305dc007905a25201dc429eecc808d7c3fb853e5bdae25"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.173755 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtlhx\" (UniqueName: \"kubernetes.io/projected/9b843a14-9586-4f29-bfba-8413f4e52104-kube-api-access-rtlhx\") pod \"marketplace-operator-79b997595-9qf9f\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.174807 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" event={"ID":"411cf3c7-6262-4458-a3bd-245b2c002260","Type":"ContainerStarted","Data":"f0757603334712cda1e85100afc256ad3564908c9e98d386ba358c316d728704"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.180760 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" event={"ID":"b93c351c-80f1-4568-8251-14d138884828","Type":"ContainerStarted","Data":"95926123efe537b551864c8e3c35143de3a8c28a559207895337ad7492135ebf"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.184334 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfkpf\" (UniqueName: \"kubernetes.io/projected/378633a1-a84b-4040-afb9-b3a549797a86-kube-api-access-kfkpf\") pod \"machine-config-controller-84d6567774-5zmxd\" (UID: \"378633a1-a84b-4040-afb9-b3a549797a86\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.190426 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" event={"ID":"90b4f63b-4ce2-46db-8f97-ea117c763799","Type":"ContainerStarted","Data":"6970df8793844183b14b1f3ba41484005d87158d41e2f2edd6cd8a241b583501"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.192214 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" event={"ID":"5c9a6a73-8690-43ac-928f-153713a523f6","Type":"ContainerStarted","Data":"55360fbbf363b861c0edba3006b3cd27f20a5914d5ccbde6a439c732ebc87558"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.193704 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" event={"ID":"32ebcc4f-f0f3-4c1b-9efd-510363f3138a","Type":"ContainerStarted","Data":"801c59ec5b2d2ff0a6e7747361cadf61643cfd24d3a2da291a9b510e9bd52813"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.194966 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" event={"ID":"d14b0df6-0f54-47fb-8584-1dcdcd54f735","Type":"ContainerStarted","Data":"a5e82f9e9f95fd15e5787c5721d47336e3d8d1910987e38e5280458ec5abe196"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.196832 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" event={"ID":"81f10398-4c3a-4642-9e4f-df746612a9b2","Type":"ContainerStarted","Data":"3695b53a8e5987d19edff543607abef52fe7fb6651489911c6015726fbe08193"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.201073 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ck7d\" (UniqueName: \"kubernetes.io/projected/07b2922a-b293-4d76-8178-e73813f70ea0-kube-api-access-5ck7d\") pod \"cluster-image-registry-operator-dc59b4c8b-k274b\" (UID: \"07b2922a-b293-4d76-8178-e73813f70ea0\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.216267 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.219015 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rrcbj" event={"ID":"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe","Type":"ContainerStarted","Data":"e8177f6827103c753925dddaf7bdd3375f537f4a3bad5ee23d7c27df15b9af0d"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.219090 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rrcbj" event={"ID":"8ea11877-f9ce-4ea1-b885-a4da7bcb63fe","Type":"ContainerStarted","Data":"8d675d23db4e5f31c06cf2292ff5fdab408d0dac6b7be3b714eafec5ebefcc65"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.219427 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.221970 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r5cx2" event={"ID":"b51a2671-2dea-418e-8f99-512e4ba4e34c","Type":"ContainerStarted","Data":"d32c7a64124df0faedcc97b91c960c3bcc579e6f30b205831e9757413888ae7b"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.225282 4978 patch_prober.go:28] interesting pod/console-operator-58897d9998-rrcbj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.226634 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rrcbj" podUID="8ea11877-f9ce-4ea1-b885-a4da7bcb63fe" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.227579 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" event={"ID":"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f","Type":"ContainerStarted","Data":"daab048b388dc747c710f14beeb448b15a0db6cb1c8971f68bd0d70ab271324e"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.228159 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.230658 4978 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qcr2x container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.230736 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" podUID="2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.231096 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" event={"ID":"3494cb07-ccde-443f-8286-f73a5d9ab58f","Type":"ContainerStarted","Data":"869a93fafe8f80d5d0edca5760b26b81882eee0bd1ecee1359dd401042e75103"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.235937 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" event={"ID":"39e6256a-c532-4534-bcda-7531865d965c","Type":"ContainerStarted","Data":"6a66ed57f010a229c27415630ffd5919ed83ccb7e86a946bf0ab98afb3b2c234"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.238445 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" event={"ID":"3414da22-9e7f-4ae1-953e-b56ec8e78a2d","Type":"ContainerStarted","Data":"ec3ae86d2d86e194843c05a65c862d5d96a797e510faf938789bab5546a4ec06"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.239874 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.240264 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:03.740247684 +0000 UTC m=+117.179504143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.242872 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" event={"ID":"e6150f41-517c-4408-9a8e-a553c4f28ecb","Type":"ContainerStarted","Data":"796e7fb483f414483b397115f435b712dc7fd842855ac9b5fc085797cba34a06"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.243803 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.250330 4978 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-cmkt5 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.250384 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" podUID="e6150f41-517c-4408-9a8e-a553c4f28ecb" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.255089 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-jmhxz" event={"ID":"028ba3b7-0bca-4541-a5be-7f63b26bdea4","Type":"ContainerStarted","Data":"56ed3c3ac57ab8fd086d7ce13f1cf27761daf130bc147ddc51a54247d297871d"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.255904 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-jmhxz" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.258694 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" event={"ID":"1fc2a49b-47ae-4601-a397-c96d56fd10cb","Type":"ContainerStarted","Data":"243e5664c8deec6dcc8571b8e01a3fa3c6990cf4846e773e0d2879341a0a7775"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.259933 4978 patch_prober.go:28] interesting pod/downloads-7954f5f757-jmhxz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.259968 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jmhxz" podUID="028ba3b7-0bca-4541-a5be-7f63b26bdea4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.260264 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" event={"ID":"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf","Type":"ContainerStarted","Data":"5f38d8460e7c3ea7a6d69e0775735785291f7a4d874ec5fdff01a06daab70f19"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.262203 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-6gpsj" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.284696 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" event={"ID":"daa8133c-0938-4871-a573-f7025bee1103","Type":"ContainerStarted","Data":"c89bc0eec90dea59b00902a677b0d189bd4f406d247f081fe153a6e1fef545c8"} Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.300836 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.344842 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.346659 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:03.846642633 +0000 UTC m=+117.285899092 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.445768 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.447857 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:03.947839664 +0000 UTC m=+117.387096123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.448809 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.460686 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.505521 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-rrcbj" podStartSLOduration=65.505503194 podStartE2EDuration="1m5.505503194s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:03.504118492 +0000 UTC m=+116.943374961" watchObservedRunningTime="2026-02-25 06:47:03.505503194 +0000 UTC m=+116.944759653" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.552780 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.553534 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.053519011 +0000 UTC m=+117.492775460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.580232 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" podStartSLOduration=64.580213011 podStartE2EDuration="1m4.580213011s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:03.550755817 +0000 UTC m=+116.990012276" watchObservedRunningTime="2026-02-25 06:47:03.580213011 +0000 UTC m=+117.019469470" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.582130 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-wqj9s"] Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.587940 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-jmhxz" podStartSLOduration=65.587924755 podStartE2EDuration="1m5.587924755s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:03.586680447 +0000 UTC m=+117.025936906" watchObservedRunningTime="2026-02-25 06:47:03.587924755 +0000 UTC m=+117.027181214" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.648552 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" podStartSLOduration=65.648533544 podStartE2EDuration="1m5.648533544s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:03.621958687 +0000 UTC m=+117.061215156" watchObservedRunningTime="2026-02-25 06:47:03.648533544 +0000 UTC m=+117.087789993" Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.653863 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.654066 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.154037641 +0000 UTC m=+117.593294100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.654164 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.654564 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.154554817 +0000 UTC m=+117.593811276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: W0225 06:47:03.684081 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded458a40_82d4_48ce_b81c_c5a8e09e5296.slice/crio-7a9c4cda0fea9164c04724d18b1a1decf6f576c5d9f79d97b4f63f3be87ea932 WatchSource:0}: Error finding container 7a9c4cda0fea9164c04724d18b1a1decf6f576c5d9f79d97b4f63f3be87ea932: Status 404 returned error can't find the container with id 7a9c4cda0fea9164c04724d18b1a1decf6f576c5d9f79d97b4f63f3be87ea932 Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.719598 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt"] Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.755800 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.756024 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.255968915 +0000 UTC m=+117.695225374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.756118 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.756479 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.256466999 +0000 UTC m=+117.695723458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.857284 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.857447 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.357431553 +0000 UTC m=+117.796688012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.857481 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.857739 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.357732473 +0000 UTC m=+117.796988932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.930868 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dqrgj"] Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.934245 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j"] Feb 25 06:47:03 crc kubenswrapper[4978]: I0225 06:47:03.957922 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:03 crc kubenswrapper[4978]: E0225 06:47:03.958307 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.458290574 +0000 UTC m=+117.897547023 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.059577 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.060496 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.560465915 +0000 UTC m=+117.999722374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.161718 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.161918 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.661902024 +0000 UTC m=+118.101158483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.162658 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.263029 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.263464 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.763443434 +0000 UTC m=+118.202699893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.299287 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" event={"ID":"b93c351c-80f1-4568-8251-14d138884828","Type":"ContainerStarted","Data":"f65948c4e26c216eb14e12e70bc6c1ec14a6a5dda69f6e2b3f6d81667714b729"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.303332 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" event={"ID":"ea47c114-e38b-424a-9a46-c0c67dce2947","Type":"ContainerStarted","Data":"1a3a59832efba7e2862cb1518c030401ab6f8c1e4afa4455f003ca193510199d"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.313571 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" event={"ID":"411cf3c7-6262-4458-a3bd-245b2c002260","Type":"ContainerStarted","Data":"0409c7739c5def8abbb57a7c7d62de941850dcaf96ff7c227af155d6f11aeafd"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.317035 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-4zmq2" podStartSLOduration=66.317022 podStartE2EDuration="1m6.317022s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:04.314843224 +0000 UTC m=+117.754099703" watchObservedRunningTime="2026-02-25 06:47:04.317022 +0000 UTC m=+117.756278459" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.329093 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" event={"ID":"d14b0df6-0f54-47fb-8584-1dcdcd54f735","Type":"ContainerStarted","Data":"8331f72399b402b3e0367174e5698e3683b319218fade1fd5b1d25bb8b8f5cb1"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.329829 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-svrlg" podStartSLOduration=65.329804169 podStartE2EDuration="1m5.329804169s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:04.328140308 +0000 UTC m=+117.767396767" watchObservedRunningTime="2026-02-25 06:47:04.329804169 +0000 UTC m=+117.769060628" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.346017 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kgnwp" podStartSLOduration=66.34598771 podStartE2EDuration="1m6.34598771s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:04.344599847 +0000 UTC m=+117.783856346" watchObservedRunningTime="2026-02-25 06:47:04.34598771 +0000 UTC m=+117.785244189" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.357955 4978 generic.go:334] "Generic (PLEG): container finished" podID="81f10398-4c3a-4642-9e4f-df746612a9b2" containerID="93346fc571b52cda87b0b2d3ba9e2c8a57fcd32d40ec1e2f2db60c5d5f5299fa" exitCode=0 Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.358088 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" event={"ID":"81f10398-4c3a-4642-9e4f-df746612a9b2","Type":"ContainerDied","Data":"93346fc571b52cda87b0b2d3ba9e2c8a57fcd32d40ec1e2f2db60c5d5f5299fa"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.363477 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.363676 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.863654565 +0000 UTC m=+118.302911024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.364101 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.365062 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.865041738 +0000 UTC m=+118.304298197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.367362 4978 generic.go:334] "Generic (PLEG): container finished" podID="1fc2a49b-47ae-4601-a397-c96d56fd10cb" containerID="0c3b1aff86cc0e1f83cb72d342bf9f80da4aaf394e64a27d9aa5e90bc1b6c707" exitCode=0 Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.367451 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" event={"ID":"1fc2a49b-47ae-4601-a397-c96d56fd10cb","Type":"ContainerDied","Data":"0c3b1aff86cc0e1f83cb72d342bf9f80da4aaf394e64a27d9aa5e90bc1b6c707"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.369657 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" event={"ID":"3494cb07-ccde-443f-8286-f73a5d9ab58f","Type":"ContainerStarted","Data":"f26f144dd90d99146f3bc099e517057b00cf1224b1ac80420db090ef40ebea9d"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.370590 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.372097 4978 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bgbqw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.372137 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" podUID="3494cb07-ccde-443f-8286-f73a5d9ab58f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.386874 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" event={"ID":"3dc93bf1-0eef-4e20-ac7b-9ccd14459bbf","Type":"ContainerStarted","Data":"c20d81b31512b68a229138eca752de4c994753261630a12e047e574e87e26cb9"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.390354 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" event={"ID":"daa8133c-0938-4871-a573-f7025bee1103","Type":"ContainerStarted","Data":"2701f22eb728273783c09269745c414371d911f904ff450f4153a907cecd0bc6"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.392540 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r5cx2" event={"ID":"b51a2671-2dea-418e-8f99-512e4ba4e34c","Type":"ContainerStarted","Data":"b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.393673 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-mxdv7" event={"ID":"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e","Type":"ContainerStarted","Data":"13d237eb0b0be4a42a1bd1a2e3a46de078648a5474e4e3853651b04012e9582d"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.397621 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dqrgj" event={"ID":"0b5a34af-02b0-488a-8e20-2d71c48a9c5e","Type":"ContainerStarted","Data":"863aa8146e8ace5726dbfa3c8cb3fb8d11685f83fbbe593a10b99ab6dad7dd80"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.409840 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6580d4d41fe80865194e955564456b9447cfb3f51c7d75cf5d995fa264466e93"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.432872 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" podStartSLOduration=66.432853596 podStartE2EDuration="1m6.432853596s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:04.431142323 +0000 UTC m=+117.870398782" watchObservedRunningTime="2026-02-25 06:47:04.432853596 +0000 UTC m=+117.872110065" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.465437 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.472465 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:04.972426187 +0000 UTC m=+118.411682646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.483271 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-tqzfq" podStartSLOduration=66.483224054 podStartE2EDuration="1m6.483224054s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:04.470782656 +0000 UTC m=+117.910039125" watchObservedRunningTime="2026-02-25 06:47:04.483224054 +0000 UTC m=+117.922480513" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.494236 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" event={"ID":"441c14a4-92eb-4853-a964-2a8e1bd67415","Type":"ContainerStarted","Data":"ac1561db0e48d541c43ff1532d117d267a344a8660bd3037df181f4ee9e62cc5"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.507019 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-r5cx2" podStartSLOduration=66.506989055 podStartE2EDuration="1m6.506989055s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:04.49989166 +0000 UTC m=+117.939148139" watchObservedRunningTime="2026-02-25 06:47:04.506989055 +0000 UTC m=+117.946245514" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.518191 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" event={"ID":"ed458a40-82d4-48ce-b81c-c5a8e09e5296","Type":"ContainerStarted","Data":"7a9c4cda0fea9164c04724d18b1a1decf6f576c5d9f79d97b4f63f3be87ea932"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.519316 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" event={"ID":"d4b61b96-afd0-47a6-be02-ead379d948b2","Type":"ContainerStarted","Data":"8fe7f26c241f3bd3ddb05d527bd655801d4df799a7722216d24665f72817b21a"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.567778 4978 generic.go:334] "Generic (PLEG): container finished" podID="5c9a6a73-8690-43ac-928f-153713a523f6" containerID="fe2bf3cecb6979d3d56a4814f09085c7e092bd788b6eec8c52a4bec5ca89dfa3" exitCode=0 Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.568173 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" event={"ID":"5c9a6a73-8690-43ac-928f-153713a523f6","Type":"ContainerDied","Data":"fe2bf3cecb6979d3d56a4814f09085c7e092bd788b6eec8c52a4bec5ca89dfa3"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.569331 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.578651 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.078627 +0000 UTC m=+118.517883979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.590124 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" event={"ID":"7cb83e77-c92b-4736-91f2-61c8a1241eb0","Type":"ContainerStarted","Data":"f8e177876a0c2b1fc06cc59b4f7447b215d4265dc51d7be82d320a81b790fa3a"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.634751 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ml6mx" event={"ID":"811afd11-d673-4ad8-b6ba-4d766d6affe5","Type":"ContainerStarted","Data":"14c10ac57b7a349c2d158d5dfb042a4c353c39b8776605a7fd4555845701a237"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.640503 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" event={"ID":"32ebcc4f-f0f3-4c1b-9efd-510363f3138a","Type":"ContainerStarted","Data":"41667a1721bd10f348b8ff251a6487c9167a005445012a57472104e65557e981"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.643918 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" event={"ID":"90b4f63b-4ce2-46db-8f97-ea117c763799","Type":"ContainerStarted","Data":"3f562506b820385d01159f320bf71bfbb31e6f71fd44b5adc6098ef2ce3fc7bf"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.658892 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" event={"ID":"39e6256a-c532-4534-bcda-7531865d965c","Type":"ContainerStarted","Data":"a5240c01a8d0d43de91b2c8c4e4d8aeac41132f8945c7f9f9e2313784e95f150"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.671149 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.678596 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.178574692 +0000 UTC m=+118.617831151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.693254 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-ml6mx" podStartSLOduration=66.693237218 podStartE2EDuration="1m6.693237218s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:04.663620408 +0000 UTC m=+118.102876867" watchObservedRunningTime="2026-02-25 06:47:04.693237218 +0000 UTC m=+118.132493667" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.693947 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" event={"ID":"3414da22-9e7f-4ae1-953e-b56ec8e78a2d","Type":"ContainerStarted","Data":"ed9b2e066129ee605527af7f784ddd7c88e0cdad7c69c71e94cbd9f591fecab5"} Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.694689 4978 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qcr2x container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.694755 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" podUID="2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.695509 4978 patch_prober.go:28] interesting pod/downloads-7954f5f757-jmhxz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.695552 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jmhxz" podUID="028ba3b7-0bca-4541-a5be-7f63b26bdea4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.695887 4978 patch_prober.go:28] interesting pod/console-operator-58897d9998-rrcbj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.696115 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rrcbj" podUID="8ea11877-f9ce-4ea1-b885-a4da7bcb63fe" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.699769 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-2bjgq"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.699818 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.707154 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-nhkg4" podStartSLOduration=66.70714446 podStartE2EDuration="1m6.70714446s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:04.7058625 +0000 UTC m=+118.145118959" watchObservedRunningTime="2026-02-25 06:47:04.70714446 +0000 UTC m=+118.146400919" Feb 25 06:47:04 crc kubenswrapper[4978]: W0225 06:47:04.724207 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20061d67_2059_4022_97ad_607c9f85540c.slice/crio-cf9f3139549dcd66bec98c1f0737afb78860807679afd9719d4b5629642593bd WatchSource:0}: Error finding container cf9f3139549dcd66bec98c1f0737afb78860807679afd9719d4b5629642593bd: Status 404 returned error can't find the container with id cf9f3139549dcd66bec98c1f0737afb78860807679afd9719d4b5629642593bd Feb 25 06:47:04 crc kubenswrapper[4978]: W0225 06:47:04.726443 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69bc42e9_8bb2_4dfa_b91b_0d0f2a72addb.slice/crio-8266aee39616d3928d9552f244f6c11eb668319aded0c67f908df5a795033823 WatchSource:0}: Error finding container 8266aee39616d3928d9552f244f6c11eb668319aded0c67f908df5a795033823: Status 404 returned error can't find the container with id 8266aee39616d3928d9552f244f6c11eb668319aded0c67f908df5a795033823 Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.732118 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2fbhg" podStartSLOduration=66.732100647 podStartE2EDuration="1m6.732100647s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:04.7315682 +0000 UTC m=+118.170824649" watchObservedRunningTime="2026-02-25 06:47:04.732100647 +0000 UTC m=+118.171357106" Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.772047 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.773769 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.273757311 +0000 UTC m=+118.713013770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.842754 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qf9f"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.873258 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.874776 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.374759246 +0000 UTC m=+118.814015705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.889540 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w"] Feb 25 06:47:04 crc kubenswrapper[4978]: W0225 06:47:04.898028 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b843a14_9586_4f29_bfba_8413f4e52104.slice/crio-826f5f8329047bb1483d89d2a9a156c5e63fb37b32383340c4781bae3e34d2a8 WatchSource:0}: Error finding container 826f5f8329047bb1483d89d2a9a156c5e63fb37b32383340c4781bae3e34d2a8: Status 404 returned error can't find the container with id 826f5f8329047bb1483d89d2a9a156c5e63fb37b32383340c4781bae3e34d2a8 Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.914401 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.953818 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.962769 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.966613 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-6gpsj"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.968841 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.970960 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.975038 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:04 crc kubenswrapper[4978]: E0225 06:47:04.975705 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.475674638 +0000 UTC m=+118.914931137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.978077 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-crqpx"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.981232 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4"] Feb 25 06:47:04 crc kubenswrapper[4978]: W0225 06:47:04.984110 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fd9f3d7_d711_43cb_a896_7c77507621bc.slice/crio-dcb3ee7dbd229617a864f6eeb76da93e8cb1ec91eec81d655cb966eee9feaebb WatchSource:0}: Error finding container dcb3ee7dbd229617a864f6eeb76da93e8cb1ec91eec81d655cb966eee9feaebb: Status 404 returned error can't find the container with id dcb3ee7dbd229617a864f6eeb76da93e8cb1ec91eec81d655cb966eee9feaebb Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.986842 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.990009 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6j76w"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.993050 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9"] Feb 25 06:47:04 crc kubenswrapper[4978]: I0225 06:47:04.993118 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt"] Feb 25 06:47:05 crc kubenswrapper[4978]: W0225 06:47:05.007639 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod208a8072_c385_4aa0_be49_2e4f915b75fa.slice/crio-94002cb3ecc95d71a33ee4db06a49c52d0d1644c078f6dca8a6fc9405590f3ba WatchSource:0}: Error finding container 94002cb3ecc95d71a33ee4db06a49c52d0d1644c078f6dca8a6fc9405590f3ba: Status 404 returned error can't find the container with id 94002cb3ecc95d71a33ee4db06a49c52d0d1644c078f6dca8a6fc9405590f3ba Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.054527 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.076220 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.076858 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.576836388 +0000 UTC m=+119.016092847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.147779 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.163005 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.163068 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.178571 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.179425 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.679362509 +0000 UTC m=+119.118618968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.281088 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.281208 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.781180159 +0000 UTC m=+119.220436628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.281784 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.282078 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.782064856 +0000 UTC m=+119.221321315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.382845 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.383238 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.883220065 +0000 UTC m=+119.322476524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.485114 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.485514 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:05.98549709 +0000 UTC m=+119.424753549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.585931 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.586122 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.086095153 +0000 UTC m=+119.525351612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.586576 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.586910 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.086898347 +0000 UTC m=+119.526154806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.694790 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.695197 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.195149282 +0000 UTC m=+119.634405741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.706315 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" event={"ID":"21e98fa4-f620-4124-952b-55ff3782a39a","Type":"ContainerStarted","Data":"d189abd910bd024ba3f556f6c8b6cb8087b0629705b1dc0da92122026a228214"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.708688 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" event={"ID":"81f10398-4c3a-4642-9e4f-df746612a9b2","Type":"ContainerStarted","Data":"0525b95aee73f020990148db48661354a4751e1f51778f55f4d1a4901931ab99"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.710107 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-ml6mx" event={"ID":"811afd11-d673-4ad8-b6ba-4d766d6affe5","Type":"ContainerStarted","Data":"fe1a3c513af29ef05e7b4b35eb2cf2268cd4d677f03ab7a957ddd6fd16237c8e"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.716180 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" event={"ID":"07b2922a-b293-4d76-8178-e73813f70ea0","Type":"ContainerStarted","Data":"e7ced0cfc2d9ebab835b43e6303d9d7f2b9ece59520f5a5a70c60b9c27629486"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.717844 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"a74c0a4152ba3e7037480189833382d2e84d0954b2c9089cf155392cb975d6aa"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.718943 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9" event={"ID":"208a8072-c385-4aa0-be49-2e4f915b75fa","Type":"ContainerStarted","Data":"94002cb3ecc95d71a33ee4db06a49c52d0d1644c078f6dca8a6fc9405590f3ba"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.722573 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2bjgq" event={"ID":"69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb","Type":"ContainerStarted","Data":"34072daa997304b2eba8f3f28cbd8b2300e84ecca0d0034a9af8d859ba9681a0"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.722629 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-2bjgq" event={"ID":"69bc42e9-8bb2-4dfa-b91b-0d0f2a72addb","Type":"ContainerStarted","Data":"8266aee39616d3928d9552f244f6c11eb668319aded0c67f908df5a795033823"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.725174 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" event={"ID":"bb0ed3e8-663a-4605-9c73-78369c878589","Type":"ContainerStarted","Data":"5926be50339e50e4cd0ceed04216d1b1bd65b35a4bd09eacbdcd9a78164ece60"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.727235 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-mxdv7" event={"ID":"ea45d21b-6a71-4a4a-a21f-3f0050a5fe6e","Type":"ContainerStarted","Data":"961a04d3fa96ae9780755a4424b74d63a3194e4e83ff8890581e51f7243ee4fa"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.727897 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" podStartSLOduration=66.727884905 podStartE2EDuration="1m6.727884905s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:05.727483693 +0000 UTC m=+119.166740162" watchObservedRunningTime="2026-02-25 06:47:05.727884905 +0000 UTC m=+119.167141364" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.750129 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-mxdv7" podStartSLOduration=7.750112239 podStartE2EDuration="7.750112239s" podCreationTimestamp="2026-02-25 06:46:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:05.747520151 +0000 UTC m=+119.186776610" watchObservedRunningTime="2026-02-25 06:47:05.750112239 +0000 UTC m=+119.189368688" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.751021 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" event={"ID":"1fc2a49b-47ae-4601-a397-c96d56fd10cb","Type":"ContainerStarted","Data":"bf8131051f594f3633d780ad87a8800036d6e65ec665661f76ad43d70839529f"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.751158 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.753913 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" event={"ID":"daa8133c-0938-4871-a573-f7025bee1103","Type":"ContainerStarted","Data":"82580e302d864fabeea199fd659225bf518b81555208bc53446564029b95028a"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.755907 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dqrgj" event={"ID":"0b5a34af-02b0-488a-8e20-2d71c48a9c5e","Type":"ContainerStarted","Data":"1c34f21d38c6d3a97e92ba2205057828ca6c975548cd3909296479ccda7d2a68"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.757153 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" event={"ID":"d4b61b96-afd0-47a6-be02-ead379d948b2","Type":"ContainerStarted","Data":"d1f74ccb5d08a120911df7fe0776139ca762f28d6ab4b4e2960b266690b37d7a"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.757406 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.758179 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6j76w" event={"ID":"31337b76-7c44-42ab-a947-24c2932e2110","Type":"ContainerStarted","Data":"dd1c32d522ddec1dd51b93f7580979d630093989760d0d0a5922f8da4ea231aa"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.758645 4978 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-9d8lt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.758704 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" podUID="d4b61b96-afd0-47a6-be02-ead379d948b2" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.767593 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-2bjgq" podStartSLOduration=7.76757334 podStartE2EDuration="7.76757334s" podCreationTimestamp="2026-02-25 06:46:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:05.761631359 +0000 UTC m=+119.200887838" watchObservedRunningTime="2026-02-25 06:47:05.76757334 +0000 UTC m=+119.206829799" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.783822 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.784102 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.790982 4978 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-g9xwf container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.791480 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" podUID="81f10398-4c3a-4642-9e4f-df746612a9b2" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.794818 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6gpsj" event={"ID":"6b8950bf-8499-44e3-88e1-6ec33500c0c3","Type":"ContainerStarted","Data":"e4ec42f91fc50b37b18bc9bfd202d273ab7b6d81b7a87a31db8ee9bc3edd44ba"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.797557 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.798241 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-vc5bs" podStartSLOduration=67.79822425 podStartE2EDuration="1m7.79822425s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:05.796277211 +0000 UTC m=+119.235533670" watchObservedRunningTime="2026-02-25 06:47:05.79822425 +0000 UTC m=+119.237480699" Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.799472 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.299456417 +0000 UTC m=+119.738712866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.807576 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" event={"ID":"3414da22-9e7f-4ae1-953e-b56ec8e78a2d","Type":"ContainerStarted","Data":"f7289aeab8b79eecf56fe06039fcd84b784baaee4feae4282f628782a1b2074f"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.817182 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" podStartSLOduration=66.817163394 podStartE2EDuration="1m6.817163394s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:05.816735841 +0000 UTC m=+119.255992300" watchObservedRunningTime="2026-02-25 06:47:05.817163394 +0000 UTC m=+119.256419853" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.835250 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" event={"ID":"9bde1ae3-e13f-4020-a11d-73a1c15d1759","Type":"ContainerStarted","Data":"c67db7c979aee62648ca654457eb5bc2c753aaa308c04252394164d8e0676342"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.835296 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" event={"ID":"9bde1ae3-e13f-4020-a11d-73a1c15d1759","Type":"ContainerStarted","Data":"40d9a8e80be09bab6d9506c613af13a6d42b91a5228d795fe3461214a39a143a"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.848198 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" event={"ID":"b0726188-e635-49f0-97ce-2cad95e63c13","Type":"ContainerStarted","Data":"922c6a1d9afabf2fb5af3d838531c678a78cae967df139ff1096f6861684baa7"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.853325 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" podStartSLOduration=67.853308331 podStartE2EDuration="1m7.853308331s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:05.850241308 +0000 UTC m=+119.289497767" watchObservedRunningTime="2026-02-25 06:47:05.853308331 +0000 UTC m=+119.292564780" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.872143 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-grzdq" podStartSLOduration=67.872128443 podStartE2EDuration="1m7.872128443s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:05.867180602 +0000 UTC m=+119.306437061" watchObservedRunningTime="2026-02-25 06:47:05.872128443 +0000 UTC m=+119.311384902" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.874834 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" event={"ID":"7cb83e77-c92b-4736-91f2-61c8a1241eb0","Type":"ContainerStarted","Data":"95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.875679 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.883293 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" event={"ID":"6779444a-7630-434c-9d59-036633f57a09","Type":"ContainerStarted","Data":"2d69ae7a5ca7a2f7da3c4e3cf996e221c69c55fe8bd99a84c0cfe24a993c4909"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.884505 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.885819 4978 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pzb7w container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.885960 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" podUID="6779444a-7630-434c-9d59-036633f57a09" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.898127 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" podStartSLOduration=7.89811024 podStartE2EDuration="7.89811024s" podCreationTimestamp="2026-02-25 06:46:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:05.894277985 +0000 UTC m=+119.333534444" watchObservedRunningTime="2026-02-25 06:47:05.89811024 +0000 UTC m=+119.337366699" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.898856 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.899192 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.399164603 +0000 UTC m=+119.838421082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.899660 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:05 crc kubenswrapper[4978]: E0225 06:47:05.900729 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.40071737 +0000 UTC m=+119.839973909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.915069 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" event={"ID":"ed458a40-82d4-48ce-b81c-c5a8e09e5296","Type":"ContainerStarted","Data":"13fd67ce1b81937bfed0c29bc6fe4b2210127725ce1ccf834051318ef51ee962"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.936818 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" podStartSLOduration=66.936793835 podStartE2EDuration="1m6.936793835s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:05.936659011 +0000 UTC m=+119.375915470" watchObservedRunningTime="2026-02-25 06:47:05.936793835 +0000 UTC m=+119.376050294" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.940888 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" event={"ID":"5c9a6a73-8690-43ac-928f-153713a523f6","Type":"ContainerStarted","Data":"89a16a8f805b4324ca3094d363dd5e41f39499a49d74fa700bee4ba195ef41af"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.962836 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" event={"ID":"20061d67-2059-4022-97ad-607c9f85540c","Type":"ContainerStarted","Data":"d605a6cb928c4d0eb9235e1561f9d96f506cc723d9589f9d9f3e3b1cb59c7a90"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.963335 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" event={"ID":"20061d67-2059-4022-97ad-607c9f85540c","Type":"ContainerStarted","Data":"cf9f3139549dcd66bec98c1f0737afb78860807679afd9719d4b5629642593bd"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.967976 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8dd5227782c84dd2cc5ab8b175fbd8fb2e4f97dd1adca56f96da9d5df43dc5e4"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.968045 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0d401eed15bb3c305a10838bc8d099c6f716a5711769ed02bfbba6fbe3133a7c"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.968624 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.969842 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" event={"ID":"7fd9f3d7-d711-43cb-a896-7c77507621bc","Type":"ContainerStarted","Data":"dcb3ee7dbd229617a864f6eeb76da93e8cb1ec91eec81d655cb966eee9feaebb"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.986801 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" event={"ID":"d8f8ca91-000f-46a3-b4ac-b04b834530e4","Type":"ContainerStarted","Data":"9c35c2946f94834efd4e7b217cb8d410d120336d252ad3259489ca9bcf9bc514"} Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.995320 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:47:05 crc kubenswrapper[4978]: I0225 06:47:05.999855 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" event={"ID":"39e6256a-c532-4534-bcda-7531865d965c","Type":"ContainerStarted","Data":"93bd1abfb65e0ff77566baccb7ed54031ff30142e4d943f0f9c4e1cbb56eaa08"} Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.001969 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.003004 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.502986004 +0000 UTC m=+119.942242453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.005798 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" event={"ID":"05ca1d75-53d9-4c65-9a78-124816a66f51","Type":"ContainerStarted","Data":"a3d3f0d35c96d317a2dec5a4a5d5358e86d21d20a6469abd13ba882bbdeec9e8"} Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.009055 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rwfz5" podStartSLOduration=67.009034007 podStartE2EDuration="1m7.009034007s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:05.981982256 +0000 UTC m=+119.421238715" watchObservedRunningTime="2026-02-25 06:47:06.009034007 +0000 UTC m=+119.448290476" Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.010007 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" event={"ID":"de7a95f8-3a89-4b73-b2c3-259880310e7a","Type":"ContainerStarted","Data":"ff77f19268de325e3968215532fe768763b2599596f859176ecb2e6d4cc75fd3"} Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.010056 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" event={"ID":"de7a95f8-3a89-4b73-b2c3-259880310e7a","Type":"ContainerStarted","Data":"1d743a828759d4765fe8c8dc275e599dbfc9112385795f08a9adcc906dd01b68"} Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.018419 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2d85654e5bc1058ad14b7629930c4e7a0989c43ca8cad98ef1aa9071563e98ad"} Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.018471 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e926e97f4049293b7279777800258e795f5984aa67194085ca33d4c27fa7c2a1"} Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.034158 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" podStartSLOduration=68.034130169 podStartE2EDuration="1m8.034130169s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:06.030815308 +0000 UTC m=+119.470071767" watchObservedRunningTime="2026-02-25 06:47:06.034130169 +0000 UTC m=+119.473386628" Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.036804 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" event={"ID":"378633a1-a84b-4040-afb9-b3a549797a86","Type":"ContainerStarted","Data":"eeb13892a0688c9cc79f9249d221b1787f8f0e161a73f47e0df44897015a9d2a"} Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.044739 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" event={"ID":"441c14a4-92eb-4853-a964-2a8e1bd67415","Type":"ContainerStarted","Data":"30d598a6fb80847a9b025e21b8f560275063a12beb2ee514a49c450a3bd28e36"} Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.050694 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ncnfh" podStartSLOduration=68.050673071 podStartE2EDuration="1m8.050673071s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:06.047847705 +0000 UTC m=+119.487104164" watchObservedRunningTime="2026-02-25 06:47:06.050673071 +0000 UTC m=+119.489929530" Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.051578 4978 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bgbqw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.051618 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" podUID="3494cb07-ccde-443f-8286-f73a5d9ab58f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.051660 4978 patch_prober.go:28] interesting pod/downloads-7954f5f757-jmhxz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.051720 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jmhxz" podUID="028ba3b7-0bca-4541-a5be-7f63b26bdea4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.051898 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" event={"ID":"9b843a14-9586-4f29-bfba-8413f4e52104","Type":"ContainerStarted","Data":"c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c"} Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.051926 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" event={"ID":"9b843a14-9586-4f29-bfba-8413f4e52104","Type":"ContainerStarted","Data":"826f5f8329047bb1483d89d2a9a156c5e63fb37b32383340c4781bae3e34d2a8"} Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.103104 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rkjpm" podStartSLOduration=67.103081741 podStartE2EDuration="1m7.103081741s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:06.102138443 +0000 UTC m=+119.541394902" watchObservedRunningTime="2026-02-25 06:47:06.103081741 +0000 UTC m=+119.542338190" Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.103538 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.105213 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.605198505 +0000 UTC m=+120.044454964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.146443 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.146501 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.149791 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-gkwg7" podStartSLOduration=67.149770207 podStartE2EDuration="1m7.149770207s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:06.146928142 +0000 UTC m=+119.586184611" watchObservedRunningTime="2026-02-25 06:47:06.149770207 +0000 UTC m=+119.589026666" Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.174946 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" podStartSLOduration=67.174926362 podStartE2EDuration="1m7.174926362s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:06.166040252 +0000 UTC m=+119.605296711" watchObservedRunningTime="2026-02-25 06:47:06.174926362 +0000 UTC m=+119.614182841" Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.206857 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.208044 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.707989995 +0000 UTC m=+120.147246474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.311307 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.311699 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.811686512 +0000 UTC m=+120.250942971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.339329 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tld2j" podStartSLOduration=67.3393142 podStartE2EDuration="1m7.3393142s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:06.202702875 +0000 UTC m=+119.641959344" watchObservedRunningTime="2026-02-25 06:47:06.3393142 +0000 UTC m=+119.778570659" Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.340079 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.413427 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.413929 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.913905813 +0000 UTC m=+120.353162282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.414174 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.414833 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:06.914822421 +0000 UTC m=+120.354078880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.515857 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.516215 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.016173187 +0000 UTC m=+120.455429646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.518756 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.519250 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.01923135 +0000 UTC m=+120.458487809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.620102 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.620419 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.120361828 +0000 UTC m=+120.559618287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.620832 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.621254 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.121242185 +0000 UTC m=+120.560498644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.722147 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.722700 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.222596231 +0000 UTC m=+120.661852690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.735785 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-mhnd6"] Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.823582 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.824451 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.324421232 +0000 UTC m=+120.763677691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.925533 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.925861 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.425810078 +0000 UTC m=+120.865066547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:06 crc kubenswrapper[4978]: I0225 06:47:06.926196 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:06 crc kubenswrapper[4978]: E0225 06:47:06.926809 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.426790898 +0000 UTC m=+120.866047357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.027972 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.028325 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.528308379 +0000 UTC m=+120.967564838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.060717 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" event={"ID":"5c9a6a73-8690-43ac-928f-153713a523f6","Type":"ContainerStarted","Data":"bedf3fd9048b3875023b0d7a4519b6d84a0de5a8e8456b35b9d19b1c68fd06a1"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.063757 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" event={"ID":"9bde1ae3-e13f-4020-a11d-73a1c15d1759","Type":"ContainerStarted","Data":"53c95483fbbd2643084825d92077516efe1bc19a1eb8663e94c32731205b9741"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.065620 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" event={"ID":"21e98fa4-f620-4124-952b-55ff3782a39a","Type":"ContainerStarted","Data":"26b6f5181edf7d68b03df596aa01a86aff187433a9cd68189266784eacc6d795"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.068173 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" event={"ID":"bb0ed3e8-663a-4605-9c73-78369c878589","Type":"ContainerStarted","Data":"565b57aab09789c7357da81bd28671e9a6cdf4c7c98d21bb8d74e534ac4af21b"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.070012 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" event={"ID":"ed458a40-82d4-48ce-b81c-c5a8e09e5296","Type":"ContainerStarted","Data":"deb7e9fe08f4d320aefb8d20fe903d93c443a7de3a50a9c7d97f5e188c7bf985"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.071596 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9" event={"ID":"208a8072-c385-4aa0-be49-2e4f915b75fa","Type":"ContainerStarted","Data":"1369ac8fa0abe7d56fda09e4bda518fa1345eec513ae8b80795a9310fef276c7"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.071712 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9" event={"ID":"208a8072-c385-4aa0-be49-2e4f915b75fa","Type":"ContainerStarted","Data":"6bd47e726a2aa6367cb9c9145db5767cb886d2ac03209ae0dc4ffcccb75e9d5c"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.073809 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" event={"ID":"378633a1-a84b-4040-afb9-b3a549797a86","Type":"ContainerStarted","Data":"b50079d12f78d1b88a158bab0ff6679705fc83f312dbe2689a7518511d95a728"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.073892 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" event={"ID":"378633a1-a84b-4040-afb9-b3a549797a86","Type":"ContainerStarted","Data":"907814f636c937cdf6deec9f92dc911ea8c0dadb08bb9e99a90a12e80311c7e1"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.074952 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" event={"ID":"07b2922a-b293-4d76-8178-e73813f70ea0","Type":"ContainerStarted","Data":"cfe157f9f58918b366a86daa9776ba0c6ee1d8d2cf7f085cc5d7cf31f3020051"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.076548 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" event={"ID":"d8f8ca91-000f-46a3-b4ac-b04b834530e4","Type":"ContainerStarted","Data":"5536760200bc71ae03697c688068e679b0e51eac6a3df350ab5482cbf5ca89e1"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.078833 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" event={"ID":"7fd9f3d7-d711-43cb-a896-7c77507621bc","Type":"ContainerStarted","Data":"8a9a06d5f426c85882f2a71024311846c9ece01c436f2a16be0cc489eade32d7"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.080895 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" event={"ID":"05ca1d75-53d9-4c65-9a78-124816a66f51","Type":"ContainerStarted","Data":"3ac7dc82c509f0f73275c0dcd4ee514bcdbfcb2a0ef1a2f6a959cbb9f6caa38d"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.081453 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.081587 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" event={"ID":"05ca1d75-53d9-4c65-9a78-124816a66f51","Type":"ContainerStarted","Data":"803b8e6ee20306cfcc73fa32d63ea17741de85ce61b10406954d24884a212905"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.083695 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dqrgj" event={"ID":"0b5a34af-02b0-488a-8e20-2d71c48a9c5e","Type":"ContainerStarted","Data":"a8d927b6c980668b2e59a63a37fbd6d9af741a07f616670d4ae32c9a88782dfc"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.083910 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.085582 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" event={"ID":"b0726188-e635-49f0-97ce-2cad95e63c13","Type":"ContainerStarted","Data":"db3f9564d38889aa08416bedc065f16acb7fa1dd2880f46eea02335ecae8a6fe"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.085801 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.088644 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6gpsj" event={"ID":"6b8950bf-8499-44e3-88e1-6ec33500c0c3","Type":"ContainerStarted","Data":"7cdf6bd731753bfe39e4fa0d7dd897e41cb1bca2b5271aa8b5ddd786468817b2"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.088680 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-6gpsj" event={"ID":"6b8950bf-8499-44e3-88e1-6ec33500c0c3","Type":"ContainerStarted","Data":"ce1d2e12761d56db87c66591a0a92ba2769f379b555ba71cae3f452703aa0ee3"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.089457 4978 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-8wxst container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.089495 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" podUID="b0726188-e635-49f0-97ce-2cad95e63c13" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.091503 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" event={"ID":"6779444a-7630-434c-9d59-036633f57a09","Type":"ContainerStarted","Data":"e1216bbe14fa40b60bd55813d03e4b58cd6cfd3a3a355c5dda7bf3af8aaa609e"} Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.092087 4978 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bgbqw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.092151 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" podUID="3494cb07-ccde-443f-8286-f73a5d9ab58f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.092105 4978 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-9d8lt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.092218 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" podUID="d4b61b96-afd0-47a6-be02-ead379d948b2" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.092341 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.092758 4978 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pzb7w container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.092872 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" podUID="6779444a-7630-434c-9d59-036633f57a09" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.095421 4978 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9qf9f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.095472 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" podUID="9b843a14-9586-4f29-bfba-8413f4e52104" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.108980 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" podStartSLOduration=69.108962646 podStartE2EDuration="1m9.108962646s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.10516675 +0000 UTC m=+120.544423229" watchObservedRunningTime="2026-02-25 06:47:07.108962646 +0000 UTC m=+120.548219105" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.129087 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.129267 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=1.129252222 podStartE2EDuration="1.129252222s" podCreationTimestamp="2026-02-25 06:47:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.127928611 +0000 UTC m=+120.567185070" watchObservedRunningTime="2026-02-25 06:47:07.129252222 +0000 UTC m=+120.568508671" Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.129405 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.629393666 +0000 UTC m=+121.068650125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.141218 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.141304 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.162567 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" podStartSLOduration=68.162545272 podStartE2EDuration="1m8.162545272s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.15950128 +0000 UTC m=+120.598757739" watchObservedRunningTime="2026-02-25 06:47:07.162545272 +0000 UTC m=+120.601801731" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.183354 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-crqpx" podStartSLOduration=68.183337133 podStartE2EDuration="1m8.183337133s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.182905879 +0000 UTC m=+120.622162358" watchObservedRunningTime="2026-02-25 06:47:07.183337133 +0000 UTC m=+120.622593582" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.230511 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.231904 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.731878146 +0000 UTC m=+121.171134605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.280445 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" podStartSLOduration=68.28042862 podStartE2EDuration="1m8.28042862s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.217611123 +0000 UTC m=+120.656867592" watchObservedRunningTime="2026-02-25 06:47:07.28042862 +0000 UTC m=+120.719685079" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.282429 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zmxd" podStartSLOduration=68.28242136 podStartE2EDuration="1m8.28242136s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.279575093 +0000 UTC m=+120.718831562" watchObservedRunningTime="2026-02-25 06:47:07.28242136 +0000 UTC m=+120.721677819" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.335386 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.335729 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.835715197 +0000 UTC m=+121.274971656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.344197 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-6gpsj" podStartSLOduration=69.344182194 podStartE2EDuration="1m9.344182194s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.33483716 +0000 UTC m=+120.774093619" watchObservedRunningTime="2026-02-25 06:47:07.344182194 +0000 UTC m=+120.783438643" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.420199 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-wqj9s" podStartSLOduration=68.420141889 podStartE2EDuration="1m8.420141889s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.41455662 +0000 UTC m=+120.853813099" watchObservedRunningTime="2026-02-25 06:47:07.420141889 +0000 UTC m=+120.859398368" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.438430 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.438579 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.938556788 +0000 UTC m=+121.377813257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.439133 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.439752 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:07.939723284 +0000 UTC m=+121.378979743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.488907 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dqrgj" podStartSLOduration=9.488886485 podStartE2EDuration="9.488886485s" podCreationTimestamp="2026-02-25 06:46:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.477813349 +0000 UTC m=+120.917069838" watchObservedRunningTime="2026-02-25 06:47:07.488886485 +0000 UTC m=+120.928142944" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.498787 4978 ???:1] "http: TLS handshake error from 192.168.126.11:41658: no serving certificate available for the kubelet" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.528280 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-k274b" podStartSLOduration=69.52825934 podStartE2EDuration="1m9.52825934s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.526794086 +0000 UTC m=+120.966050535" watchObservedRunningTime="2026-02-25 06:47:07.52825934 +0000 UTC m=+120.967515799" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.541950 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.542300 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.042280375 +0000 UTC m=+121.481536834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.574595 4978 ???:1] "http: TLS handshake error from 192.168.126.11:41668: no serving certificate available for the kubelet" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.597722 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hb5xc" podStartSLOduration=68.597705527 podStartE2EDuration="1m8.597705527s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.561672004 +0000 UTC m=+121.000928453" watchObservedRunningTime="2026-02-25 06:47:07.597705527 +0000 UTC m=+121.036961986" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.597974 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8cv8g" podStartSLOduration=68.597970385 podStartE2EDuration="1m8.597970385s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.59451856 +0000 UTC m=+121.033775019" watchObservedRunningTime="2026-02-25 06:47:07.597970385 +0000 UTC m=+121.037226844" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.623573 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-kdrt4" podStartSLOduration=68.623543572 podStartE2EDuration="1m8.623543572s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.622337754 +0000 UTC m=+121.061594243" watchObservedRunningTime="2026-02-25 06:47:07.623543572 +0000 UTC m=+121.062800051" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.643398 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.643759 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.143746325 +0000 UTC m=+121.583002784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.650721 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-5kcn9" podStartSLOduration=68.650685645 podStartE2EDuration="1m8.650685645s" podCreationTimestamp="2026-02-25 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:07.649786258 +0000 UTC m=+121.089042707" watchObservedRunningTime="2026-02-25 06:47:07.650685645 +0000 UTC m=+121.089942104" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.691689 4978 ???:1] "http: TLS handshake error from 192.168.126.11:41684: no serving certificate available for the kubelet" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.744632 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.745533 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.245515273 +0000 UTC m=+121.684771732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.793909 4978 ???:1] "http: TLS handshake error from 192.168.126.11:41690: no serving certificate available for the kubelet" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.847256 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.847892 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.347868089 +0000 UTC m=+121.787124548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.891342 4978 ???:1] "http: TLS handshake error from 192.168.126.11:41704: no serving certificate available for the kubelet" Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.948835 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.949118 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.449078781 +0000 UTC m=+121.888335240 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.949575 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:07 crc kubenswrapper[4978]: E0225 06:47:07.949923 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.449908975 +0000 UTC m=+121.889165434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:07 crc kubenswrapper[4978]: I0225 06:47:07.994963 4978 ???:1] "http: TLS handshake error from 192.168.126.11:41720: no serving certificate available for the kubelet" Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.050727 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.050937 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.55090978 +0000 UTC m=+121.990166239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.051049 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.051406 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.551396485 +0000 UTC m=+121.990652944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.096415 4978 ???:1] "http: TLS handshake error from 192.168.126.11:41730: no serving certificate available for the kubelet" Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.115138 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6j76w" event={"ID":"31337b76-7c44-42ab-a947-24c2932e2110","Type":"ContainerStarted","Data":"9cfdcc67be40f4de358bb67ffe635c6e1e22040a902cccbe5b5df5e174cee385"} Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.115960 4978 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pzb7w container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" start-of-body= Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.116012 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" podUID="6779444a-7630-434c-9d59-036633f57a09" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.29:5443/healthz\": dial tcp 10.217.0.29:5443: connect: connection refused" Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.116801 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" gracePeriod=30 Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.116994 4978 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9qf9f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.117025 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" podUID="9b843a14-9586-4f29-bfba-8413f4e52104" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.18:8080/healthz\": dial tcp 10.217.0.18:8080: connect: connection refused" Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.117140 4978 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-8wxst container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.117189 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" podUID="b0726188-e635-49f0-97ce-2cad95e63c13" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.150615 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:08 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:08 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:08 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.150678 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.151847 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.151995 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.651973837 +0000 UTC m=+122.091230296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.152202 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.154247 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.654233447 +0000 UTC m=+122.093489906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.253520 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.253721 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.753695605 +0000 UTC m=+122.192952064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.255158 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.271677 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.77165089 +0000 UTC m=+122.210907349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.278331 4978 ???:1] "http: TLS handshake error from 192.168.126.11:41732: no serving certificate available for the kubelet" Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.358868 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.359224 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.859208276 +0000 UTC m=+122.298464735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.459984 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.460419 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:08.960401527 +0000 UTC m=+122.399657976 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.561064 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.561408 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.061391442 +0000 UTC m=+122.500647901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.662891 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.663237 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.163225093 +0000 UTC m=+122.602481552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.764551 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.764959 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.264939689 +0000 UTC m=+122.704196148 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.865969 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.866315 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.366301895 +0000 UTC m=+122.805558354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.967484 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.967647 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.46762174 +0000 UTC m=+122.906878199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:08 crc kubenswrapper[4978]: I0225 06:47:08.967699 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:08 crc kubenswrapper[4978]: E0225 06:47:08.968004 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.467997091 +0000 UTC m=+122.907253550 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.155957 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.156120 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.656095399 +0000 UTC m=+123.095351858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.156168 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.156192 4978 ???:1] "http: TLS handshake error from 192.168.126.11:41736: no serving certificate available for the kubelet" Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.156550 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.656536872 +0000 UTC m=+123.095793331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.178046 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:09 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:09 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:09 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.178121 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.257484 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.258208 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.758192087 +0000 UTC m=+123.197448546 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.277813 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.278412 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.309100 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.315140 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.335314 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.359715 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ed029a4-c060-4c4c-b106-48ac222d1d62-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ed029a4-c060-4c4c-b106-48ac222d1d62\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.359847 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.359903 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ed029a4-c060-4c4c-b106-48ac222d1d62-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ed029a4-c060-4c4c-b106-48ac222d1d62\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.360218 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.860202493 +0000 UTC m=+123.299458942 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.461489 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.461637 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.961609521 +0000 UTC m=+123.400865980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.461833 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ed029a4-c060-4c4c-b106-48ac222d1d62-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ed029a4-c060-4c4c-b106-48ac222d1d62\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.461896 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ed029a4-c060-4c4c-b106-48ac222d1d62-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ed029a4-c060-4c4c-b106-48ac222d1d62\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.461968 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ed029a4-c060-4c4c-b106-48ac222d1d62-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9ed029a4-c060-4c4c-b106-48ac222d1d62\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.461984 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.462328 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:09.962301901 +0000 UTC m=+123.401558360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.557635 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ed029a4-c060-4c4c-b106-48ac222d1d62-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9ed029a4-c060-4c4c-b106-48ac222d1d62\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.563912 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.564214 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:10.064198614 +0000 UTC m=+123.503455073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.594841 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.644398 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qcr2x"] Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.644603 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" podUID="2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" containerName="controller-manager" containerID="cri-o://daab048b388dc747c710f14beeb448b15a0db6cb1c8971f68bd0d70ab271324e" gracePeriod=30 Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.649722 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.668314 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.668612 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:10.168595752 +0000 UTC m=+123.607852211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.750790 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5"] Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.750984 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" podUID="e6150f41-517c-4408-9a8e-a553c4f28ecb" containerName="route-controller-manager" containerID="cri-o://796e7fb483f414483b397115f435b712dc7fd842855ac9b5fc085797cba34a06" gracePeriod=30 Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.774210 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.774386 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:10.274348191 +0000 UTC m=+123.713604640 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.774493 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.774806 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:10.274792745 +0000 UTC m=+123.714049204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.876241 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.876583 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:10.376566553 +0000 UTC m=+123.815823012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:09 crc kubenswrapper[4978]: I0225 06:47:09.977664 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:09 crc kubenswrapper[4978]: E0225 06:47:09.977995 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:10.47798322 +0000 UTC m=+123.917239679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.078577 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.079360 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:10.579334436 +0000 UTC m=+124.018590895 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.113719 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.157017 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:10 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:10 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:10 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.157075 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.182188 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.182539 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:10.682527188 +0000 UTC m=+124.121783647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.224017 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6fbkc"] Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.230188 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.232750 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.253095 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6fbkc"] Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.283019 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.283193 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-utilities\") pod \"certified-operators-6fbkc\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.283214 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvlxt\" (UniqueName: \"kubernetes.io/projected/3054616f-87b0-467e-b5a7-499e3addf68a-kube-api-access-gvlxt\") pod \"certified-operators-6fbkc\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.283236 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-catalog-content\") pod \"certified-operators-6fbkc\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.283324 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:10.783310396 +0000 UTC m=+124.222566855 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.295699 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6j76w" event={"ID":"31337b76-7c44-42ab-a947-24c2932e2110","Type":"ContainerStarted","Data":"54290ab42bc5e774788fda23132407878397806de599042e1cc12dac53e04130"} Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.295742 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6j76w" event={"ID":"31337b76-7c44-42ab-a947-24c2932e2110","Type":"ContainerStarted","Data":"da56e61bf2a101967428c51da0d6918036bc52f6ccd07b8b7437bfb252d3d8e2"} Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.307261 4978 generic.go:334] "Generic (PLEG): container finished" podID="2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" containerID="daab048b388dc747c710f14beeb448b15a0db6cb1c8971f68bd0d70ab271324e" exitCode=0 Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.307313 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" event={"ID":"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f","Type":"ContainerDied","Data":"daab048b388dc747c710f14beeb448b15a0db6cb1c8971f68bd0d70ab271324e"} Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.308095 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9ed029a4-c060-4c4c-b106-48ac222d1d62","Type":"ContainerStarted","Data":"716947a062cb9a88aebb7fc627c4d3e3699b58bf2797982d5f2778582268e1a8"} Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.309114 4978 generic.go:334] "Generic (PLEG): container finished" podID="e6150f41-517c-4408-9a8e-a553c4f28ecb" containerID="796e7fb483f414483b397115f435b712dc7fd842855ac9b5fc085797cba34a06" exitCode=0 Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.309165 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" event={"ID":"e6150f41-517c-4408-9a8e-a553c4f28ecb","Type":"ContainerDied","Data":"796e7fb483f414483b397115f435b712dc7fd842855ac9b5fc085797cba34a06"} Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.332296 4978 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.372700 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.384001 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-client-ca\") pod \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.384094 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-proxy-ca-bundles\") pod \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.384117 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-serving-cert\") pod \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.384150 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-config\") pod \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.384356 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h25wg\" (UniqueName: \"kubernetes.io/projected/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-kube-api-access-h25wg\") pod \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\" (UID: \"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.384582 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.384631 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-utilities\") pod \"certified-operators-6fbkc\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.384650 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvlxt\" (UniqueName: \"kubernetes.io/projected/3054616f-87b0-467e-b5a7-499e3addf68a-kube-api-access-gvlxt\") pod \"certified-operators-6fbkc\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.384693 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-catalog-content\") pod \"certified-operators-6fbkc\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.385401 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-catalog-content\") pod \"certified-operators-6fbkc\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.386296 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-client-ca" (OuterVolumeSpecName: "client-ca") pod "2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" (UID: "2b9fb737-b4a4-4f4e-89f9-76f43468ca4f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.386837 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" (UID: "2b9fb737-b4a4-4f4e-89f9-76f43468ca4f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.389712 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:10.889696164 +0000 UTC m=+124.328952623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.390303 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-config" (OuterVolumeSpecName: "config") pod "2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" (UID: "2b9fb737-b4a4-4f4e-89f9-76f43468ca4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.390785 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-utilities\") pod \"certified-operators-6fbkc\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.416070 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-kube-api-access-h25wg" (OuterVolumeSpecName: "kube-api-access-h25wg") pod "2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" (UID: "2b9fb737-b4a4-4f4e-89f9-76f43468ca4f"). InnerVolumeSpecName "kube-api-access-h25wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.417824 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" (UID: "2b9fb737-b4a4-4f4e-89f9-76f43468ca4f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.429512 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvlxt\" (UniqueName: \"kubernetes.io/projected/3054616f-87b0-467e-b5a7-499e3addf68a-kube-api-access-gvlxt\") pod \"certified-operators-6fbkc\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.431431 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h9wx2"] Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.431717 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" containerName="controller-manager" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.431730 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" containerName="controller-manager" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.431884 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" containerName="controller-manager" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.432603 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.441732 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.446591 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h9wx2"] Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.501853 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.502072 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvwwx\" (UniqueName: \"kubernetes.io/projected/9283fdd2-ab8a-48ca-a326-1a0b107671a0-kube-api-access-gvwwx\") pod \"community-operators-h9wx2\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.502102 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-catalog-content\") pod \"community-operators-h9wx2\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.502169 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-utilities\") pod \"community-operators-h9wx2\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.502210 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.502220 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.502232 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.502242 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h25wg\" (UniqueName: \"kubernetes.io/projected/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-kube-api-access-h25wg\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.502250 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.502330 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:11.002314042 +0000 UTC m=+124.441570501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.535592 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-77b586cd9d-dlsrs"] Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.538129 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.548967 4978 ???:1] "http: TLS handshake error from 192.168.126.11:41750: no serving certificate available for the kubelet" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.564857 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-md9t9" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.564937 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.570967 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.596658 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77b586cd9d-dlsrs"] Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607187 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-config\") pod \"e6150f41-517c-4408-9a8e-a553c4f28ecb\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607316 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6150f41-517c-4408-9a8e-a553c4f28ecb-serving-cert\") pod \"e6150f41-517c-4408-9a8e-a553c4f28ecb\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607540 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-client-ca\") pod \"e6150f41-517c-4408-9a8e-a553c4f28ecb\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607595 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzm7z\" (UniqueName: \"kubernetes.io/projected/e6150f41-517c-4408-9a8e-a553c4f28ecb-kube-api-access-jzm7z\") pod \"e6150f41-517c-4408-9a8e-a553c4f28ecb\" (UID: \"e6150f41-517c-4408-9a8e-a553c4f28ecb\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607751 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-utilities\") pod \"community-operators-h9wx2\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607807 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-config\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607833 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvwwx\" (UniqueName: \"kubernetes.io/projected/9283fdd2-ab8a-48ca-a326-1a0b107671a0-kube-api-access-gvwwx\") pod \"community-operators-h9wx2\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607851 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-proxy-ca-bundles\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607874 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-serving-cert\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607895 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-catalog-content\") pod \"community-operators-h9wx2\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607919 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbcn5\" (UniqueName: \"kubernetes.io/projected/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-kube-api-access-kbcn5\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607952 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.607977 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-client-ca\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.609215 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-config" (OuterVolumeSpecName: "config") pod "e6150f41-517c-4408-9a8e-a553c4f28ecb" (UID: "e6150f41-517c-4408-9a8e-a553c4f28ecb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.611814 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-utilities\") pod \"community-operators-h9wx2\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.612153 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-catalog-content\") pod \"community-operators-h9wx2\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.612802 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:11.112785484 +0000 UTC m=+124.552041943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.614839 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-client-ca" (OuterVolumeSpecName: "client-ca") pod "e6150f41-517c-4408-9a8e-a553c4f28ecb" (UID: "e6150f41-517c-4408-9a8e-a553c4f28ecb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.642181 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6150f41-517c-4408-9a8e-a553c4f28ecb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e6150f41-517c-4408-9a8e-a553c4f28ecb" (UID: "e6150f41-517c-4408-9a8e-a553c4f28ecb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.642528 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6150f41-517c-4408-9a8e-a553c4f28ecb-kube-api-access-jzm7z" (OuterVolumeSpecName: "kube-api-access-jzm7z") pod "e6150f41-517c-4408-9a8e-a553c4f28ecb" (UID: "e6150f41-517c-4408-9a8e-a553c4f28ecb"). InnerVolumeSpecName "kube-api-access-jzm7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.635931 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h7dbj"] Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.652774 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6150f41-517c-4408-9a8e-a553c4f28ecb" containerName="route-controller-manager" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.652794 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6150f41-517c-4408-9a8e-a553c4f28ecb" containerName="route-controller-manager" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.652922 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6150f41-517c-4408-9a8e-a553c4f28ecb" containerName="route-controller-manager" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.654035 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.655573 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvwwx\" (UniqueName: \"kubernetes.io/projected/9283fdd2-ab8a-48ca-a326-1a0b107671a0-kube-api-access-gvwwx\") pod \"community-operators-h9wx2\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.668465 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h7dbj"] Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.722001 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.722560 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-config\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.722593 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-proxy-ca-bundles\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.722612 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-serving-cert\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.722634 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbcn5\" (UniqueName: \"kubernetes.io/projected/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-kube-api-access-kbcn5\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.722686 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-client-ca\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.722761 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.722970 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6150f41-517c-4408-9a8e-a553c4f28ecb-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.723698 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-client-ca\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.725747 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-config\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.726758 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-proxy-ca-bundles\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.727436 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e6150f41-517c-4408-9a8e-a553c4f28ecb-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.727456 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzm7z\" (UniqueName: \"kubernetes.io/projected/e6150f41-517c-4408-9a8e-a553c4f28ecb-kube-api-access-jzm7z\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.727504 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:11.227485905 +0000 UTC m=+124.666742364 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.732016 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-serving-cert\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.746009 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbcn5\" (UniqueName: \"kubernetes.io/projected/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-kube-api-access-kbcn5\") pod \"controller-manager-77b586cd9d-dlsrs\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.814585 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tjqjf"] Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.815701 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.816846 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.828895 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xnlb\" (UniqueName: \"kubernetes.io/projected/6a76f909-3198-47cd-adbb-9f94fb99ab48-kube-api-access-8xnlb\") pod \"certified-operators-h7dbj\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.828945 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-utilities\") pod \"community-operators-tjqjf\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.829017 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.829061 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-catalog-content\") pod \"community-operators-tjqjf\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.829108 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95x8l\" (UniqueName: \"kubernetes.io/projected/8178b540-4bca-402c-a2e0-57da7a09fa30-kube-api-access-95x8l\") pod \"community-operators-tjqjf\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.829127 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-utilities\") pod \"certified-operators-h7dbj\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.829144 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-catalog-content\") pod \"certified-operators-h7dbj\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.830006 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:11.329969375 +0000 UTC m=+124.769226004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.839551 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-g9xwf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.847689 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tjqjf"] Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.863097 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.911674 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.926813 4978 patch_prober.go:28] interesting pod/downloads-7954f5f757-jmhxz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.926852 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-jmhxz" podUID="028ba3b7-0bca-4541-a5be-7f63b26bdea4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.927192 4978 patch_prober.go:28] interesting pod/downloads-7954f5f757-jmhxz container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.927222 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-jmhxz" podUID="028ba3b7-0bca-4541-a5be-7f63b26bdea4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.932954 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.933357 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xnlb\" (UniqueName: \"kubernetes.io/projected/6a76f909-3198-47cd-adbb-9f94fb99ab48-kube-api-access-8xnlb\") pod \"certified-operators-h7dbj\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.933400 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-utilities\") pod \"community-operators-tjqjf\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.933457 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-catalog-content\") pod \"community-operators-tjqjf\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.933493 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95x8l\" (UniqueName: \"kubernetes.io/projected/8178b540-4bca-402c-a2e0-57da7a09fa30-kube-api-access-95x8l\") pod \"community-operators-tjqjf\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.933514 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-utilities\") pod \"certified-operators-h7dbj\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.933536 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-catalog-content\") pod \"certified-operators-h7dbj\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.933936 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-catalog-content\") pod \"certified-operators-h7dbj\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:10 crc kubenswrapper[4978]: E0225 06:47:10.934002 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:11.433988862 +0000 UTC m=+124.873245321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.934478 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-utilities\") pod \"community-operators-tjqjf\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.934685 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-catalog-content\") pod \"community-operators-tjqjf\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.937900 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-utilities\") pod \"certified-operators-h7dbj\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.967055 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-rrcbj" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.980219 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95x8l\" (UniqueName: \"kubernetes.io/projected/8178b540-4bca-402c-a2e0-57da7a09fa30-kube-api-access-95x8l\") pod \"community-operators-tjqjf\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:10 crc kubenswrapper[4978]: I0225 06:47:10.991113 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.020710 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xnlb\" (UniqueName: \"kubernetes.io/projected/6a76f909-3198-47cd-adbb-9f94fb99ab48-kube-api-access-8xnlb\") pod \"certified-operators-h7dbj\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.025347 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.045857 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:11 crc kubenswrapper[4978]: E0225 06:47:11.046175 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:11.546162336 +0000 UTC m=+124.985418795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.146856 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.148143 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:11 crc kubenswrapper[4978]: E0225 06:47:11.148335 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:11.648303435 +0000 UTC m=+125.087559894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.148876 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:11 crc kubenswrapper[4978]: E0225 06:47:11.150099 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-25 06:47:11.650082259 +0000 UTC m=+125.089338718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-w6cv7" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.220225 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:11 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:11 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:11 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.220505 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.251634 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:11 crc kubenswrapper[4978]: E0225 06:47:11.252107 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-25 06:47:11.752065164 +0000 UTC m=+125.191321623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.252181 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.252209 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.253883 4978 patch_prober.go:28] interesting pod/console-f9d7485db-r5cx2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.253931 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-r5cx2" podUID="b51a2671-2dea-418e-8f99-512e4ba4e34c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.295348 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.295734 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.310040 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.310001 4978 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-25T06:47:10.332337213Z","Handler":null,"Name":""} Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.330497 4978 patch_prober.go:28] interesting pod/apiserver-76f77b778f-ljw8k container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]log ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]etcd ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]poststarthook/generic-apiserver-start-informers ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]poststarthook/max-in-flight-filter ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 25 06:47:11 crc kubenswrapper[4978]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 25 06:47:11 crc kubenswrapper[4978]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Feb 25 06:47:11 crc kubenswrapper[4978]: [+]poststarthook/project.openshift.io-projectcache ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]poststarthook/openshift.io-startinformers ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 25 06:47:11 crc kubenswrapper[4978]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 25 06:47:11 crc kubenswrapper[4978]: livez check failed Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.330860 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" podUID="5c9a6a73-8690-43ac-928f-153713a523f6" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.338122 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.351649 4978 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.351690 4978 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.353341 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.382562 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.382637 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.395320 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qcr2x" event={"ID":"2b9fb737-b4a4-4f4e-89f9-76f43468ca4f","Type":"ContainerDied","Data":"d1782f842647f2c06ef2802e1ba932a4b61679dc816c347acf60cb3f28c7bd5c"} Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.395388 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6fbkc"] Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.395408 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9ed029a4-c060-4c4c-b106-48ac222d1d62","Type":"ContainerStarted","Data":"80ced6ce48f87a15f12c3ad6602bdbb1decf121699947644041615968d5e3540"} Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.395434 4978 scope.go:117] "RemoveContainer" containerID="daab048b388dc747c710f14beeb448b15a0db6cb1c8971f68bd0d70ab271324e" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.403895 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.404761 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5" event={"ID":"e6150f41-517c-4408-9a8e-a553c4f28ecb","Type":"ContainerDied","Data":"7df80d639d9c2311238ed4cf984ef9286efda80bf1927b95181e9d9bac88608a"} Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.460505 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h9wx2"] Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.462845 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.4628239499999998 podStartE2EDuration="2.46282395s" podCreationTimestamp="2026-02-25 06:47:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:11.417171425 +0000 UTC m=+124.856427884" watchObservedRunningTime="2026-02-25 06:47:11.46282395 +0000 UTC m=+124.902080409" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.464759 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-w6cv7\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.483000 4978 scope.go:117] "RemoveContainer" containerID="796e7fb483f414483b397115f435b712dc7fd842855ac9b5fc085797cba34a06" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.485573 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6j76w" event={"ID":"31337b76-7c44-42ab-a947-24c2932e2110","Type":"ContainerStarted","Data":"03d34b8567f4c013cf573f2f81fdcd5a851f3c1cce2e44e2b7b966be37df3773"} Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.536287 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.544543 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.561483 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.572497 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.627330 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-6j76w" podStartSLOduration=13.627311912 podStartE2EDuration="13.627311912s" podCreationTimestamp="2026-02-25 06:46:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:11.51391333 +0000 UTC m=+124.953169799" watchObservedRunningTime="2026-02-25 06:47:11.627311912 +0000 UTC m=+125.066568371" Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.636997 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qcr2x"] Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.641156 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qcr2x"] Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.655547 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5"] Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.661203 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-cmkt5"] Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.703858 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77b586cd9d-dlsrs"] Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.725670 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tjqjf"] Feb 25 06:47:11 crc kubenswrapper[4978]: I0225 06:47:11.779844 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h7dbj"] Feb 25 06:47:11 crc kubenswrapper[4978]: W0225 06:47:11.886664 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a76f909_3198_47cd_adbb_9f94fb99ab48.slice/crio-c3ff1f1c79ee21b7db55f0ecbf4880886400645959a36e2c971a62bab0b705c1 WatchSource:0}: Error finding container c3ff1f1c79ee21b7db55f0ecbf4880886400645959a36e2c971a62bab0b705c1: Status 404 returned error can't find the container with id c3ff1f1c79ee21b7db55f0ecbf4880886400645959a36e2c971a62bab0b705c1 Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.073967 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w6cv7"] Feb 25 06:47:12 crc kubenswrapper[4978]: W0225 06:47:12.102431 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cbd051d_53bc_4bd1_9c06_3c71859e7a7b.slice/crio-a98d99ded2c7c406b81f98a34c22fd475387a93da080fed28f540903efc001e0 WatchSource:0}: Error finding container a98d99ded2c7c406b81f98a34c22fd475387a93da080fed28f540903efc001e0: Status 404 returned error can't find the container with id a98d99ded2c7c406b81f98a34c22fd475387a93da080fed28f540903efc001e0 Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.143457 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:12 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:12 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:12 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.143523 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.376234 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.383632 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pzb7w" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.412097 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v9fsk"] Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.413190 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.415041 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 25 06:47:12 crc kubenswrapper[4978]: E0225 06:47:12.440990 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.443184 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9fsk"] Feb 25 06:47:12 crc kubenswrapper[4978]: E0225 06:47:12.444214 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:12 crc kubenswrapper[4978]: E0225 06:47:12.445644 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:12 crc kubenswrapper[4978]: E0225 06:47:12.445685 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.478260 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd"] Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.484153 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.496237 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.496391 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.496501 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.496723 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.496809 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.499153 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.504803 4978 generic.go:334] "Generic (PLEG): container finished" podID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerID="a956b032b2fa1fc6ca91347f778505b974cdf7926b68a9c2de2d1f287d10a222" exitCode=0 Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.505091 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7dbj" event={"ID":"6a76f909-3198-47cd-adbb-9f94fb99ab48","Type":"ContainerDied","Data":"a956b032b2fa1fc6ca91347f778505b974cdf7926b68a9c2de2d1f287d10a222"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.505144 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7dbj" event={"ID":"6a76f909-3198-47cd-adbb-9f94fb99ab48","Type":"ContainerStarted","Data":"c3ff1f1c79ee21b7db55f0ecbf4880886400645959a36e2c971a62bab0b705c1"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.512254 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.519492 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" event={"ID":"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0","Type":"ContainerStarted","Data":"8c322ccfbfb9a156c5e69fb9295d78bd840619315b5a55bb00933e772c646b17"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.519714 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" event={"ID":"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0","Type":"ContainerStarted","Data":"2ac12a83a0e0511826497ed7b88008799f5a3abdbb59f49267febe2b032f8c93"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.521010 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.525484 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd"] Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.541710 4978 generic.go:334] "Generic (PLEG): container finished" podID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerID="b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a" exitCode=0 Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.542015 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9wx2" event={"ID":"9283fdd2-ab8a-48ca-a326-1a0b107671a0","Type":"ContainerDied","Data":"b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.542053 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9wx2" event={"ID":"9283fdd2-ab8a-48ca-a326-1a0b107671a0","Type":"ContainerStarted","Data":"f4455be87d9d95d12c5a65362832348affe098310322398912a17232c2146f10"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.545174 4978 generic.go:334] "Generic (PLEG): container finished" podID="7fd9f3d7-d711-43cb-a896-7c77507621bc" containerID="8a9a06d5f426c85882f2a71024311846c9ece01c436f2a16be0cc489eade32d7" exitCode=0 Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.545224 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" event={"ID":"7fd9f3d7-d711-43cb-a896-7c77507621bc","Type":"ContainerDied","Data":"8a9a06d5f426c85882f2a71024311846c9ece01c436f2a16be0cc489eade32d7"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.546533 4978 generic.go:334] "Generic (PLEG): container finished" podID="9ed029a4-c060-4c4c-b106-48ac222d1d62" containerID="80ced6ce48f87a15f12c3ad6602bdbb1decf121699947644041615968d5e3540" exitCode=0 Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.546594 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9ed029a4-c060-4c4c-b106-48ac222d1d62","Type":"ContainerDied","Data":"80ced6ce48f87a15f12c3ad6602bdbb1decf121699947644041615968d5e3540"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.561652 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.561839 4978 generic.go:334] "Generic (PLEG): container finished" podID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerID="7953a2670e32437e4b3852e26b73bed4b928a3596a4e252f6af7dc80c09f9925" exitCode=0 Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.561896 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tjqjf" event={"ID":"8178b540-4bca-402c-a2e0-57da7a09fa30","Type":"ContainerDied","Data":"7953a2670e32437e4b3852e26b73bed4b928a3596a4e252f6af7dc80c09f9925"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.561911 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tjqjf" event={"ID":"8178b540-4bca-402c-a2e0-57da7a09fa30","Type":"ContainerStarted","Data":"a9f26814afc265765d15d5140f4133d7ac53d797c1578a17c58dc8af237f0080"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.570754 4978 generic.go:334] "Generic (PLEG): container finished" podID="3054616f-87b0-467e-b5a7-499e3addf68a" containerID="c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76" exitCode=0 Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.570839 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fbkc" event={"ID":"3054616f-87b0-467e-b5a7-499e3addf68a","Type":"ContainerDied","Data":"c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.570863 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fbkc" event={"ID":"3054616f-87b0-467e-b5a7-499e3addf68a","Type":"ContainerStarted","Data":"e23e2633f2ce8205b0d62b2109ade68af4e8bdc5c8f8f8d6f5c0a62134ea0a8e"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.572115 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" podStartSLOduration=2.572094752 podStartE2EDuration="2.572094752s" podCreationTimestamp="2026-02-25 06:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:12.569741921 +0000 UTC m=+126.008998380" watchObservedRunningTime="2026-02-25 06:47:12.572094752 +0000 UTC m=+126.011351211" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.578757 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsl92\" (UniqueName: \"kubernetes.io/projected/d241c721-1ac8-4bda-93cb-ae761f0075d5-kube-api-access-bsl92\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.578821 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-catalog-content\") pod \"redhat-marketplace-v9fsk\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.578870 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d241c721-1ac8-4bda-93cb-ae761f0075d5-serving-cert\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.578899 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-config\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.578928 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-utilities\") pod \"redhat-marketplace-v9fsk\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.578954 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-client-ca\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.578972 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w2qq\" (UniqueName: \"kubernetes.io/projected/8ae5f061-fdff-4fc5-b252-4734b7ad892a-kube-api-access-8w2qq\") pod \"redhat-marketplace-v9fsk\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.581301 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" event={"ID":"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b","Type":"ContainerStarted","Data":"496811eccf93eaa4d501eb231825144e91b438ae0f66a11b53f747acee76bb2a"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.581350 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" event={"ID":"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b","Type":"ContainerStarted","Data":"a98d99ded2c7c406b81f98a34c22fd475387a93da080fed28f540903efc001e0"} Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.581406 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.667907 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" podStartSLOduration=74.667890829 podStartE2EDuration="1m14.667890829s" podCreationTimestamp="2026-02-25 06:45:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:12.639650222 +0000 UTC m=+126.078906681" watchObservedRunningTime="2026-02-25 06:47:12.667890829 +0000 UTC m=+126.107147288" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.681008 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-client-ca\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.681075 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w2qq\" (UniqueName: \"kubernetes.io/projected/8ae5f061-fdff-4fc5-b252-4734b7ad892a-kube-api-access-8w2qq\") pod \"redhat-marketplace-v9fsk\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.681099 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsl92\" (UniqueName: \"kubernetes.io/projected/d241c721-1ac8-4bda-93cb-ae761f0075d5-kube-api-access-bsl92\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.681159 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-catalog-content\") pod \"redhat-marketplace-v9fsk\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.681208 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d241c721-1ac8-4bda-93cb-ae761f0075d5-serving-cert\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.681234 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-config\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.681317 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-utilities\") pod \"redhat-marketplace-v9fsk\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.682931 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-client-ca\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.685454 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-catalog-content\") pod \"redhat-marketplace-v9fsk\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.686599 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-utilities\") pod \"redhat-marketplace-v9fsk\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.687666 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-config\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.701282 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d241c721-1ac8-4bda-93cb-ae761f0075d5-serving-cert\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.733181 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsl92\" (UniqueName: \"kubernetes.io/projected/d241c721-1ac8-4bda-93cb-ae761f0075d5-kube-api-access-bsl92\") pod \"route-controller-manager-78b646bc84-xjqmd\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.734307 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w2qq\" (UniqueName: \"kubernetes.io/projected/8ae5f061-fdff-4fc5-b252-4734b7ad892a-kube-api-access-8w2qq\") pod \"redhat-marketplace-v9fsk\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.763241 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.763855 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.775844 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.776318 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.776694 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.837567 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.885284 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2439c661-055f-46ca-bbea-6be73bab6260-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2439c661-055f-46ca-bbea-6be73bab6260\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.885331 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2439c661-055f-46ca-bbea-6be73bab6260-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2439c661-055f-46ca-bbea-6be73bab6260\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.987006 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2439c661-055f-46ca-bbea-6be73bab6260-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2439c661-055f-46ca-bbea-6be73bab6260\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.987045 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2439c661-055f-46ca-bbea-6be73bab6260-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2439c661-055f-46ca-bbea-6be73bab6260\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 06:47:12 crc kubenswrapper[4978]: I0225 06:47:12.987108 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2439c661-055f-46ca-bbea-6be73bab6260-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2439c661-055f-46ca-bbea-6be73bab6260\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.028008 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.331898 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:13 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:13 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:13 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.331958 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.323750 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.422911 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b9fb737-b4a4-4f4e-89f9-76f43468ca4f" path="/var/lib/kubelet/pods/2b9fb737-b4a4-4f4e-89f9-76f43468ca4f/volumes" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.423705 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.424203 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6150f41-517c-4408-9a8e-a553c4f28ecb" path="/var/lib/kubelet/pods/e6150f41-517c-4408-9a8e-a553c4f28ecb/volumes" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.425420 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8wxst" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.425447 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w9qqz"] Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.426350 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9qqz"] Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.427761 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.428559 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2439c661-055f-46ca-bbea-6be73bab6260-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2439c661-055f-46ca-bbea-6be73bab6260\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.434598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-catalog-content\") pod \"redhat-marketplace-w9qqz\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.437357 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr85c\" (UniqueName: \"kubernetes.io/projected/ca2405db-a877-4fda-a24c-ff39a0069758-kube-api-access-tr85c\") pod \"redhat-marketplace-w9qqz\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.437668 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-utilities\") pod \"redhat-marketplace-w9qqz\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.442763 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cghmv"] Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.446021 4978 ???:1] "http: TLS handshake error from 192.168.126.11:57490: no serving certificate available for the kubelet" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.450098 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.460986 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.485463 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cghmv"] Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.502064 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.540280 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-utilities\") pod \"redhat-operators-cghmv\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.540333 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-utilities\") pod \"redhat-marketplace-w9qqz\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.540401 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-catalog-content\") pod \"redhat-operators-cghmv\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.540434 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-catalog-content\") pod \"redhat-marketplace-w9qqz\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.540456 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr85c\" (UniqueName: \"kubernetes.io/projected/ca2405db-a877-4fda-a24c-ff39a0069758-kube-api-access-tr85c\") pod \"redhat-marketplace-w9qqz\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.540482 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctx5l\" (UniqueName: \"kubernetes.io/projected/6f6ed9c0-cb59-4aac-af72-2633f18378c9-kube-api-access-ctx5l\") pod \"redhat-operators-cghmv\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.541641 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-utilities\") pod \"redhat-marketplace-w9qqz\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.541722 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-catalog-content\") pod \"redhat-marketplace-w9qqz\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.584564 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr85c\" (UniqueName: \"kubernetes.io/projected/ca2405db-a877-4fda-a24c-ff39a0069758-kube-api-access-tr85c\") pod \"redhat-marketplace-w9qqz\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.644482 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-utilities\") pod \"redhat-operators-cghmv\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.644558 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-catalog-content\") pod \"redhat-operators-cghmv\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.644603 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctx5l\" (UniqueName: \"kubernetes.io/projected/6f6ed9c0-cb59-4aac-af72-2633f18378c9-kube-api-access-ctx5l\") pod \"redhat-operators-cghmv\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.645219 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-utilities\") pod \"redhat-operators-cghmv\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.645719 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-catalog-content\") pod \"redhat-operators-cghmv\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.665596 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctx5l\" (UniqueName: \"kubernetes.io/projected/6f6ed9c0-cb59-4aac-af72-2633f18378c9-kube-api-access-ctx5l\") pod \"redhat-operators-cghmv\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.712824 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.806226 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zzswn"] Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.807482 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.808385 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.818743 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zzswn"] Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.852264 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.950851 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-catalog-content\") pod \"redhat-operators-zzswn\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.950902 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-utilities\") pod \"redhat-operators-zzswn\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:13 crc kubenswrapper[4978]: I0225 06:47:13.950954 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxzv2\" (UniqueName: \"kubernetes.io/projected/d6342947-b76f-4c86-8e34-d42a780c6bf5-kube-api-access-hxzv2\") pod \"redhat-operators-zzswn\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.011529 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9fsk"] Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.053470 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxzv2\" (UniqueName: \"kubernetes.io/projected/d6342947-b76f-4c86-8e34-d42a780c6bf5-kube-api-access-hxzv2\") pod \"redhat-operators-zzswn\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.053541 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-catalog-content\") pod \"redhat-operators-zzswn\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.053570 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-utilities\") pod \"redhat-operators-zzswn\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.053957 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-utilities\") pod \"redhat-operators-zzswn\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.054417 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-catalog-content\") pod \"redhat-operators-zzswn\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.080847 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd"] Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.091599 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxzv2\" (UniqueName: \"kubernetes.io/projected/d6342947-b76f-4c86-8e34-d42a780c6bf5-kube-api-access-hxzv2\") pod \"redhat-operators-zzswn\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.116261 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 06:47:14 crc kubenswrapper[4978]: W0225 06:47:14.125629 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd241c721_1ac8_4bda_93cb_ae761f0075d5.slice/crio-17d322426d1d7d25c3e0a8d152f700a347a17ec274184346278704f2ba474c82 WatchSource:0}: Error finding container 17d322426d1d7d25c3e0a8d152f700a347a17ec274184346278704f2ba474c82: Status 404 returned error can't find the container with id 17d322426d1d7d25c3e0a8d152f700a347a17ec274184346278704f2ba474c82 Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.140320 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.151541 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:14 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:14 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:14 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.151582 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.154424 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.201279 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.263821 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fd9f3d7-d711-43cb-a896-7c77507621bc-config-volume\") pod \"7fd9f3d7-d711-43cb-a896-7c77507621bc\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.263883 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ed029a4-c060-4c4c-b106-48ac222d1d62-kubelet-dir\") pod \"9ed029a4-c060-4c4c-b106-48ac222d1d62\" (UID: \"9ed029a4-c060-4c4c-b106-48ac222d1d62\") " Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.263943 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fd9f3d7-d711-43cb-a896-7c77507621bc-secret-volume\") pod \"7fd9f3d7-d711-43cb-a896-7c77507621bc\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.263990 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7xkz\" (UniqueName: \"kubernetes.io/projected/7fd9f3d7-d711-43cb-a896-7c77507621bc-kube-api-access-n7xkz\") pod \"7fd9f3d7-d711-43cb-a896-7c77507621bc\" (UID: \"7fd9f3d7-d711-43cb-a896-7c77507621bc\") " Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.264013 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ed029a4-c060-4c4c-b106-48ac222d1d62-kube-api-access\") pod \"9ed029a4-c060-4c4c-b106-48ac222d1d62\" (UID: \"9ed029a4-c060-4c4c-b106-48ac222d1d62\") " Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.265421 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ed029a4-c060-4c4c-b106-48ac222d1d62-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9ed029a4-c060-4c4c-b106-48ac222d1d62" (UID: "9ed029a4-c060-4c4c-b106-48ac222d1d62"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.273142 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fd9f3d7-d711-43cb-a896-7c77507621bc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7fd9f3d7-d711-43cb-a896-7c77507621bc" (UID: "7fd9f3d7-d711-43cb-a896-7c77507621bc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.276301 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fd9f3d7-d711-43cb-a896-7c77507621bc-config-volume" (OuterVolumeSpecName: "config-volume") pod "7fd9f3d7-d711-43cb-a896-7c77507621bc" (UID: "7fd9f3d7-d711-43cb-a896-7c77507621bc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.290034 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed029a4-c060-4c4c-b106-48ac222d1d62-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9ed029a4-c060-4c4c-b106-48ac222d1d62" (UID: "9ed029a4-c060-4c4c-b106-48ac222d1d62"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.290510 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd9f3d7-d711-43cb-a896-7c77507621bc-kube-api-access-n7xkz" (OuterVolumeSpecName: "kube-api-access-n7xkz") pod "7fd9f3d7-d711-43cb-a896-7c77507621bc" (UID: "7fd9f3d7-d711-43cb-a896-7c77507621bc"). InnerVolumeSpecName "kube-api-access-n7xkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.365157 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7fd9f3d7-d711-43cb-a896-7c77507621bc-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.365513 4978 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9ed029a4-c060-4c4c-b106-48ac222d1d62-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.365522 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7fd9f3d7-d711-43cb-a896-7c77507621bc-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.365532 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7xkz\" (UniqueName: \"kubernetes.io/projected/7fd9f3d7-d711-43cb-a896-7c77507621bc-kube-api-access-n7xkz\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.365542 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9ed029a4-c060-4c4c-b106-48ac222d1d62-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.469944 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cghmv"] Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.497391 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9qqz"] Feb 25 06:47:14 crc kubenswrapper[4978]: W0225 06:47:14.507236 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f6ed9c0_cb59_4aac_af72_2633f18378c9.slice/crio-671793f646d39d34d80a7eaf6dff6f51fd8cf43ef70f7de1cb983ba7d9261b52 WatchSource:0}: Error finding container 671793f646d39d34d80a7eaf6dff6f51fd8cf43ef70f7de1cb983ba7d9261b52: Status 404 returned error can't find the container with id 671793f646d39d34d80a7eaf6dff6f51fd8cf43ef70f7de1cb983ba7d9261b52 Feb 25 06:47:14 crc kubenswrapper[4978]: W0225 06:47:14.538562 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca2405db_a877_4fda_a24c_ff39a0069758.slice/crio-f60379f97bc3bfd556c1a8759fb483182bdd7cf2441356c77eb3a0e9914870ef WatchSource:0}: Error finding container f60379f97bc3bfd556c1a8759fb483182bdd7cf2441356c77eb3a0e9914870ef: Status 404 returned error can't find the container with id f60379f97bc3bfd556c1a8759fb483182bdd7cf2441356c77eb3a0e9914870ef Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.719104 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.719458 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk" event={"ID":"7fd9f3d7-d711-43cb-a896-7c77507621bc","Type":"ContainerDied","Data":"dcb3ee7dbd229617a864f6eeb76da93e8cb1ec91eec81d655cb966eee9feaebb"} Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.719902 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcb3ee7dbd229617a864f6eeb76da93e8cb1ec91eec81d655cb966eee9feaebb" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.725942 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9qqz" event={"ID":"ca2405db-a877-4fda-a24c-ff39a0069758","Type":"ContainerStarted","Data":"f60379f97bc3bfd556c1a8759fb483182bdd7cf2441356c77eb3a0e9914870ef"} Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.728645 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cghmv" event={"ID":"6f6ed9c0-cb59-4aac-af72-2633f18378c9","Type":"ContainerStarted","Data":"671793f646d39d34d80a7eaf6dff6f51fd8cf43ef70f7de1cb983ba7d9261b52"} Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.730549 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2439c661-055f-46ca-bbea-6be73bab6260","Type":"ContainerStarted","Data":"5f5f3657286c1e6c5e554a5aa0e259594a0c14c9c45d755a7d19eba92dda60ef"} Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.733343 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" event={"ID":"d241c721-1ac8-4bda-93cb-ae761f0075d5","Type":"ContainerStarted","Data":"bfd1ff0a503d419ae3deaccdd1c2ec0baafbf653e0292b9c185c5965edcf6226"} Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.733428 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" event={"ID":"d241c721-1ac8-4bda-93cb-ae761f0075d5","Type":"ContainerStarted","Data":"17d322426d1d7d25c3e0a8d152f700a347a17ec274184346278704f2ba474c82"} Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.734571 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.741608 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9ed029a4-c060-4c4c-b106-48ac222d1d62","Type":"ContainerDied","Data":"716947a062cb9a88aebb7fc627c4d3e3699b58bf2797982d5f2778582268e1a8"} Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.741650 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="716947a062cb9a88aebb7fc627c4d3e3699b58bf2797982d5f2778582268e1a8" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.741739 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.742517 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zzswn"] Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.748942 4978 generic.go:334] "Generic (PLEG): container finished" podID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerID="33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5" exitCode=0 Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.751965 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9fsk" event={"ID":"8ae5f061-fdff-4fc5-b252-4734b7ad892a","Type":"ContainerDied","Data":"33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5"} Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.752011 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9fsk" event={"ID":"8ae5f061-fdff-4fc5-b252-4734b7ad892a","Type":"ContainerStarted","Data":"007e2ab9c79f9fab639dfdf6d2fedcdb60166a5acdb28a35d1aa9d22a3c18ec7"} Feb 25 06:47:14 crc kubenswrapper[4978]: I0225 06:47:14.754491 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" podStartSLOduration=4.75447612 podStartE2EDuration="4.75447612s" podCreationTimestamp="2026-02-25 06:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:14.752514311 +0000 UTC m=+128.191770790" watchObservedRunningTime="2026-02-25 06:47:14.75447612 +0000 UTC m=+128.193732579" Feb 25 06:47:14 crc kubenswrapper[4978]: W0225 06:47:14.814294 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6342947_b76f_4c86_8e34_d42a780c6bf5.slice/crio-e6526a36d32af4b3b556d40482ae7d3c9abcff2ed30a29e98bb7687ad3926b14 WatchSource:0}: Error finding container e6526a36d32af4b3b556d40482ae7d3c9abcff2ed30a29e98bb7687ad3926b14: Status 404 returned error can't find the container with id e6526a36d32af4b3b556d40482ae7d3c9abcff2ed30a29e98bb7687ad3926b14 Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.017037 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.142947 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:15 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:15 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:15 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.143011 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.756068 4978 generic.go:334] "Generic (PLEG): container finished" podID="2439c661-055f-46ca-bbea-6be73bab6260" containerID="a836c42d91c62e23ed1a8831cf03d0fc91c5f6efa189b574cb6e8b8b3cdc59b6" exitCode=0 Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.756573 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2439c661-055f-46ca-bbea-6be73bab6260","Type":"ContainerDied","Data":"a836c42d91c62e23ed1a8831cf03d0fc91c5f6efa189b574cb6e8b8b3cdc59b6"} Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.759637 4978 generic.go:334] "Generic (PLEG): container finished" podID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerID="32a86d94a98bf96554a0b8e20a6003e2c9c01b97626c822cfe5cee85af8ef71c" exitCode=0 Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.759676 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzswn" event={"ID":"d6342947-b76f-4c86-8e34-d42a780c6bf5","Type":"ContainerDied","Data":"32a86d94a98bf96554a0b8e20a6003e2c9c01b97626c822cfe5cee85af8ef71c"} Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.759692 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzswn" event={"ID":"d6342947-b76f-4c86-8e34-d42a780c6bf5","Type":"ContainerStarted","Data":"e6526a36d32af4b3b556d40482ae7d3c9abcff2ed30a29e98bb7687ad3926b14"} Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.766616 4978 generic.go:334] "Generic (PLEG): container finished" podID="ca2405db-a877-4fda-a24c-ff39a0069758" containerID="38462a0fa2ea01aea1677a7e7e80488587f32dac33403ba064064b919263ce79" exitCode=0 Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.766886 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9qqz" event={"ID":"ca2405db-a877-4fda-a24c-ff39a0069758","Type":"ContainerDied","Data":"38462a0fa2ea01aea1677a7e7e80488587f32dac33403ba064064b919263ce79"} Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.796175 4978 generic.go:334] "Generic (PLEG): container finished" podID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerID="5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab" exitCode=0 Feb 25 06:47:15 crc kubenswrapper[4978]: I0225 06:47:15.796862 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cghmv" event={"ID":"6f6ed9c0-cb59-4aac-af72-2633f18378c9","Type":"ContainerDied","Data":"5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab"} Feb 25 06:47:16 crc kubenswrapper[4978]: I0225 06:47:16.145219 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:16 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:16 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:16 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:16 crc kubenswrapper[4978]: I0225 06:47:16.145830 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:16 crc kubenswrapper[4978]: I0225 06:47:16.297720 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:47:16 crc kubenswrapper[4978]: I0225 06:47:16.304507 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-ljw8k" Feb 25 06:47:16 crc kubenswrapper[4978]: I0225 06:47:16.597394 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:47:16 crc kubenswrapper[4978]: I0225 06:47:16.868513 4978 ???:1] "http: TLS handshake error from 192.168.126.11:57504: no serving certificate available for the kubelet" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.049807 4978 ???:1] "http: TLS handshake error from 192.168.126.11:57506: no serving certificate available for the kubelet" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.135648 4978 ???:1] "http: TLS handshake error from 192.168.126.11:57508: no serving certificate available for the kubelet" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.142656 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:17 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:17 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:17 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.142705 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.217632 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.347189 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2439c661-055f-46ca-bbea-6be73bab6260-kubelet-dir\") pod \"2439c661-055f-46ca-bbea-6be73bab6260\" (UID: \"2439c661-055f-46ca-bbea-6be73bab6260\") " Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.347244 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2439c661-055f-46ca-bbea-6be73bab6260-kube-api-access\") pod \"2439c661-055f-46ca-bbea-6be73bab6260\" (UID: \"2439c661-055f-46ca-bbea-6be73bab6260\") " Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.347311 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2439c661-055f-46ca-bbea-6be73bab6260-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2439c661-055f-46ca-bbea-6be73bab6260" (UID: "2439c661-055f-46ca-bbea-6be73bab6260"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.347635 4978 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2439c661-055f-46ca-bbea-6be73bab6260-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.352485 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.367719 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2439c661-055f-46ca-bbea-6be73bab6260-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2439c661-055f-46ca-bbea-6be73bab6260" (UID: "2439c661-055f-46ca-bbea-6be73bab6260"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.448512 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2439c661-055f-46ca-bbea-6be73bab6260-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.490281 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dqrgj" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.524553 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=0.524532191 podStartE2EDuration="524.532191ms" podCreationTimestamp="2026-02-25 06:47:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:47:17.523606503 +0000 UTC m=+130.962862962" watchObservedRunningTime="2026-02-25 06:47:17.524532191 +0000 UTC m=+130.963788650" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.829249 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.829299 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2439c661-055f-46ca-bbea-6be73bab6260","Type":"ContainerDied","Data":"5f5f3657286c1e6c5e554a5aa0e259594a0c14c9c45d755a7d19eba92dda60ef"} Feb 25 06:47:17 crc kubenswrapper[4978]: I0225 06:47:17.829326 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f5f3657286c1e6c5e554a5aa0e259594a0c14c9c45d755a7d19eba92dda60ef" Feb 25 06:47:18 crc kubenswrapper[4978]: I0225 06:47:18.143640 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:18 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:18 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:18 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:18 crc kubenswrapper[4978]: I0225 06:47:18.144043 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:18 crc kubenswrapper[4978]: I0225 06:47:18.601003 4978 ???:1] "http: TLS handshake error from 192.168.126.11:57518: no serving certificate available for the kubelet" Feb 25 06:47:19 crc kubenswrapper[4978]: I0225 06:47:19.142934 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:19 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:19 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:19 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:19 crc kubenswrapper[4978]: I0225 06:47:19.143014 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:20 crc kubenswrapper[4978]: I0225 06:47:20.143821 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:20 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:20 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:20 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:20 crc kubenswrapper[4978]: I0225 06:47:20.144182 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:20 crc kubenswrapper[4978]: I0225 06:47:20.942863 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-jmhxz" Feb 25 06:47:21 crc kubenswrapper[4978]: I0225 06:47:21.143152 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:21 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:21 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:21 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:21 crc kubenswrapper[4978]: I0225 06:47:21.143224 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:21 crc kubenswrapper[4978]: I0225 06:47:21.247008 4978 patch_prober.go:28] interesting pod/console-f9d7485db-r5cx2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Feb 25 06:47:21 crc kubenswrapper[4978]: I0225 06:47:21.247073 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-r5cx2" podUID="b51a2671-2dea-418e-8f99-512e4ba4e34c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.20:8443/health\": dial tcp 10.217.0.20:8443: connect: connection refused" Feb 25 06:47:22 crc kubenswrapper[4978]: I0225 06:47:22.141955 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:22 crc kubenswrapper[4978]: [-]has-synced failed: reason withheld Feb 25 06:47:22 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:22 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:22 crc kubenswrapper[4978]: I0225 06:47:22.142029 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:22 crc kubenswrapper[4978]: E0225 06:47:22.455961 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:22 crc kubenswrapper[4978]: E0225 06:47:22.469639 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:22 crc kubenswrapper[4978]: E0225 06:47:22.475845 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:22 crc kubenswrapper[4978]: E0225 06:47:22.475918 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" Feb 25 06:47:23 crc kubenswrapper[4978]: I0225 06:47:23.141992 4978 patch_prober.go:28] interesting pod/router-default-5444994796-ml6mx container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 25 06:47:23 crc kubenswrapper[4978]: [+]has-synced ok Feb 25 06:47:23 crc kubenswrapper[4978]: [+]process-running ok Feb 25 06:47:23 crc kubenswrapper[4978]: healthz check failed Feb 25 06:47:23 crc kubenswrapper[4978]: I0225 06:47:23.142043 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-ml6mx" podUID="811afd11-d673-4ad8-b6ba-4d766d6affe5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 25 06:47:24 crc kubenswrapper[4978]: I0225 06:47:24.143282 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:24 crc kubenswrapper[4978]: I0225 06:47:24.147868 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-ml6mx" Feb 25 06:47:28 crc kubenswrapper[4978]: I0225 06:47:28.486096 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-77b586cd9d-dlsrs"] Feb 25 06:47:28 crc kubenswrapper[4978]: I0225 06:47:28.486643 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" podUID="616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" containerName="controller-manager" containerID="cri-o://8c322ccfbfb9a156c5e69fb9295d78bd840619315b5a55bb00933e772c646b17" gracePeriod=30 Feb 25 06:47:28 crc kubenswrapper[4978]: I0225 06:47:28.524267 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd"] Feb 25 06:47:28 crc kubenswrapper[4978]: I0225 06:47:28.524633 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" podUID="d241c721-1ac8-4bda-93cb-ae761f0075d5" containerName="route-controller-manager" containerID="cri-o://bfd1ff0a503d419ae3deaccdd1c2ec0baafbf653e0292b9c185c5965edcf6226" gracePeriod=30 Feb 25 06:47:29 crc kubenswrapper[4978]: I0225 06:47:29.917293 4978 generic.go:334] "Generic (PLEG): container finished" podID="616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" containerID="8c322ccfbfb9a156c5e69fb9295d78bd840619315b5a55bb00933e772c646b17" exitCode=0 Feb 25 06:47:29 crc kubenswrapper[4978]: I0225 06:47:29.917644 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" event={"ID":"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0","Type":"ContainerDied","Data":"8c322ccfbfb9a156c5e69fb9295d78bd840619315b5a55bb00933e772c646b17"} Feb 25 06:47:29 crc kubenswrapper[4978]: I0225 06:47:29.920566 4978 generic.go:334] "Generic (PLEG): container finished" podID="d241c721-1ac8-4bda-93cb-ae761f0075d5" containerID="bfd1ff0a503d419ae3deaccdd1c2ec0baafbf653e0292b9c185c5965edcf6226" exitCode=0 Feb 25 06:47:29 crc kubenswrapper[4978]: I0225 06:47:29.920634 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" event={"ID":"d241c721-1ac8-4bda-93cb-ae761f0075d5","Type":"ContainerDied","Data":"bfd1ff0a503d419ae3deaccdd1c2ec0baafbf653e0292b9c185c5965edcf6226"} Feb 25 06:47:30 crc kubenswrapper[4978]: I0225 06:47:30.913445 4978 patch_prober.go:28] interesting pod/controller-manager-77b586cd9d-dlsrs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.47:8443/healthz\": dial tcp 10.217.0.47:8443: connect: connection refused" start-of-body= Feb 25 06:47:30 crc kubenswrapper[4978]: I0225 06:47:30.913528 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" podUID="616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.47:8443/healthz\": dial tcp 10.217.0.47:8443: connect: connection refused" Feb 25 06:47:31 crc kubenswrapper[4978]: I0225 06:47:31.252133 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:47:31 crc kubenswrapper[4978]: I0225 06:47:31.257565 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 06:47:31 crc kubenswrapper[4978]: I0225 06:47:31.549549 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:47:32 crc kubenswrapper[4978]: E0225 06:47:32.442775 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:32 crc kubenswrapper[4978]: E0225 06:47:32.445176 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:32 crc kubenswrapper[4978]: E0225 06:47:32.447491 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:32 crc kubenswrapper[4978]: E0225 06:47:32.447548 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" Feb 25 06:47:32 crc kubenswrapper[4978]: I0225 06:47:32.839500 4978 patch_prober.go:28] interesting pod/route-controller-manager-78b646bc84-xjqmd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": dial tcp 10.217.0.51:8443: connect: connection refused" start-of-body= Feb 25 06:47:32 crc kubenswrapper[4978]: I0225 06:47:32.839572 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" podUID="d241c721-1ac8-4bda-93cb-ae761f0075d5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": dial tcp 10.217.0.51:8443: connect: connection refused" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.793702 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.825025 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6686fbcbd7-4zltb"] Feb 25 06:47:37 crc kubenswrapper[4978]: E0225 06:47:37.825273 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed029a4-c060-4c4c-b106-48ac222d1d62" containerName="pruner" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.825289 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed029a4-c060-4c4c-b106-48ac222d1d62" containerName="pruner" Feb 25 06:47:37 crc kubenswrapper[4978]: E0225 06:47:37.825305 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" containerName="controller-manager" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.825316 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" containerName="controller-manager" Feb 25 06:47:37 crc kubenswrapper[4978]: E0225 06:47:37.825326 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd9f3d7-d711-43cb-a896-7c77507621bc" containerName="collect-profiles" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.825336 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd9f3d7-d711-43cb-a896-7c77507621bc" containerName="collect-profiles" Feb 25 06:47:37 crc kubenswrapper[4978]: E0225 06:47:37.825348 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2439c661-055f-46ca-bbea-6be73bab6260" containerName="pruner" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.825356 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2439c661-055f-46ca-bbea-6be73bab6260" containerName="pruner" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.825520 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" containerName="controller-manager" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.825534 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed029a4-c060-4c4c-b106-48ac222d1d62" containerName="pruner" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.825544 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2439c661-055f-46ca-bbea-6be73bab6260" containerName="pruner" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.825558 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd9f3d7-d711-43cb-a896-7c77507621bc" containerName="collect-profiles" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.825980 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.852846 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6686fbcbd7-4zltb"] Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.944195 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-serving-cert\") pod \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.944239 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-proxy-ca-bundles\") pod \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.944331 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-client-ca\") pod \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.944354 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbcn5\" (UniqueName: \"kubernetes.io/projected/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-kube-api-access-kbcn5\") pod \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.944412 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-config\") pod \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\" (UID: \"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0\") " Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.944874 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-proxy-ca-bundles\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.944926 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94cf6\" (UniqueName: \"kubernetes.io/projected/ccd5faea-8284-4164-8383-72800aea2048-kube-api-access-94cf6\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.945087 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-client-ca\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.945293 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd5faea-8284-4164-8383-72800aea2048-serving-cert\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.945380 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-config\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.945464 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-config" (OuterVolumeSpecName: "config") pod "616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" (UID: "616ed7b7-eb6d-4079-bb20-5e797ab5dfc0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.946231 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" (UID: "616ed7b7-eb6d-4079-bb20-5e797ab5dfc0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.946323 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-client-ca" (OuterVolumeSpecName: "client-ca") pod "616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" (UID: "616ed7b7-eb6d-4079-bb20-5e797ab5dfc0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.950967 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-kube-api-access-kbcn5" (OuterVolumeSpecName: "kube-api-access-kbcn5") pod "616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" (UID: "616ed7b7-eb6d-4079-bb20-5e797ab5dfc0"). InnerVolumeSpecName "kube-api-access-kbcn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.956182 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" (UID: "616ed7b7-eb6d-4079-bb20-5e797ab5dfc0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.976910 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" event={"ID":"616ed7b7-eb6d-4079-bb20-5e797ab5dfc0","Type":"ContainerDied","Data":"2ac12a83a0e0511826497ed7b88008799f5a3abdbb59f49267febe2b032f8c93"} Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.976978 4978 scope.go:117] "RemoveContainer" containerID="8c322ccfbfb9a156c5e69fb9295d78bd840619315b5a55bb00933e772c646b17" Feb 25 06:47:37 crc kubenswrapper[4978]: I0225 06:47:37.976987 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77b586cd9d-dlsrs" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.018261 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-77b586cd9d-dlsrs"] Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.021478 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-77b586cd9d-dlsrs"] Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.046092 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-client-ca\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.046172 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd5faea-8284-4164-8383-72800aea2048-serving-cert\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.046213 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-config\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.046265 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-proxy-ca-bundles\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.046304 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94cf6\" (UniqueName: \"kubernetes.io/projected/ccd5faea-8284-4164-8383-72800aea2048-kube-api-access-94cf6\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.046354 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.046423 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.046438 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.046450 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbcn5\" (UniqueName: \"kubernetes.io/projected/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-kube-api-access-kbcn5\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.046462 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.048507 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-config\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.048601 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-client-ca\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.048825 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-proxy-ca-bundles\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.052312 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd5faea-8284-4164-8383-72800aea2048-serving-cert\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.063487 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94cf6\" (UniqueName: \"kubernetes.io/projected/ccd5faea-8284-4164-8383-72800aea2048-kube-api-access-94cf6\") pod \"controller-manager-6686fbcbd7-4zltb\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.156245 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:47:38 crc kubenswrapper[4978]: E0225 06:47:38.307577 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7cb83e77_c92b_4736_91f2_61c8a1241eb0.slice/crio-conmon-95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b.scope\": RecentStats: unable to find data in memory cache]" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.986556 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-mhnd6_7cb83e77-c92b-4736-91f2-61c8a1241eb0/kube-multus-additional-cni-plugins/0.log" Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.986605 4978 generic.go:334] "Generic (PLEG): container finished" podID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" exitCode=137 Feb 25 06:47:38 crc kubenswrapper[4978]: I0225 06:47:38.986635 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" event={"ID":"7cb83e77-c92b-4736-91f2-61c8a1241eb0","Type":"ContainerDied","Data":"95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b"} Feb 25 06:47:39 crc kubenswrapper[4978]: I0225 06:47:39.340817 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="616ed7b7-eb6d-4079-bb20-5e797ab5dfc0" path="/var/lib/kubelet/pods/616ed7b7-eb6d-4079-bb20-5e797ab5dfc0/volumes" Feb 25 06:47:42 crc kubenswrapper[4978]: E0225 06:47:42.439598 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:42 crc kubenswrapper[4978]: E0225 06:47:42.440462 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:42 crc kubenswrapper[4978]: E0225 06:47:42.441015 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:42 crc kubenswrapper[4978]: E0225 06:47:42.441116 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" Feb 25 06:47:42 crc kubenswrapper[4978]: I0225 06:47:42.472698 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 25 06:47:43 crc kubenswrapper[4978]: I0225 06:47:43.144323 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fvmbt" Feb 25 06:47:43 crc kubenswrapper[4978]: I0225 06:47:43.839570 4978 patch_prober.go:28] interesting pod/route-controller-manager-78b646bc84-xjqmd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 06:47:43 crc kubenswrapper[4978]: I0225 06:47:43.839654 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" podUID="d241c721-1ac8-4bda-93cb-ae761f0075d5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.548015 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.550836 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.559502 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.559612 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.567566 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.576571 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62081d94-2fb1-481a-b45f-abd2267878fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"62081d94-2fb1-481a-b45f-abd2267878fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.576698 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62081d94-2fb1-481a-b45f-abd2267878fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"62081d94-2fb1-481a-b45f-abd2267878fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.677853 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62081d94-2fb1-481a-b45f-abd2267878fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"62081d94-2fb1-481a-b45f-abd2267878fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.677944 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62081d94-2fb1-481a-b45f-abd2267878fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"62081d94-2fb1-481a-b45f-abd2267878fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.678139 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62081d94-2fb1-481a-b45f-abd2267878fb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"62081d94-2fb1-481a-b45f-abd2267878fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.708677 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62081d94-2fb1-481a-b45f-abd2267878fb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"62081d94-2fb1-481a-b45f-abd2267878fb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 06:47:46 crc kubenswrapper[4978]: I0225 06:47:46.901774 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 06:47:48 crc kubenswrapper[4978]: I0225 06:47:48.456240 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6686fbcbd7-4zltb"] Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.549506 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.552746 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.565899 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.654916 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.655502 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-var-lock\") pod \"installer-9-crc\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.655772 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7451e170-1aa9-4967-b33e-b8b42707ecc6-kube-api-access\") pod \"installer-9-crc\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.757172 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.757286 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.757585 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-var-lock\") pod \"installer-9-crc\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.757679 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7451e170-1aa9-4967-b33e-b8b42707ecc6-kube-api-access\") pod \"installer-9-crc\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.757830 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-var-lock\") pod \"installer-9-crc\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.780391 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7451e170-1aa9-4967-b33e-b8b42707ecc6-kube-api-access\") pod \"installer-9-crc\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:51 crc kubenswrapper[4978]: I0225 06:47:51.884059 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:47:52 crc kubenswrapper[4978]: E0225 06:47:52.440051 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:52 crc kubenswrapper[4978]: E0225 06:47:52.440722 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:52 crc kubenswrapper[4978]: E0225 06:47:52.441314 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:47:52 crc kubenswrapper[4978]: E0225 06:47:52.441358 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" Feb 25 06:47:53 crc kubenswrapper[4978]: I0225 06:47:53.838319 4978 patch_prober.go:28] interesting pod/route-controller-manager-78b646bc84-xjqmd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 06:47:53 crc kubenswrapper[4978]: I0225 06:47:53.838458 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" podUID="d241c721-1ac8-4bda-93cb-ae761f0075d5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 06:47:55 crc kubenswrapper[4978]: E0225 06:47:55.821176 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 25 06:47:55 crc kubenswrapper[4978]: E0225 06:47:55.821454 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ctx5l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-cghmv_openshift-marketplace(6f6ed9c0-cb59-4aac-af72-2633f18378c9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 06:47:55 crc kubenswrapper[4978]: E0225 06:47:55.823551 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-cghmv" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" Feb 25 06:47:57 crc kubenswrapper[4978]: E0225 06:47:57.039716 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-cghmv" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" Feb 25 06:47:59 crc kubenswrapper[4978]: E0225 06:47:59.909141 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 25 06:47:59 crc kubenswrapper[4978]: E0225 06:47:59.909819 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tr85c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-w9qqz_openshift-marketplace(ca2405db-a877-4fda-a24c-ff39a0069758): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 06:47:59 crc kubenswrapper[4978]: E0225 06:47:59.911357 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-w9qqz" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" Feb 25 06:48:00 crc kubenswrapper[4978]: I0225 06:48:00.141737 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533368-gn4dn"] Feb 25 06:48:00 crc kubenswrapper[4978]: I0225 06:48:00.143400 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533368-gn4dn" Feb 25 06:48:00 crc kubenswrapper[4978]: I0225 06:48:00.148915 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 06:48:00 crc kubenswrapper[4978]: I0225 06:48:00.150052 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 06:48:00 crc kubenswrapper[4978]: I0225 06:48:00.150308 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 06:48:00 crc kubenswrapper[4978]: I0225 06:48:00.151685 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533368-gn4dn"] Feb 25 06:48:00 crc kubenswrapper[4978]: I0225 06:48:00.286564 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9wbd\" (UniqueName: \"kubernetes.io/projected/dec7193e-a91c-45b7-a8b1-236d626f0540-kube-api-access-v9wbd\") pod \"auto-csr-approver-29533368-gn4dn\" (UID: \"dec7193e-a91c-45b7-a8b1-236d626f0540\") " pod="openshift-infra/auto-csr-approver-29533368-gn4dn" Feb 25 06:48:00 crc kubenswrapper[4978]: I0225 06:48:00.388447 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9wbd\" (UniqueName: \"kubernetes.io/projected/dec7193e-a91c-45b7-a8b1-236d626f0540-kube-api-access-v9wbd\") pod \"auto-csr-approver-29533368-gn4dn\" (UID: \"dec7193e-a91c-45b7-a8b1-236d626f0540\") " pod="openshift-infra/auto-csr-approver-29533368-gn4dn" Feb 25 06:48:00 crc kubenswrapper[4978]: I0225 06:48:00.409242 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9wbd\" (UniqueName: \"kubernetes.io/projected/dec7193e-a91c-45b7-a8b1-236d626f0540-kube-api-access-v9wbd\") pod \"auto-csr-approver-29533368-gn4dn\" (UID: \"dec7193e-a91c-45b7-a8b1-236d626f0540\") " pod="openshift-infra/auto-csr-approver-29533368-gn4dn" Feb 25 06:48:00 crc kubenswrapper[4978]: I0225 06:48:00.471899 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533368-gn4dn" Feb 25 06:48:02 crc kubenswrapper[4978]: E0225 06:48:02.440271 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:48:02 crc kubenswrapper[4978]: E0225 06:48:02.441478 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:48:02 crc kubenswrapper[4978]: E0225 06:48:02.441966 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:48:02 crc kubenswrapper[4978]: E0225 06:48:02.442238 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" Feb 25 06:48:03 crc kubenswrapper[4978]: I0225 06:48:03.839614 4978 patch_prober.go:28] interesting pod/route-controller-manager-78b646bc84-xjqmd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 06:48:03 crc kubenswrapper[4978]: I0225 06:48:03.840677 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" podUID="d241c721-1ac8-4bda-93cb-ae761f0075d5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 06:48:05 crc kubenswrapper[4978]: E0225 06:48:05.942634 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-w9qqz" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" Feb 25 06:48:06 crc kubenswrapper[4978]: E0225 06:48:06.827623 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 25 06:48:06 crc kubenswrapper[4978]: E0225 06:48:06.828207 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gvlxt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-6fbkc_openshift-marketplace(3054616f-87b0-467e-b5a7-499e3addf68a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 06:48:06 crc kubenswrapper[4978]: E0225 06:48:06.829532 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-6fbkc" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" Feb 25 06:48:07 crc kubenswrapper[4978]: E0225 06:48:07.396823 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 25 06:48:07 crc kubenswrapper[4978]: E0225 06:48:07.397005 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8w2qq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-v9fsk_openshift-marketplace(8ae5f061-fdff-4fc5-b252-4734b7ad892a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 06:48:07 crc kubenswrapper[4978]: E0225 06:48:07.398550 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-v9fsk" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" Feb 25 06:48:12 crc kubenswrapper[4978]: E0225 06:48:12.440283 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:48:12 crc kubenswrapper[4978]: E0225 06:48:12.442476 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:48:12 crc kubenswrapper[4978]: E0225 06:48:12.443204 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 25 06:48:12 crc kubenswrapper[4978]: E0225 06:48:12.443324 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" Feb 25 06:48:13 crc kubenswrapper[4978]: E0225 06:48:13.754502 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-v9fsk" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" Feb 25 06:48:13 crc kubenswrapper[4978]: E0225 06:48:13.754586 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-6fbkc" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" Feb 25 06:48:13 crc kubenswrapper[4978]: I0225 06:48:13.838919 4978 patch_prober.go:28] interesting pod/route-controller-manager-78b646bc84-xjqmd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 06:48:13 crc kubenswrapper[4978]: I0225 06:48:13.839045 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" podUID="d241c721-1ac8-4bda-93cb-ae761f0075d5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 25 06:48:13 crc kubenswrapper[4978]: I0225 06:48:13.858915 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:48:13 crc kubenswrapper[4978]: I0225 06:48:13.898904 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns"] Feb 25 06:48:13 crc kubenswrapper[4978]: E0225 06:48:13.899495 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d241c721-1ac8-4bda-93cb-ae761f0075d5" containerName="route-controller-manager" Feb 25 06:48:13 crc kubenswrapper[4978]: I0225 06:48:13.899515 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d241c721-1ac8-4bda-93cb-ae761f0075d5" containerName="route-controller-manager" Feb 25 06:48:13 crc kubenswrapper[4978]: I0225 06:48:13.899650 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d241c721-1ac8-4bda-93cb-ae761f0075d5" containerName="route-controller-manager" Feb 25 06:48:13 crc kubenswrapper[4978]: I0225 06:48:13.902541 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:13 crc kubenswrapper[4978]: I0225 06:48:13.904045 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns"] Feb 25 06:48:13 crc kubenswrapper[4978]: E0225 06:48:13.924859 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 25 06:48:13 crc kubenswrapper[4978]: E0225 06:48:13.925235 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-95x8l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tjqjf_openshift-marketplace(8178b540-4bca-402c-a2e0-57da7a09fa30): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 06:48:13 crc kubenswrapper[4978]: E0225 06:48:13.927582 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tjqjf" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" Feb 25 06:48:13 crc kubenswrapper[4978]: E0225 06:48:13.952020 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 25 06:48:13 crc kubenswrapper[4978]: E0225 06:48:13.952327 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gvwwx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-h9wx2_openshift-marketplace(9283fdd2-ab8a-48ca-a326-1a0b107671a0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 06:48:13 crc kubenswrapper[4978]: E0225 06:48:13.953634 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-h9wx2" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" Feb 25 06:48:13 crc kubenswrapper[4978]: I0225 06:48:13.968190 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-mhnd6_7cb83e77-c92b-4736-91f2-61c8a1241eb0/kube-multus-additional-cni-plugins/0.log" Feb 25 06:48:13 crc kubenswrapper[4978]: I0225 06:48:13.968255 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.015609 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-client-ca\") pod \"d241c721-1ac8-4bda-93cb-ae761f0075d5\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.016123 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsl92\" (UniqueName: \"kubernetes.io/projected/d241c721-1ac8-4bda-93cb-ae761f0075d5-kube-api-access-bsl92\") pod \"d241c721-1ac8-4bda-93cb-ae761f0075d5\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.016194 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-config\") pod \"d241c721-1ac8-4bda-93cb-ae761f0075d5\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.016239 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7cb83e77-c92b-4736-91f2-61c8a1241eb0-tuning-conf-dir\") pod \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.016271 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7cb83e77-c92b-4736-91f2-61c8a1241eb0-cni-sysctl-allowlist\") pod \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.016300 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d241c721-1ac8-4bda-93cb-ae761f0075d5-serving-cert\") pod \"d241c721-1ac8-4bda-93cb-ae761f0075d5\" (UID: \"d241c721-1ac8-4bda-93cb-ae761f0075d5\") " Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.016754 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-client-ca" (OuterVolumeSpecName: "client-ca") pod "d241c721-1ac8-4bda-93cb-ae761f0075d5" (UID: "d241c721-1ac8-4bda-93cb-ae761f0075d5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.016983 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-serving-cert\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.017030 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n8kp\" (UniqueName: \"kubernetes.io/projected/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-kube-api-access-6n8kp\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.017088 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-config\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.017134 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-client-ca\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.017178 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.017203 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-config" (OuterVolumeSpecName: "config") pod "d241c721-1ac8-4bda-93cb-ae761f0075d5" (UID: "d241c721-1ac8-4bda-93cb-ae761f0075d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.017272 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7cb83e77-c92b-4736-91f2-61c8a1241eb0-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "7cb83e77-c92b-4736-91f2-61c8a1241eb0" (UID: "7cb83e77-c92b-4736-91f2-61c8a1241eb0"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.018789 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb83e77-c92b-4736-91f2-61c8a1241eb0-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7cb83e77-c92b-4736-91f2-61c8a1241eb0" (UID: "7cb83e77-c92b-4736-91f2-61c8a1241eb0"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:14 crc kubenswrapper[4978]: E0225 06:48:14.024191 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 25 06:48:14 crc kubenswrapper[4978]: E0225 06:48:14.024351 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8xnlb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-h7dbj_openshift-marketplace(6a76f909-3198-47cd-adbb-9f94fb99ab48): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.024487 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d241c721-1ac8-4bda-93cb-ae761f0075d5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d241c721-1ac8-4bda-93cb-ae761f0075d5" (UID: "d241c721-1ac8-4bda-93cb-ae761f0075d5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:48:14 crc kubenswrapper[4978]: E0225 06:48:14.025571 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-h7dbj" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.028524 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d241c721-1ac8-4bda-93cb-ae761f0075d5-kube-api-access-bsl92" (OuterVolumeSpecName: "kube-api-access-bsl92") pod "d241c721-1ac8-4bda-93cb-ae761f0075d5" (UID: "d241c721-1ac8-4bda-93cb-ae761f0075d5"). InnerVolumeSpecName "kube-api-access-bsl92". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:14 crc kubenswrapper[4978]: E0225 06:48:14.070020 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 25 06:48:14 crc kubenswrapper[4978]: E0225 06:48:14.070185 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hxzv2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zzswn_openshift-marketplace(d6342947-b76f-4c86-8e34-d42a780c6bf5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 25 06:48:14 crc kubenswrapper[4978]: E0225 06:48:14.071409 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zzswn" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.117845 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/7cb83e77-c92b-4736-91f2-61c8a1241eb0-ready\") pod \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.117970 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwhtv\" (UniqueName: \"kubernetes.io/projected/7cb83e77-c92b-4736-91f2-61c8a1241eb0-kube-api-access-gwhtv\") pod \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\" (UID: \"7cb83e77-c92b-4736-91f2-61c8a1241eb0\") " Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.118218 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-config\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.118265 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-client-ca\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.118322 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-serving-cert\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.118351 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n8kp\" (UniqueName: \"kubernetes.io/projected/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-kube-api-access-6n8kp\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.118410 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsl92\" (UniqueName: \"kubernetes.io/projected/d241c721-1ac8-4bda-93cb-ae761f0075d5-kube-api-access-bsl92\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.118423 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d241c721-1ac8-4bda-93cb-ae761f0075d5-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.118434 4978 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7cb83e77-c92b-4736-91f2-61c8a1241eb0-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.118442 4978 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7cb83e77-c92b-4736-91f2-61c8a1241eb0-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.118451 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d241c721-1ac8-4bda-93cb-ae761f0075d5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.118437 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cb83e77-c92b-4736-91f2-61c8a1241eb0-ready" (OuterVolumeSpecName: "ready") pod "7cb83e77-c92b-4736-91f2-61c8a1241eb0" (UID: "7cb83e77-c92b-4736-91f2-61c8a1241eb0"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.119758 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-client-ca\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.119986 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-config\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.124769 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cb83e77-c92b-4736-91f2-61c8a1241eb0-kube-api-access-gwhtv" (OuterVolumeSpecName: "kube-api-access-gwhtv") pod "7cb83e77-c92b-4736-91f2-61c8a1241eb0" (UID: "7cb83e77-c92b-4736-91f2-61c8a1241eb0"). InnerVolumeSpecName "kube-api-access-gwhtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.125133 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-serving-cert\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.136802 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n8kp\" (UniqueName: \"kubernetes.io/projected/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-kube-api-access-6n8kp\") pod \"route-controller-manager-59b44d8784-lfzns\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.219197 4978 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/7cb83e77-c92b-4736-91f2-61c8a1241eb0-ready\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.219232 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwhtv\" (UniqueName: \"kubernetes.io/projected/7cb83e77-c92b-4736-91f2-61c8a1241eb0-kube-api-access-gwhtv\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.231265 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-mhnd6_7cb83e77-c92b-4736-91f2-61c8a1241eb0/kube-multus-additional-cni-plugins/0.log" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.231359 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.231469 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-mhnd6" event={"ID":"7cb83e77-c92b-4736-91f2-61c8a1241eb0","Type":"ContainerDied","Data":"f8e177876a0c2b1fc06cc59b4f7447b215d4265dc51d7be82d320a81b790fa3a"} Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.231545 4978 scope.go:117] "RemoveContainer" containerID="95647acce8a4fe3928f03b6ef88a0ffd469b67c380af238c34a8fba53a7a1f9b" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.234853 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" event={"ID":"d241c721-1ac8-4bda-93cb-ae761f0075d5","Type":"ContainerDied","Data":"17d322426d1d7d25c3e0a8d152f700a347a17ec274184346278704f2ba474c82"} Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.235244 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd" Feb 25 06:48:14 crc kubenswrapper[4978]: E0225 06:48:14.237494 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tjqjf" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" Feb 25 06:48:14 crc kubenswrapper[4978]: E0225 06:48:14.237505 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zzswn" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" Feb 25 06:48:14 crc kubenswrapper[4978]: E0225 06:48:14.237621 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-h7dbj" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" Feb 25 06:48:14 crc kubenswrapper[4978]: E0225 06:48:14.237696 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-h9wx2" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.271155 4978 scope.go:117] "RemoveContainer" containerID="bfd1ff0a503d419ae3deaccdd1c2ec0baafbf653e0292b9c185c5965edcf6226" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.288270 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd"] Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.288944 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.291736 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78b646bc84-xjqmd"] Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.318538 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-mhnd6"] Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.322289 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-mhnd6"] Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.336583 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6686fbcbd7-4zltb"] Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.365542 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533368-gn4dn"] Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.367790 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.432293 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 25 06:48:14 crc kubenswrapper[4978]: I0225 06:48:14.771260 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns"] Feb 25 06:48:14 crc kubenswrapper[4978]: W0225 06:48:14.796045 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7feb6319_f64c_47a4_8090_4dcf75f4d3f7.slice/crio-9cc08e0a56e5647c5cca5b5fbcd08a6088c5f1b4e85709b7c0f63e27e1e9dbea WatchSource:0}: Error finding container 9cc08e0a56e5647c5cca5b5fbcd08a6088c5f1b4e85709b7c0f63e27e1e9dbea: Status 404 returned error can't find the container with id 9cc08e0a56e5647c5cca5b5fbcd08a6088c5f1b4e85709b7c0f63e27e1e9dbea Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.241196 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7451e170-1aa9-4967-b33e-b8b42707ecc6","Type":"ContainerStarted","Data":"df379516bc997db8ed5d0d739adf199d56d26d183e198aabf605118471da47a5"} Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.241864 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7451e170-1aa9-4967-b33e-b8b42707ecc6","Type":"ContainerStarted","Data":"a90098f92d9735dd7148c0954689bec5ab757b83f9230f759001d9180e3d4c4f"} Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.244342 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" event={"ID":"ccd5faea-8284-4164-8383-72800aea2048","Type":"ContainerStarted","Data":"7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6"} Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.244998 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.245112 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" event={"ID":"ccd5faea-8284-4164-8383-72800aea2048","Type":"ContainerStarted","Data":"084f608377aa0f523e6fb53ab8d7aeb8e002fbf3dfb1d71707a12e7fba8cb085"} Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.244966 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" podUID="ccd5faea-8284-4164-8383-72800aea2048" containerName="controller-manager" containerID="cri-o://7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6" gracePeriod=30 Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.250413 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.251674 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" event={"ID":"7feb6319-f64c-47a4-8090-4dcf75f4d3f7","Type":"ContainerStarted","Data":"81fcefb665a61468a87a293b4733f7a494926b838956fa87fc95b6892b281f99"} Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.251730 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" event={"ID":"7feb6319-f64c-47a4-8090-4dcf75f4d3f7","Type":"ContainerStarted","Data":"9cc08e0a56e5647c5cca5b5fbcd08a6088c5f1b4e85709b7c0f63e27e1e9dbea"} Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.252859 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.254275 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"62081d94-2fb1-481a-b45f-abd2267878fb","Type":"ContainerStarted","Data":"3c0e1440a76cca54a3411a9c022d77d117ae7ab146d913dd25ff0bd88c3f4359"} Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.254298 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"62081d94-2fb1-481a-b45f-abd2267878fb","Type":"ContainerStarted","Data":"7c6557489e21c2734ee88527e4b6b946fd90181326bbe6c0b6d2954264289df7"} Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.265636 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=24.265612933 podStartE2EDuration="24.265612933s" podCreationTimestamp="2026-02-25 06:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:48:15.263706593 +0000 UTC m=+188.702963062" watchObservedRunningTime="2026-02-25 06:48:15.265612933 +0000 UTC m=+188.704869392" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.268899 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533368-gn4dn" event={"ID":"dec7193e-a91c-45b7-a8b1-236d626f0540","Type":"ContainerStarted","Data":"2a6647e8be050d2a230e46240cc927aab379647ce5f82f660b0dd8ae50a1d921"} Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.286524 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cghmv" event={"ID":"6f6ed9c0-cb59-4aac-af72-2633f18378c9","Type":"ContainerStarted","Data":"6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09"} Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.335251 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" podStartSLOduration=47.33523362 podStartE2EDuration="47.33523362s" podCreationTimestamp="2026-02-25 06:47:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:48:15.298901654 +0000 UTC m=+188.738158113" watchObservedRunningTime="2026-02-25 06:48:15.33523362 +0000 UTC m=+188.774490069" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.337235 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" podStartSLOduration=27.337226672 podStartE2EDuration="27.337226672s" podCreationTimestamp="2026-02-25 06:47:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:48:15.33398952 +0000 UTC m=+188.773245989" watchObservedRunningTime="2026-02-25 06:48:15.337226672 +0000 UTC m=+188.776483131" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.339836 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" path="/var/lib/kubelet/pods/7cb83e77-c92b-4736-91f2-61c8a1241eb0/volumes" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.341173 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d241c721-1ac8-4bda-93cb-ae761f0075d5" path="/var/lib/kubelet/pods/d241c721-1ac8-4bda-93cb-ae761f0075d5/volumes" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.359058 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.361312 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=29.361293121 podStartE2EDuration="29.361293121s" podCreationTimestamp="2026-02-25 06:47:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:48:15.358847824 +0000 UTC m=+188.798104303" watchObservedRunningTime="2026-02-25 06:48:15.361293121 +0000 UTC m=+188.800549580" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.645444 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.753928 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94cf6\" (UniqueName: \"kubernetes.io/projected/ccd5faea-8284-4164-8383-72800aea2048-kube-api-access-94cf6\") pod \"ccd5faea-8284-4164-8383-72800aea2048\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.754058 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-config\") pod \"ccd5faea-8284-4164-8383-72800aea2048\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.754085 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd5faea-8284-4164-8383-72800aea2048-serving-cert\") pod \"ccd5faea-8284-4164-8383-72800aea2048\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.754115 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-client-ca\") pod \"ccd5faea-8284-4164-8383-72800aea2048\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.754154 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-proxy-ca-bundles\") pod \"ccd5faea-8284-4164-8383-72800aea2048\" (UID: \"ccd5faea-8284-4164-8383-72800aea2048\") " Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.755099 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ccd5faea-8284-4164-8383-72800aea2048" (UID: "ccd5faea-8284-4164-8383-72800aea2048"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.756175 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-config" (OuterVolumeSpecName: "config") pod "ccd5faea-8284-4164-8383-72800aea2048" (UID: "ccd5faea-8284-4164-8383-72800aea2048"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.756196 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-client-ca" (OuterVolumeSpecName: "client-ca") pod "ccd5faea-8284-4164-8383-72800aea2048" (UID: "ccd5faea-8284-4164-8383-72800aea2048"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.760582 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccd5faea-8284-4164-8383-72800aea2048-kube-api-access-94cf6" (OuterVolumeSpecName: "kube-api-access-94cf6") pod "ccd5faea-8284-4164-8383-72800aea2048" (UID: "ccd5faea-8284-4164-8383-72800aea2048"). InnerVolumeSpecName "kube-api-access-94cf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.760915 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd5faea-8284-4164-8383-72800aea2048-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ccd5faea-8284-4164-8383-72800aea2048" (UID: "ccd5faea-8284-4164-8383-72800aea2048"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.856105 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.856358 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94cf6\" (UniqueName: \"kubernetes.io/projected/ccd5faea-8284-4164-8383-72800aea2048-kube-api-access-94cf6\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.856382 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.856390 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccd5faea-8284-4164-8383-72800aea2048-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:15 crc kubenswrapper[4978]: I0225 06:48:15.856398 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccd5faea-8284-4164-8383-72800aea2048-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.295879 4978 generic.go:334] "Generic (PLEG): container finished" podID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerID="6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09" exitCode=0 Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.295970 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cghmv" event={"ID":"6f6ed9c0-cb59-4aac-af72-2633f18378c9","Type":"ContainerDied","Data":"6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09"} Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.297658 4978 generic.go:334] "Generic (PLEG): container finished" podID="ccd5faea-8284-4164-8383-72800aea2048" containerID="7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6" exitCode=0 Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.297708 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.297714 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" event={"ID":"ccd5faea-8284-4164-8383-72800aea2048","Type":"ContainerDied","Data":"7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6"} Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.297739 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6686fbcbd7-4zltb" event={"ID":"ccd5faea-8284-4164-8383-72800aea2048","Type":"ContainerDied","Data":"084f608377aa0f523e6fb53ab8d7aeb8e002fbf3dfb1d71707a12e7fba8cb085"} Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.297755 4978 scope.go:117] "RemoveContainer" containerID="7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.300869 4978 generic.go:334] "Generic (PLEG): container finished" podID="62081d94-2fb1-481a-b45f-abd2267878fb" containerID="3c0e1440a76cca54a3411a9c022d77d117ae7ab146d913dd25ff0bd88c3f4359" exitCode=0 Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.301513 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"62081d94-2fb1-481a-b45f-abd2267878fb","Type":"ContainerDied","Data":"3c0e1440a76cca54a3411a9c022d77d117ae7ab146d913dd25ff0bd88c3f4359"} Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.319687 4978 scope.go:117] "RemoveContainer" containerID="7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6" Feb 25 06:48:16 crc kubenswrapper[4978]: E0225 06:48:16.320302 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6\": container with ID starting with 7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6 not found: ID does not exist" containerID="7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.320339 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6"} err="failed to get container status \"7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6\": rpc error: code = NotFound desc = could not find container \"7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6\": container with ID starting with 7ec404bf38f02b440f6fe323836aa5fdd1c956f3f4a97d4c5953ab3b764480e6 not found: ID does not exist" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.343515 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6686fbcbd7-4zltb"] Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.347002 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6686fbcbd7-4zltb"] Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.503038 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-756b67b69f-5d94g"] Feb 25 06:48:16 crc kubenswrapper[4978]: E0225 06:48:16.503311 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd5faea-8284-4164-8383-72800aea2048" containerName="controller-manager" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.503334 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd5faea-8284-4164-8383-72800aea2048" containerName="controller-manager" Feb 25 06:48:16 crc kubenswrapper[4978]: E0225 06:48:16.503353 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.503362 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.503503 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd5faea-8284-4164-8383-72800aea2048" containerName="controller-manager" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.503534 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cb83e77-c92b-4736-91f2-61c8a1241eb0" containerName="kube-multus-additional-cni-plugins" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.503997 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.506592 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.506804 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.506941 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.507414 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.507918 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.507936 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.516197 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.518566 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-756b67b69f-5d94g"] Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.564026 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-config\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.564076 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/652ab496-ffe4-4e83-8cf4-9da4516b559a-serving-cert\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.564123 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-client-ca\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.564141 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-proxy-ca-bundles\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.564162 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsv8h\" (UniqueName: \"kubernetes.io/projected/652ab496-ffe4-4e83-8cf4-9da4516b559a-kube-api-access-vsv8h\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.665059 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-config\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.665520 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/652ab496-ffe4-4e83-8cf4-9da4516b559a-serving-cert\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.665664 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-client-ca\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.665684 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-proxy-ca-bundles\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.665719 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsv8h\" (UniqueName: \"kubernetes.io/projected/652ab496-ffe4-4e83-8cf4-9da4516b559a-kube-api-access-vsv8h\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.666705 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-client-ca\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.667010 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-config\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.667471 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-proxy-ca-bundles\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.670711 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/652ab496-ffe4-4e83-8cf4-9da4516b559a-serving-cert\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.684166 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsv8h\" (UniqueName: \"kubernetes.io/projected/652ab496-ffe4-4e83-8cf4-9da4516b559a-kube-api-access-vsv8h\") pod \"controller-manager-756b67b69f-5d94g\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:16 crc kubenswrapper[4978]: I0225 06:48:16.828975 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:17 crc kubenswrapper[4978]: I0225 06:48:17.018330 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-756b67b69f-5d94g"] Feb 25 06:48:17 crc kubenswrapper[4978]: I0225 06:48:17.312710 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cghmv" event={"ID":"6f6ed9c0-cb59-4aac-af72-2633f18378c9","Type":"ContainerStarted","Data":"157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9"} Feb 25 06:48:17 crc kubenswrapper[4978]: I0225 06:48:17.333676 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cghmv" podStartSLOduration=3.405438919 podStartE2EDuration="1m4.333648041s" podCreationTimestamp="2026-02-25 06:47:13 +0000 UTC" firstStartedPulling="2026-02-25 06:47:15.798031778 +0000 UTC m=+129.237288237" lastFinishedPulling="2026-02-25 06:48:16.7262409 +0000 UTC m=+190.165497359" observedRunningTime="2026-02-25 06:48:17.330808171 +0000 UTC m=+190.770064640" watchObservedRunningTime="2026-02-25 06:48:17.333648041 +0000 UTC m=+190.772904520" Feb 25 06:48:17 crc kubenswrapper[4978]: I0225 06:48:17.340968 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccd5faea-8284-4164-8383-72800aea2048" path="/var/lib/kubelet/pods/ccd5faea-8284-4164-8383-72800aea2048/volumes" Feb 25 06:48:20 crc kubenswrapper[4978]: W0225 06:48:20.043870 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod652ab496_ffe4_4e83_8cf4_9da4516b559a.slice/crio-8e1ecee9bb876be84911c6434724bc22b2d9d5e2300b730ad789efe2755e2c47 WatchSource:0}: Error finding container 8e1ecee9bb876be84911c6434724bc22b2d9d5e2300b730ad789efe2755e2c47: Status 404 returned error can't find the container with id 8e1ecee9bb876be84911c6434724bc22b2d9d5e2300b730ad789efe2755e2c47 Feb 25 06:48:20 crc kubenswrapper[4978]: I0225 06:48:20.335429 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" event={"ID":"652ab496-ffe4-4e83-8cf4-9da4516b559a","Type":"ContainerStarted","Data":"8e1ecee9bb876be84911c6434724bc22b2d9d5e2300b730ad789efe2755e2c47"} Feb 25 06:48:21 crc kubenswrapper[4978]: I0225 06:48:21.705884 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 06:48:21 crc kubenswrapper[4978]: I0225 06:48:21.862020 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62081d94-2fb1-481a-b45f-abd2267878fb-kube-api-access\") pod \"62081d94-2fb1-481a-b45f-abd2267878fb\" (UID: \"62081d94-2fb1-481a-b45f-abd2267878fb\") " Feb 25 06:48:21 crc kubenswrapper[4978]: I0225 06:48:21.862183 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62081d94-2fb1-481a-b45f-abd2267878fb-kubelet-dir\") pod \"62081d94-2fb1-481a-b45f-abd2267878fb\" (UID: \"62081d94-2fb1-481a-b45f-abd2267878fb\") " Feb 25 06:48:21 crc kubenswrapper[4978]: I0225 06:48:21.862393 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62081d94-2fb1-481a-b45f-abd2267878fb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "62081d94-2fb1-481a-b45f-abd2267878fb" (UID: "62081d94-2fb1-481a-b45f-abd2267878fb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:48:21 crc kubenswrapper[4978]: I0225 06:48:21.863187 4978 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62081d94-2fb1-481a-b45f-abd2267878fb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:21 crc kubenswrapper[4978]: I0225 06:48:21.870927 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62081d94-2fb1-481a-b45f-abd2267878fb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "62081d94-2fb1-481a-b45f-abd2267878fb" (UID: "62081d94-2fb1-481a-b45f-abd2267878fb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:21 crc kubenswrapper[4978]: I0225 06:48:21.965297 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62081d94-2fb1-481a-b45f-abd2267878fb-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.346559 4978 generic.go:334] "Generic (PLEG): container finished" podID="ca2405db-a877-4fda-a24c-ff39a0069758" containerID="0a7d15a3e54f9c3eac29e05a04be57324d0ce2ec03b144595900d48eabb28695" exitCode=0 Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.346772 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9qqz" event={"ID":"ca2405db-a877-4fda-a24c-ff39a0069758","Type":"ContainerDied","Data":"0a7d15a3e54f9c3eac29e05a04be57324d0ce2ec03b144595900d48eabb28695"} Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.349193 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" event={"ID":"652ab496-ffe4-4e83-8cf4-9da4516b559a","Type":"ContainerStarted","Data":"96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32"} Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.349407 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.351685 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"62081d94-2fb1-481a-b45f-abd2267878fb","Type":"ContainerDied","Data":"7c6557489e21c2734ee88527e4b6b946fd90181326bbe6c0b6d2954264289df7"} Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.351721 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c6557489e21c2734ee88527e4b6b946fd90181326bbe6c0b6d2954264289df7" Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.351724 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.356081 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.358847 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533368-gn4dn" event={"ID":"dec7193e-a91c-45b7-a8b1-236d626f0540","Type":"ContainerStarted","Data":"b4d31aab45082a45f8ae53c881d441c93ca3611ac7427910de1c0f13ced0d0d1"} Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.390556 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" podStartSLOduration=34.390536932 podStartE2EDuration="34.390536932s" podCreationTimestamp="2026-02-25 06:47:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:48:22.38855365 +0000 UTC m=+195.827810119" watchObservedRunningTime="2026-02-25 06:48:22.390536932 +0000 UTC m=+195.829793401" Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.411048 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533368-gn4dn" podStartSLOduration=15.182587424 podStartE2EDuration="22.411011478s" podCreationTimestamp="2026-02-25 06:48:00 +0000 UTC" firstStartedPulling="2026-02-25 06:48:14.395627179 +0000 UTC m=+187.834883638" lastFinishedPulling="2026-02-25 06:48:21.624051233 +0000 UTC m=+195.063307692" observedRunningTime="2026-02-25 06:48:22.408415256 +0000 UTC m=+195.847671725" watchObservedRunningTime="2026-02-25 06:48:22.411011478 +0000 UTC m=+195.850267977" Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.481465 4978 csr.go:261] certificate signing request csr-djhck is approved, waiting to be issued Feb 25 06:48:22 crc kubenswrapper[4978]: I0225 06:48:22.487660 4978 csr.go:257] certificate signing request csr-djhck is issued Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.364655 4978 generic.go:334] "Generic (PLEG): container finished" podID="dec7193e-a91c-45b7-a8b1-236d626f0540" containerID="b4d31aab45082a45f8ae53c881d441c93ca3611ac7427910de1c0f13ced0d0d1" exitCode=0 Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.364770 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533368-gn4dn" event={"ID":"dec7193e-a91c-45b7-a8b1-236d626f0540","Type":"ContainerDied","Data":"b4d31aab45082a45f8ae53c881d441c93ca3611ac7427910de1c0f13ced0d0d1"} Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.367532 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9qqz" event={"ID":"ca2405db-a877-4fda-a24c-ff39a0069758","Type":"ContainerStarted","Data":"a2e224d8f3961151a0c3cc30c6499394c70fdf59e1a19502e385117e36754236"} Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.401273 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w9qqz" podStartSLOduration=4.420400095 podStartE2EDuration="1m11.40125336s" podCreationTimestamp="2026-02-25 06:47:12 +0000 UTC" firstStartedPulling="2026-02-25 06:47:15.789691205 +0000 UTC m=+129.228947664" lastFinishedPulling="2026-02-25 06:48:22.77054447 +0000 UTC m=+196.209800929" observedRunningTime="2026-02-25 06:48:23.395099981 +0000 UTC m=+196.834356450" watchObservedRunningTime="2026-02-25 06:48:23.40125336 +0000 UTC m=+196.840509819" Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.488522 4978 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-28 20:52:19.500841711 +0000 UTC Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.488576 4978 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7358h3m56.012268476s for next certificate rotation Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.809352 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.809839 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.852776 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.853213 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:48:23 crc kubenswrapper[4978]: I0225 06:48:23.946569 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:48:24 crc kubenswrapper[4978]: I0225 06:48:24.422039 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:48:24 crc kubenswrapper[4978]: I0225 06:48:24.490971 4978 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-29 21:32:57.557334659 +0000 UTC Feb 25 06:48:24 crc kubenswrapper[4978]: I0225 06:48:24.491021 4978 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7382h44m33.066316488s for next certificate rotation Feb 25 06:48:24 crc kubenswrapper[4978]: I0225 06:48:24.650445 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533368-gn4dn" Feb 25 06:48:24 crc kubenswrapper[4978]: I0225 06:48:24.706330 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9wbd\" (UniqueName: \"kubernetes.io/projected/dec7193e-a91c-45b7-a8b1-236d626f0540-kube-api-access-v9wbd\") pod \"dec7193e-a91c-45b7-a8b1-236d626f0540\" (UID: \"dec7193e-a91c-45b7-a8b1-236d626f0540\") " Feb 25 06:48:24 crc kubenswrapper[4978]: I0225 06:48:24.717982 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dec7193e-a91c-45b7-a8b1-236d626f0540-kube-api-access-v9wbd" (OuterVolumeSpecName: "kube-api-access-v9wbd") pod "dec7193e-a91c-45b7-a8b1-236d626f0540" (UID: "dec7193e-a91c-45b7-a8b1-236d626f0540"). InnerVolumeSpecName "kube-api-access-v9wbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:24 crc kubenswrapper[4978]: I0225 06:48:24.808803 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9wbd\" (UniqueName: \"kubernetes.io/projected/dec7193e-a91c-45b7-a8b1-236d626f0540-kube-api-access-v9wbd\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:24 crc kubenswrapper[4978]: I0225 06:48:24.936519 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-w9qqz" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" containerName="registry-server" probeResult="failure" output=< Feb 25 06:48:24 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 06:48:24 crc kubenswrapper[4978]: > Feb 25 06:48:25 crc kubenswrapper[4978]: I0225 06:48:25.382308 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533368-gn4dn" event={"ID":"dec7193e-a91c-45b7-a8b1-236d626f0540","Type":"ContainerDied","Data":"2a6647e8be050d2a230e46240cc927aab379647ce5f82f660b0dd8ae50a1d921"} Feb 25 06:48:25 crc kubenswrapper[4978]: I0225 06:48:25.382354 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a6647e8be050d2a230e46240cc927aab379647ce5f82f660b0dd8ae50a1d921" Feb 25 06:48:25 crc kubenswrapper[4978]: I0225 06:48:25.382592 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533368-gn4dn" Feb 25 06:48:31 crc kubenswrapper[4978]: I0225 06:48:31.419850 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzswn" event={"ID":"d6342947-b76f-4c86-8e34-d42a780c6bf5","Type":"ContainerStarted","Data":"60263caf6fc0e64898c66ea8ebe25b84631be5691d76916c94d6d5d87b486834"} Feb 25 06:48:31 crc kubenswrapper[4978]: I0225 06:48:31.423142 4978 generic.go:334] "Generic (PLEG): container finished" podID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerID="271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c" exitCode=0 Feb 25 06:48:31 crc kubenswrapper[4978]: I0225 06:48:31.423228 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9fsk" event={"ID":"8ae5f061-fdff-4fc5-b252-4734b7ad892a","Type":"ContainerDied","Data":"271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c"} Feb 25 06:48:31 crc kubenswrapper[4978]: I0225 06:48:31.426041 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tjqjf" event={"ID":"8178b540-4bca-402c-a2e0-57da7a09fa30","Type":"ContainerStarted","Data":"2ffdbfb1ce9eb22010ad85731983cdcc58587ab8da0f95eb4107368f5e0f3bbc"} Feb 25 06:48:31 crc kubenswrapper[4978]: I0225 06:48:31.429225 4978 generic.go:334] "Generic (PLEG): container finished" podID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerID="cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd" exitCode=0 Feb 25 06:48:31 crc kubenswrapper[4978]: I0225 06:48:31.429604 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9wx2" event={"ID":"9283fdd2-ab8a-48ca-a326-1a0b107671a0","Type":"ContainerDied","Data":"cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd"} Feb 25 06:48:31 crc kubenswrapper[4978]: I0225 06:48:31.433102 4978 generic.go:334] "Generic (PLEG): container finished" podID="3054616f-87b0-467e-b5a7-499e3addf68a" containerID="0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a" exitCode=0 Feb 25 06:48:31 crc kubenswrapper[4978]: I0225 06:48:31.433176 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fbkc" event={"ID":"3054616f-87b0-467e-b5a7-499e3addf68a","Type":"ContainerDied","Data":"0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a"} Feb 25 06:48:31 crc kubenswrapper[4978]: I0225 06:48:31.435593 4978 generic.go:334] "Generic (PLEG): container finished" podID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerID="f4003d9ca43f946dbfeece01826fe402cbff432cc55930e7677389473783011a" exitCode=0 Feb 25 06:48:31 crc kubenswrapper[4978]: I0225 06:48:31.435626 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7dbj" event={"ID":"6a76f909-3198-47cd-adbb-9f94fb99ab48","Type":"ContainerDied","Data":"f4003d9ca43f946dbfeece01826fe402cbff432cc55930e7677389473783011a"} Feb 25 06:48:32 crc kubenswrapper[4978]: I0225 06:48:32.443321 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7dbj" event={"ID":"6a76f909-3198-47cd-adbb-9f94fb99ab48","Type":"ContainerStarted","Data":"7be8c9231564fad9f8b2c31a2e22639620182af3243c73c8fcc13a6d139914ad"} Feb 25 06:48:32 crc kubenswrapper[4978]: I0225 06:48:32.445448 4978 generic.go:334] "Generic (PLEG): container finished" podID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerID="60263caf6fc0e64898c66ea8ebe25b84631be5691d76916c94d6d5d87b486834" exitCode=0 Feb 25 06:48:32 crc kubenswrapper[4978]: I0225 06:48:32.445484 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzswn" event={"ID":"d6342947-b76f-4c86-8e34-d42a780c6bf5","Type":"ContainerDied","Data":"60263caf6fc0e64898c66ea8ebe25b84631be5691d76916c94d6d5d87b486834"} Feb 25 06:48:32 crc kubenswrapper[4978]: I0225 06:48:32.447308 4978 generic.go:334] "Generic (PLEG): container finished" podID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerID="2ffdbfb1ce9eb22010ad85731983cdcc58587ab8da0f95eb4107368f5e0f3bbc" exitCode=0 Feb 25 06:48:32 crc kubenswrapper[4978]: I0225 06:48:32.447395 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tjqjf" event={"ID":"8178b540-4bca-402c-a2e0-57da7a09fa30","Type":"ContainerDied","Data":"2ffdbfb1ce9eb22010ad85731983cdcc58587ab8da0f95eb4107368f5e0f3bbc"} Feb 25 06:48:32 crc kubenswrapper[4978]: I0225 06:48:32.449605 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9wx2" event={"ID":"9283fdd2-ab8a-48ca-a326-1a0b107671a0","Type":"ContainerStarted","Data":"84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31"} Feb 25 06:48:32 crc kubenswrapper[4978]: I0225 06:48:32.452288 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fbkc" event={"ID":"3054616f-87b0-467e-b5a7-499e3addf68a","Type":"ContainerStarted","Data":"ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b"} Feb 25 06:48:32 crc kubenswrapper[4978]: I0225 06:48:32.462544 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h7dbj" podStartSLOduration=2.84977869 podStartE2EDuration="1m22.462523727s" podCreationTimestamp="2026-02-25 06:47:10 +0000 UTC" firstStartedPulling="2026-02-25 06:47:12.511614137 +0000 UTC m=+125.950870596" lastFinishedPulling="2026-02-25 06:48:32.124359164 +0000 UTC m=+205.563615633" observedRunningTime="2026-02-25 06:48:32.460626688 +0000 UTC m=+205.899883157" watchObservedRunningTime="2026-02-25 06:48:32.462523727 +0000 UTC m=+205.901780186" Feb 25 06:48:32 crc kubenswrapper[4978]: I0225 06:48:32.475809 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h9wx2" podStartSLOduration=3.039976107 podStartE2EDuration="1m22.475797021s" podCreationTimestamp="2026-02-25 06:47:10 +0000 UTC" firstStartedPulling="2026-02-25 06:47:12.544387521 +0000 UTC m=+125.983643980" lastFinishedPulling="2026-02-25 06:48:31.980208435 +0000 UTC m=+205.419464894" observedRunningTime="2026-02-25 06:48:32.474124352 +0000 UTC m=+205.913380811" watchObservedRunningTime="2026-02-25 06:48:32.475797021 +0000 UTC m=+205.915053480" Feb 25 06:48:32 crc kubenswrapper[4978]: I0225 06:48:32.515397 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6fbkc" podStartSLOduration=3.276293017 podStartE2EDuration="1m22.515360147s" podCreationTimestamp="2026-02-25 06:47:10 +0000 UTC" firstStartedPulling="2026-02-25 06:47:12.575195696 +0000 UTC m=+126.014452155" lastFinishedPulling="2026-02-25 06:48:31.814262826 +0000 UTC m=+205.253519285" observedRunningTime="2026-02-25 06:48:32.513965107 +0000 UTC m=+205.953221576" watchObservedRunningTime="2026-02-25 06:48:32.515360147 +0000 UTC m=+205.954616606" Feb 25 06:48:33 crc kubenswrapper[4978]: I0225 06:48:33.462620 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzswn" event={"ID":"d6342947-b76f-4c86-8e34-d42a780c6bf5","Type":"ContainerStarted","Data":"5c192f9f8a9f9f6953443a655ff78b96a0c437a957da0fb51aac12bfc3693353"} Feb 25 06:48:33 crc kubenswrapper[4978]: I0225 06:48:33.466217 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9fsk" event={"ID":"8ae5f061-fdff-4fc5-b252-4734b7ad892a","Type":"ContainerStarted","Data":"f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8"} Feb 25 06:48:33 crc kubenswrapper[4978]: I0225 06:48:33.503451 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zzswn" podStartSLOduration=3.241939034 podStartE2EDuration="1m20.503432528s" podCreationTimestamp="2026-02-25 06:47:13 +0000 UTC" firstStartedPulling="2026-02-25 06:47:15.761136028 +0000 UTC m=+129.200392487" lastFinishedPulling="2026-02-25 06:48:33.022629522 +0000 UTC m=+206.461885981" observedRunningTime="2026-02-25 06:48:33.501998818 +0000 UTC m=+206.941255277" watchObservedRunningTime="2026-02-25 06:48:33.503432528 +0000 UTC m=+206.942688987" Feb 25 06:48:33 crc kubenswrapper[4978]: I0225 06:48:33.523883 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v9fsk" podStartSLOduration=3.736325993 podStartE2EDuration="1m21.52386062s" podCreationTimestamp="2026-02-25 06:47:12 +0000 UTC" firstStartedPulling="2026-02-25 06:47:14.753633994 +0000 UTC m=+128.192890453" lastFinishedPulling="2026-02-25 06:48:32.541168621 +0000 UTC m=+205.980425080" observedRunningTime="2026-02-25 06:48:33.521707332 +0000 UTC m=+206.960963811" watchObservedRunningTime="2026-02-25 06:48:33.52386062 +0000 UTC m=+206.963117079" Feb 25 06:48:33 crc kubenswrapper[4978]: I0225 06:48:33.855255 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:48:33 crc kubenswrapper[4978]: I0225 06:48:33.899304 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:48:34 crc kubenswrapper[4978]: I0225 06:48:34.155239 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:48:34 crc kubenswrapper[4978]: I0225 06:48:34.155332 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:48:34 crc kubenswrapper[4978]: I0225 06:48:34.476241 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tjqjf" event={"ID":"8178b540-4bca-402c-a2e0-57da7a09fa30","Type":"ContainerStarted","Data":"c992ebe17f4605fd789a6971c6f524883a9154e1379aaf53f85ea185e3e54adc"} Feb 25 06:48:34 crc kubenswrapper[4978]: I0225 06:48:34.495577 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tjqjf" podStartSLOduration=3.620516402 podStartE2EDuration="1m24.495557705s" podCreationTimestamp="2026-02-25 06:47:10 +0000 UTC" firstStartedPulling="2026-02-25 06:47:12.566703458 +0000 UTC m=+126.005959917" lastFinishedPulling="2026-02-25 06:48:33.441744761 +0000 UTC m=+206.881001220" observedRunningTime="2026-02-25 06:48:34.493654706 +0000 UTC m=+207.932911175" watchObservedRunningTime="2026-02-25 06:48:34.495557705 +0000 UTC m=+207.934814174" Feb 25 06:48:35 crc kubenswrapper[4978]: I0225 06:48:35.191606 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zzswn" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerName="registry-server" probeResult="failure" output=< Feb 25 06:48:35 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 06:48:35 crc kubenswrapper[4978]: > Feb 25 06:48:35 crc kubenswrapper[4978]: I0225 06:48:35.579842 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9qqz"] Feb 25 06:48:35 crc kubenswrapper[4978]: I0225 06:48:35.580157 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w9qqz" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" containerName="registry-server" containerID="cri-o://a2e224d8f3961151a0c3cc30c6499394c70fdf59e1a19502e385117e36754236" gracePeriod=2 Feb 25 06:48:37 crc kubenswrapper[4978]: I0225 06:48:37.494161 4978 generic.go:334] "Generic (PLEG): container finished" podID="ca2405db-a877-4fda-a24c-ff39a0069758" containerID="a2e224d8f3961151a0c3cc30c6499394c70fdf59e1a19502e385117e36754236" exitCode=0 Feb 25 06:48:37 crc kubenswrapper[4978]: I0225 06:48:37.494233 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9qqz" event={"ID":"ca2405db-a877-4fda-a24c-ff39a0069758","Type":"ContainerDied","Data":"a2e224d8f3961151a0c3cc30c6499394c70fdf59e1a19502e385117e36754236"} Feb 25 06:48:37 crc kubenswrapper[4978]: I0225 06:48:37.999771 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.087488 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-catalog-content\") pod \"ca2405db-a877-4fda-a24c-ff39a0069758\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.087848 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr85c\" (UniqueName: \"kubernetes.io/projected/ca2405db-a877-4fda-a24c-ff39a0069758-kube-api-access-tr85c\") pod \"ca2405db-a877-4fda-a24c-ff39a0069758\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.087945 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-utilities\") pod \"ca2405db-a877-4fda-a24c-ff39a0069758\" (UID: \"ca2405db-a877-4fda-a24c-ff39a0069758\") " Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.088783 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-utilities" (OuterVolumeSpecName: "utilities") pod "ca2405db-a877-4fda-a24c-ff39a0069758" (UID: "ca2405db-a877-4fda-a24c-ff39a0069758"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.101403 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2405db-a877-4fda-a24c-ff39a0069758-kube-api-access-tr85c" (OuterVolumeSpecName: "kube-api-access-tr85c") pod "ca2405db-a877-4fda-a24c-ff39a0069758" (UID: "ca2405db-a877-4fda-a24c-ff39a0069758"). InnerVolumeSpecName "kube-api-access-tr85c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.113165 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca2405db-a877-4fda-a24c-ff39a0069758" (UID: "ca2405db-a877-4fda-a24c-ff39a0069758"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.189811 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr85c\" (UniqueName: \"kubernetes.io/projected/ca2405db-a877-4fda-a24c-ff39a0069758-kube-api-access-tr85c\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.189850 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.189861 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca2405db-a877-4fda-a24c-ff39a0069758-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.501887 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w9qqz" event={"ID":"ca2405db-a877-4fda-a24c-ff39a0069758","Type":"ContainerDied","Data":"f60379f97bc3bfd556c1a8759fb483182bdd7cf2441356c77eb3a0e9914870ef"} Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.501971 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w9qqz" Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.502212 4978 scope.go:117] "RemoveContainer" containerID="a2e224d8f3961151a0c3cc30c6499394c70fdf59e1a19502e385117e36754236" Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.532658 4978 scope.go:117] "RemoveContainer" containerID="0a7d15a3e54f9c3eac29e05a04be57324d0ce2ec03b144595900d48eabb28695" Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.549053 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9qqz"] Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.559063 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w9qqz"] Feb 25 06:48:38 crc kubenswrapper[4978]: I0225 06:48:38.561283 4978 scope.go:117] "RemoveContainer" containerID="38462a0fa2ea01aea1677a7e7e80488587f32dac33403ba064064b919263ce79" Feb 25 06:48:39 crc kubenswrapper[4978]: I0225 06:48:39.339777 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" path="/var/lib/kubelet/pods/ca2405db-a877-4fda-a24c-ff39a0069758/volumes" Feb 25 06:48:40 crc kubenswrapper[4978]: I0225 06:48:40.565212 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:48:40 crc kubenswrapper[4978]: I0225 06:48:40.565283 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:48:40 crc kubenswrapper[4978]: I0225 06:48:40.612016 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:48:40 crc kubenswrapper[4978]: I0225 06:48:40.864901 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:48:40 crc kubenswrapper[4978]: I0225 06:48:40.864960 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:48:40 crc kubenswrapper[4978]: I0225 06:48:40.916383 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:48:41 crc kubenswrapper[4978]: I0225 06:48:41.026263 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:48:41 crc kubenswrapper[4978]: I0225 06:48:41.026391 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:48:41 crc kubenswrapper[4978]: I0225 06:48:41.083580 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:48:41 crc kubenswrapper[4978]: I0225 06:48:41.147680 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:48:41 crc kubenswrapper[4978]: I0225 06:48:41.147815 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:48:41 crc kubenswrapper[4978]: I0225 06:48:41.208956 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:48:41 crc kubenswrapper[4978]: I0225 06:48:41.584219 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:48:41 crc kubenswrapper[4978]: I0225 06:48:41.588935 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:48:41 crc kubenswrapper[4978]: I0225 06:48:41.589839 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:48:41 crc kubenswrapper[4978]: I0225 06:48:41.598924 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:48:42 crc kubenswrapper[4978]: I0225 06:48:42.583992 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h7dbj"] Feb 25 06:48:43 crc kubenswrapper[4978]: I0225 06:48:43.028341 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:48:43 crc kubenswrapper[4978]: I0225 06:48:43.028454 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:48:43 crc kubenswrapper[4978]: I0225 06:48:43.107702 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:48:43 crc kubenswrapper[4978]: I0225 06:48:43.180081 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tjqjf"] Feb 25 06:48:43 crc kubenswrapper[4978]: I0225 06:48:43.538634 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tjqjf" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerName="registry-server" containerID="cri-o://c992ebe17f4605fd789a6971c6f524883a9154e1379aaf53f85ea185e3e54adc" gracePeriod=2 Feb 25 06:48:43 crc kubenswrapper[4978]: I0225 06:48:43.539099 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h7dbj" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerName="registry-server" containerID="cri-o://7be8c9231564fad9f8b2c31a2e22639620182af3243c73c8fcc13a6d139914ad" gracePeriod=2 Feb 25 06:48:43 crc kubenswrapper[4978]: I0225 06:48:43.612945 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:48:44 crc kubenswrapper[4978]: I0225 06:48:44.223745 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:48:44 crc kubenswrapper[4978]: I0225 06:48:44.291730 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:48:46 crc kubenswrapper[4978]: I0225 06:48:46.540653 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:48:46 crc kubenswrapper[4978]: I0225 06:48:46.540753 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:48:46 crc kubenswrapper[4978]: I0225 06:48:46.566488 4978 generic.go:334] "Generic (PLEG): container finished" podID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerID="7be8c9231564fad9f8b2c31a2e22639620182af3243c73c8fcc13a6d139914ad" exitCode=0 Feb 25 06:48:46 crc kubenswrapper[4978]: I0225 06:48:46.566544 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7dbj" event={"ID":"6a76f909-3198-47cd-adbb-9f94fb99ab48","Type":"ContainerDied","Data":"7be8c9231564fad9f8b2c31a2e22639620182af3243c73c8fcc13a6d139914ad"} Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.506448 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.577604 4978 generic.go:334] "Generic (PLEG): container finished" podID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerID="c992ebe17f4605fd789a6971c6f524883a9154e1379aaf53f85ea185e3e54adc" exitCode=0 Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.577688 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tjqjf" event={"ID":"8178b540-4bca-402c-a2e0-57da7a09fa30","Type":"ContainerDied","Data":"c992ebe17f4605fd789a6971c6f524883a9154e1379aaf53f85ea185e3e54adc"} Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.580938 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zzswn"] Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.581186 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zzswn" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerName="registry-server" containerID="cri-o://5c192f9f8a9f9f6953443a655ff78b96a0c437a957da0fb51aac12bfc3693353" gracePeriod=2 Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.586778 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h7dbj" event={"ID":"6a76f909-3198-47cd-adbb-9f94fb99ab48","Type":"ContainerDied","Data":"c3ff1f1c79ee21b7db55f0ecbf4880886400645959a36e2c971a62bab0b705c1"} Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.587031 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h7dbj" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.599469 4978 scope.go:117] "RemoveContainer" containerID="7be8c9231564fad9f8b2c31a2e22639620182af3243c73c8fcc13a6d139914ad" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.637044 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xnlb\" (UniqueName: \"kubernetes.io/projected/6a76f909-3198-47cd-adbb-9f94fb99ab48-kube-api-access-8xnlb\") pod \"6a76f909-3198-47cd-adbb-9f94fb99ab48\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.637201 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-utilities\") pod \"6a76f909-3198-47cd-adbb-9f94fb99ab48\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.637253 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-catalog-content\") pod \"6a76f909-3198-47cd-adbb-9f94fb99ab48\" (UID: \"6a76f909-3198-47cd-adbb-9f94fb99ab48\") " Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.638347 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-utilities" (OuterVolumeSpecName: "utilities") pod "6a76f909-3198-47cd-adbb-9f94fb99ab48" (UID: "6a76f909-3198-47cd-adbb-9f94fb99ab48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.640678 4978 scope.go:117] "RemoveContainer" containerID="f4003d9ca43f946dbfeece01826fe402cbff432cc55930e7677389473783011a" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.650470 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a76f909-3198-47cd-adbb-9f94fb99ab48-kube-api-access-8xnlb" (OuterVolumeSpecName: "kube-api-access-8xnlb") pod "6a76f909-3198-47cd-adbb-9f94fb99ab48" (UID: "6a76f909-3198-47cd-adbb-9f94fb99ab48"). InnerVolumeSpecName "kube-api-access-8xnlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.683704 4978 scope.go:117] "RemoveContainer" containerID="a956b032b2fa1fc6ca91347f778505b974cdf7926b68a9c2de2d1f287d10a222" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.697688 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a76f909-3198-47cd-adbb-9f94fb99ab48" (UID: "6a76f909-3198-47cd-adbb-9f94fb99ab48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.738945 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xnlb\" (UniqueName: \"kubernetes.io/projected/6a76f909-3198-47cd-adbb-9f94fb99ab48-kube-api-access-8xnlb\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.738995 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.739028 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a76f909-3198-47cd-adbb-9f94fb99ab48-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.800771 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.937861 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h7dbj"] Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.941180 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-catalog-content\") pod \"8178b540-4bca-402c-a2e0-57da7a09fa30\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.941426 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-utilities\") pod \"8178b540-4bca-402c-a2e0-57da7a09fa30\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.941526 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95x8l\" (UniqueName: \"kubernetes.io/projected/8178b540-4bca-402c-a2e0-57da7a09fa30-kube-api-access-95x8l\") pod \"8178b540-4bca-402c-a2e0-57da7a09fa30\" (UID: \"8178b540-4bca-402c-a2e0-57da7a09fa30\") " Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.941575 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h7dbj"] Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.942232 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-utilities" (OuterVolumeSpecName: "utilities") pod "8178b540-4bca-402c-a2e0-57da7a09fa30" (UID: "8178b540-4bca-402c-a2e0-57da7a09fa30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:48:47 crc kubenswrapper[4978]: I0225 06:48:47.946774 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8178b540-4bca-402c-a2e0-57da7a09fa30-kube-api-access-95x8l" (OuterVolumeSpecName: "kube-api-access-95x8l") pod "8178b540-4bca-402c-a2e0-57da7a09fa30" (UID: "8178b540-4bca-402c-a2e0-57da7a09fa30"). InnerVolumeSpecName "kube-api-access-95x8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.008213 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8178b540-4bca-402c-a2e0-57da7a09fa30" (UID: "8178b540-4bca-402c-a2e0-57da7a09fa30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.043271 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.043316 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95x8l\" (UniqueName: \"kubernetes.io/projected/8178b540-4bca-402c-a2e0-57da7a09fa30-kube-api-access-95x8l\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.043336 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8178b540-4bca-402c-a2e0-57da7a09fa30-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.505182 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-756b67b69f-5d94g"] Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.506040 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" podUID="652ab496-ffe4-4e83-8cf4-9da4516b559a" containerName="controller-manager" containerID="cri-o://96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32" gracePeriod=30 Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.602526 4978 generic.go:334] "Generic (PLEG): container finished" podID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerID="5c192f9f8a9f9f6953443a655ff78b96a0c437a957da0fb51aac12bfc3693353" exitCode=0 Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.602618 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzswn" event={"ID":"d6342947-b76f-4c86-8e34-d42a780c6bf5","Type":"ContainerDied","Data":"5c192f9f8a9f9f6953443a655ff78b96a0c437a957da0fb51aac12bfc3693353"} Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.605796 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns"] Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.606032 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" podUID="7feb6319-f64c-47a4-8090-4dcf75f4d3f7" containerName="route-controller-manager" containerID="cri-o://81fcefb665a61468a87a293b4733f7a494926b838956fa87fc95b6892b281f99" gracePeriod=30 Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.617388 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tjqjf" event={"ID":"8178b540-4bca-402c-a2e0-57da7a09fa30","Type":"ContainerDied","Data":"a9f26814afc265765d15d5140f4133d7ac53d797c1578a17c58dc8af237f0080"} Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.617438 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tjqjf" Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.617452 4978 scope.go:117] "RemoveContainer" containerID="c992ebe17f4605fd789a6971c6f524883a9154e1379aaf53f85ea185e3e54adc" Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.658171 4978 scope.go:117] "RemoveContainer" containerID="2ffdbfb1ce9eb22010ad85731983cdcc58587ab8da0f95eb4107368f5e0f3bbc" Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.664822 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tjqjf"] Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.668015 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tjqjf"] Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.681801 4978 scope.go:117] "RemoveContainer" containerID="7953a2670e32437e4b3852e26b73bed4b928a3596a4e252f6af7dc80c09f9925" Feb 25 06:48:48 crc kubenswrapper[4978]: I0225 06:48:48.868356 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.051108 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.061050 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxzv2\" (UniqueName: \"kubernetes.io/projected/d6342947-b76f-4c86-8e34-d42a780c6bf5-kube-api-access-hxzv2\") pod \"d6342947-b76f-4c86-8e34-d42a780c6bf5\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.061131 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-catalog-content\") pod \"d6342947-b76f-4c86-8e34-d42a780c6bf5\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.061239 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-utilities\") pod \"d6342947-b76f-4c86-8e34-d42a780c6bf5\" (UID: \"d6342947-b76f-4c86-8e34-d42a780c6bf5\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.062965 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-utilities" (OuterVolumeSpecName: "utilities") pod "d6342947-b76f-4c86-8e34-d42a780c6bf5" (UID: "d6342947-b76f-4c86-8e34-d42a780c6bf5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.066666 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6342947-b76f-4c86-8e34-d42a780c6bf5-kube-api-access-hxzv2" (OuterVolumeSpecName: "kube-api-access-hxzv2") pod "d6342947-b76f-4c86-8e34-d42a780c6bf5" (UID: "d6342947-b76f-4c86-8e34-d42a780c6bf5"). InnerVolumeSpecName "kube-api-access-hxzv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.162824 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-config\") pod \"652ab496-ffe4-4e83-8cf4-9da4516b559a\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.162927 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/652ab496-ffe4-4e83-8cf4-9da4516b559a-serving-cert\") pod \"652ab496-ffe4-4e83-8cf4-9da4516b559a\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.162969 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-client-ca\") pod \"652ab496-ffe4-4e83-8cf4-9da4516b559a\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.163030 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsv8h\" (UniqueName: \"kubernetes.io/projected/652ab496-ffe4-4e83-8cf4-9da4516b559a-kube-api-access-vsv8h\") pod \"652ab496-ffe4-4e83-8cf4-9da4516b559a\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.163075 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-proxy-ca-bundles\") pod \"652ab496-ffe4-4e83-8cf4-9da4516b559a\" (UID: \"652ab496-ffe4-4e83-8cf4-9da4516b559a\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.163463 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.163496 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxzv2\" (UniqueName: \"kubernetes.io/projected/d6342947-b76f-4c86-8e34-d42a780c6bf5-kube-api-access-hxzv2\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.164047 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-client-ca" (OuterVolumeSpecName: "client-ca") pod "652ab496-ffe4-4e83-8cf4-9da4516b559a" (UID: "652ab496-ffe4-4e83-8cf4-9da4516b559a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.164084 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "652ab496-ffe4-4e83-8cf4-9da4516b559a" (UID: "652ab496-ffe4-4e83-8cf4-9da4516b559a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.164864 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-config" (OuterVolumeSpecName: "config") pod "652ab496-ffe4-4e83-8cf4-9da4516b559a" (UID: "652ab496-ffe4-4e83-8cf4-9da4516b559a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.168695 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/652ab496-ffe4-4e83-8cf4-9da4516b559a-kube-api-access-vsv8h" (OuterVolumeSpecName: "kube-api-access-vsv8h") pod "652ab496-ffe4-4e83-8cf4-9da4516b559a" (UID: "652ab496-ffe4-4e83-8cf4-9da4516b559a"). InnerVolumeSpecName "kube-api-access-vsv8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.170400 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/652ab496-ffe4-4e83-8cf4-9da4516b559a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "652ab496-ffe4-4e83-8cf4-9da4516b559a" (UID: "652ab496-ffe4-4e83-8cf4-9da4516b559a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.211227 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6342947-b76f-4c86-8e34-d42a780c6bf5" (UID: "d6342947-b76f-4c86-8e34-d42a780c6bf5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.265233 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6342947-b76f-4c86-8e34-d42a780c6bf5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.265571 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.265585 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/652ab496-ffe4-4e83-8cf4-9da4516b559a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.265596 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.265610 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsv8h\" (UniqueName: \"kubernetes.io/projected/652ab496-ffe4-4e83-8cf4-9da4516b559a-kube-api-access-vsv8h\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.265625 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/652ab496-ffe4-4e83-8cf4-9da4516b559a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.340093 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" path="/var/lib/kubelet/pods/6a76f909-3198-47cd-adbb-9f94fb99ab48/volumes" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.341413 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" path="/var/lib/kubelet/pods/8178b540-4bca-402c-a2e0-57da7a09fa30/volumes" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.646723 4978 generic.go:334] "Generic (PLEG): container finished" podID="652ab496-ffe4-4e83-8cf4-9da4516b559a" containerID="96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32" exitCode=0 Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.646800 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" event={"ID":"652ab496-ffe4-4e83-8cf4-9da4516b559a","Type":"ContainerDied","Data":"96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32"} Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.646830 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" event={"ID":"652ab496-ffe4-4e83-8cf4-9da4516b559a","Type":"ContainerDied","Data":"8e1ecee9bb876be84911c6434724bc22b2d9d5e2300b730ad789efe2755e2c47"} Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.646861 4978 scope.go:117] "RemoveContainer" containerID="96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.647066 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-756b67b69f-5d94g" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.649784 4978 generic.go:334] "Generic (PLEG): container finished" podID="7feb6319-f64c-47a4-8090-4dcf75f4d3f7" containerID="81fcefb665a61468a87a293b4733f7a494926b838956fa87fc95b6892b281f99" exitCode=0 Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.649837 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" event={"ID":"7feb6319-f64c-47a4-8090-4dcf75f4d3f7","Type":"ContainerDied","Data":"81fcefb665a61468a87a293b4733f7a494926b838956fa87fc95b6892b281f99"} Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.649853 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" event={"ID":"7feb6319-f64c-47a4-8090-4dcf75f4d3f7","Type":"ContainerDied","Data":"9cc08e0a56e5647c5cca5b5fbcd08a6088c5f1b4e85709b7c0f63e27e1e9dbea"} Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.649865 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cc08e0a56e5647c5cca5b5fbcd08a6088c5f1b4e85709b7c0f63e27e1e9dbea" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.653259 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzswn" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.653322 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzswn" event={"ID":"d6342947-b76f-4c86-8e34-d42a780c6bf5","Type":"ContainerDied","Data":"e6526a36d32af4b3b556d40482ae7d3c9abcff2ed30a29e98bb7687ad3926b14"} Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.684482 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.699246 4978 scope.go:117] "RemoveContainer" containerID="96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32" Feb 25 06:48:49 crc kubenswrapper[4978]: E0225 06:48:49.699869 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32\": container with ID starting with 96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32 not found: ID does not exist" containerID="96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.699978 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32"} err="failed to get container status \"96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32\": rpc error: code = NotFound desc = could not find container \"96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32\": container with ID starting with 96c9230c77e5ba3f206d70f75b3de17171f61d059d1d9a9acae3d524600a1c32 not found: ID does not exist" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.700057 4978 scope.go:117] "RemoveContainer" containerID="5c192f9f8a9f9f6953443a655ff78b96a0c437a957da0fb51aac12bfc3693353" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.703629 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-756b67b69f-5d94g"] Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.709042 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-756b67b69f-5d94g"] Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.716604 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zzswn"] Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.719350 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zzswn"] Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.728613 4978 scope.go:117] "RemoveContainer" containerID="60263caf6fc0e64898c66ea8ebe25b84631be5691d76916c94d6d5d87b486834" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.765048 4978 scope.go:117] "RemoveContainer" containerID="32a86d94a98bf96554a0b8e20a6003e2c9c01b97626c822cfe5cee85af8ef71c" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.872900 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n8kp\" (UniqueName: \"kubernetes.io/projected/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-kube-api-access-6n8kp\") pod \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.872999 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-config\") pod \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.873062 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-client-ca\") pod \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.873145 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-serving-cert\") pod \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\" (UID: \"7feb6319-f64c-47a4-8090-4dcf75f4d3f7\") " Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.874117 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-client-ca" (OuterVolumeSpecName: "client-ca") pod "7feb6319-f64c-47a4-8090-4dcf75f4d3f7" (UID: "7feb6319-f64c-47a4-8090-4dcf75f4d3f7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.874176 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-config" (OuterVolumeSpecName: "config") pod "7feb6319-f64c-47a4-8090-4dcf75f4d3f7" (UID: "7feb6319-f64c-47a4-8090-4dcf75f4d3f7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.878926 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7feb6319-f64c-47a4-8090-4dcf75f4d3f7" (UID: "7feb6319-f64c-47a4-8090-4dcf75f4d3f7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.879528 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-kube-api-access-6n8kp" (OuterVolumeSpecName: "kube-api-access-6n8kp") pod "7feb6319-f64c-47a4-8090-4dcf75f4d3f7" (UID: "7feb6319-f64c-47a4-8090-4dcf75f4d3f7"). InnerVolumeSpecName "kube-api-access-6n8kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.974357 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.974488 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n8kp\" (UniqueName: \"kubernetes.io/projected/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-kube-api-access-6n8kp\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.974510 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:49 crc kubenswrapper[4978]: I0225 06:48:49.974529 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7feb6319-f64c-47a4-8090-4dcf75f4d3f7-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.534409 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz"] Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.534810 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerName="extract-utilities" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.534840 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerName="extract-utilities" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.534870 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerName="extract-content" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.534885 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerName="extract-content" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.534911 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62081d94-2fb1-481a-b45f-abd2267878fb" containerName="pruner" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.534929 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="62081d94-2fb1-481a-b45f-abd2267878fb" containerName="pruner" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.534949 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerName="extract-content" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.534965 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerName="extract-content" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.534984 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535000 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535024 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerName="extract-content" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535041 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerName="extract-content" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535062 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerName="extract-utilities" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535079 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerName="extract-utilities" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535100 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerName="extract-utilities" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535116 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerName="extract-utilities" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535147 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535164 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535191 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" containerName="extract-utilities" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535208 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" containerName="extract-utilities" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535234 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535249 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535269 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" containerName="extract-content" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535286 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" containerName="extract-content" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535307 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="652ab496-ffe4-4e83-8cf4-9da4516b559a" containerName="controller-manager" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535323 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="652ab496-ffe4-4e83-8cf4-9da4516b559a" containerName="controller-manager" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535349 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7feb6319-f64c-47a4-8090-4dcf75f4d3f7" containerName="route-controller-manager" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535469 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7feb6319-f64c-47a4-8090-4dcf75f4d3f7" containerName="route-controller-manager" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535491 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dec7193e-a91c-45b7-a8b1-236d626f0540" containerName="oc" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535505 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="dec7193e-a91c-45b7-a8b1-236d626f0540" containerName="oc" Feb 25 06:48:50 crc kubenswrapper[4978]: E0225 06:48:50.535529 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535546 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535738 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8178b540-4bca-402c-a2e0-57da7a09fa30" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535762 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="dec7193e-a91c-45b7-a8b1-236d626f0540" containerName="oc" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535782 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535798 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a76f909-3198-47cd-adbb-9f94fb99ab48" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535817 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="652ab496-ffe4-4e83-8cf4-9da4516b559a" containerName="controller-manager" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535830 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7feb6319-f64c-47a4-8090-4dcf75f4d3f7" containerName="route-controller-manager" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535848 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca2405db-a877-4fda-a24c-ff39a0069758" containerName="registry-server" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.535864 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="62081d94-2fb1-481a-b45f-abd2267878fb" containerName="pruner" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.536836 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.539819 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.540898 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.541513 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.541740 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.542271 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.542912 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.547168 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6"] Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.548517 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.553162 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.560882 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6"] Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.566213 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz"] Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.614939 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-client-ca\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.614983 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-config\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.615022 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv9v8\" (UniqueName: \"kubernetes.io/projected/8dfc6055-a97f-4a1c-80de-6e93fe154912-kube-api-access-pv9v8\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.615066 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-proxy-ca-bundles\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.615199 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dfc6055-a97f-4a1c-80de-6e93fe154912-serving-cert\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.668322 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.698788 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns"] Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.703299 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59b44d8784-lfzns"] Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.716388 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-proxy-ca-bundles\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.716439 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dfc6055-a97f-4a1c-80de-6e93fe154912-serving-cert\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.716469 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-client-ca\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.716488 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pdcb\" (UniqueName: \"kubernetes.io/projected/5093e120-5216-40be-8ebb-8d3c5154d98d-kube-api-access-2pdcb\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.716522 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-config\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.716546 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-client-ca\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.716564 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-config\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.716590 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv9v8\" (UniqueName: \"kubernetes.io/projected/8dfc6055-a97f-4a1c-80de-6e93fe154912-kube-api-access-pv9v8\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.716606 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5093e120-5216-40be-8ebb-8d3c5154d98d-serving-cert\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.719170 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-config\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.719329 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-client-ca\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.719334 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-proxy-ca-bundles\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.725539 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dfc6055-a97f-4a1c-80de-6e93fe154912-serving-cert\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.735990 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv9v8\" (UniqueName: \"kubernetes.io/projected/8dfc6055-a97f-4a1c-80de-6e93fe154912-kube-api-access-pv9v8\") pod \"controller-manager-75df7cc8f4-gtrqz\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.818149 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5093e120-5216-40be-8ebb-8d3c5154d98d-serving-cert\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.818309 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-client-ca\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.818357 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pdcb\" (UniqueName: \"kubernetes.io/projected/5093e120-5216-40be-8ebb-8d3c5154d98d-kube-api-access-2pdcb\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.818448 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-config\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.819198 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-client-ca\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.820681 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-config\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.824998 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5093e120-5216-40be-8ebb-8d3c5154d98d-serving-cert\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.838480 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pdcb\" (UniqueName: \"kubernetes.io/projected/5093e120-5216-40be-8ebb-8d3c5154d98d-kube-api-access-2pdcb\") pod \"route-controller-manager-884dcd8c4-ngxj6\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.929388 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:50 crc kubenswrapper[4978]: I0225 06:48:50.941649 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.351921 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="652ab496-ffe4-4e83-8cf4-9da4516b559a" path="/var/lib/kubelet/pods/652ab496-ffe4-4e83-8cf4-9da4516b559a/volumes" Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.353365 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7feb6319-f64c-47a4-8090-4dcf75f4d3f7" path="/var/lib/kubelet/pods/7feb6319-f64c-47a4-8090-4dcf75f4d3f7/volumes" Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.354477 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6342947-b76f-4c86-8e34-d42a780c6bf5" path="/var/lib/kubelet/pods/d6342947-b76f-4c86-8e34-d42a780c6bf5/volumes" Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.389128 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6"] Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.440484 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz"] Feb 25 06:48:51 crc kubenswrapper[4978]: W0225 06:48:51.462320 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dfc6055_a97f_4a1c_80de_6e93fe154912.slice/crio-bb6fddcb3a36667a9750f5e992238d373ef10b8e556e10bec70421bbb9e48137 WatchSource:0}: Error finding container bb6fddcb3a36667a9750f5e992238d373ef10b8e556e10bec70421bbb9e48137: Status 404 returned error can't find the container with id bb6fddcb3a36667a9750f5e992238d373ef10b8e556e10bec70421bbb9e48137 Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.474998 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bgbqw"] Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.673134 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" event={"ID":"5093e120-5216-40be-8ebb-8d3c5154d98d","Type":"ContainerStarted","Data":"b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f"} Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.673207 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" event={"ID":"5093e120-5216-40be-8ebb-8d3c5154d98d","Type":"ContainerStarted","Data":"326b1e7f919bc9482fd5375ed75bd32fb87f5049fbae8ed1784090ca7e1dd1f9"} Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.673633 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.674678 4978 patch_prober.go:28] interesting pod/route-controller-manager-884dcd8c4-ngxj6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" start-of-body= Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.674760 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.675156 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" event={"ID":"8dfc6055-a97f-4a1c-80de-6e93fe154912","Type":"ContainerStarted","Data":"08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8"} Feb 25 06:48:51 crc kubenswrapper[4978]: I0225 06:48:51.675190 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" event={"ID":"8dfc6055-a97f-4a1c-80de-6e93fe154912","Type":"ContainerStarted","Data":"bb6fddcb3a36667a9750f5e992238d373ef10b8e556e10bec70421bbb9e48137"} Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.324658 4978 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.326156 4978 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.326317 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.326699 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41" gracePeriod=15 Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.326741 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b" gracePeriod=15 Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.326758 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228" gracePeriod=15 Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.326793 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19" gracePeriod=15 Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.326859 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8" gracePeriod=15 Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327403 4978 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.327590 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327614 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.327627 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327637 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.327646 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327653 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.327662 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327670 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.327681 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327689 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.327700 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327709 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.327716 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327724 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.327739 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327746 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327874 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327888 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327901 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327911 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327920 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327930 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.327938 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.328047 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.328058 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.328068 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.328075 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.328185 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.328441 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.352649 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" podStartSLOduration=4.352630336 podStartE2EDuration="4.352630336s" podCreationTimestamp="2026-02-25 06:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:48:51.696217215 +0000 UTC m=+225.135473674" watchObservedRunningTime="2026-02-25 06:48:52.352630336 +0000 UTC m=+225.791886795" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.354208 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.441622 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.441676 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.441713 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.441733 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.441915 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.441967 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.442005 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.442037 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.542960 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543024 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543058 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543062 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543110 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543080 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543205 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543231 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543248 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543193 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543075 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543313 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543343 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543312 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543519 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.543604 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.653916 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:48:52 crc kubenswrapper[4978]: W0225 06:48:52.682617 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-b11e53eb820035557044998c8dda64a34dbbc82a1a21a5383f072eb0169deace WatchSource:0}: Error finding container b11e53eb820035557044998c8dda64a34dbbc82a1a21a5383f072eb0169deace: Status 404 returned error can't find the container with id b11e53eb820035557044998c8dda64a34dbbc82a1a21a5383f072eb0169deace Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.683159 4978 generic.go:334] "Generic (PLEG): container finished" podID="7451e170-1aa9-4967-b33e-b8b42707ecc6" containerID="df379516bc997db8ed5d0d739adf199d56d26d183e198aabf605118471da47a5" exitCode=0 Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.683259 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7451e170-1aa9-4967-b33e-b8b42707ecc6","Type":"ContainerDied","Data":"df379516bc997db8ed5d0d739adf199d56d26d183e198aabf605118471da47a5"} Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.684725 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.685675 4978 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: E0225 06:48:52.685643 4978 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18976a8d1bedb6c3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:48:52.685076163 +0000 UTC m=+226.124332682,LastTimestamp:2026-02-25 06:48:52.685076163 +0000 UTC m=+226.124332682,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.685989 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.686874 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.688545 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.689787 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19" exitCode=0 Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.689821 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b" exitCode=0 Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.689840 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228" exitCode=0 Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.689860 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8" exitCode=2 Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.689936 4978 scope.go:117] "RemoveContainer" containerID="4a117066f9e7ba92e2a4e5a74d22bd9428fa871c643f02b6fa126e35144c44a5" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.690353 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.690605 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.690841 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.691318 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.692087 4978 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.697708 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.697966 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.698594 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.698622 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.699062 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.699345 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.699813 4978 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.700014 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.700318 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.700885 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.701464 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:52 crc kubenswrapper[4978]: I0225 06:48:52.701858 4978 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:53 crc kubenswrapper[4978]: I0225 06:48:53.704648 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 25 06:48:53 crc kubenswrapper[4978]: I0225 06:48:53.709329 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f"} Feb 25 06:48:53 crc kubenswrapper[4978]: I0225 06:48:53.709430 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b11e53eb820035557044998c8dda64a34dbbc82a1a21a5383f072eb0169deace"} Feb 25 06:48:53 crc kubenswrapper[4978]: I0225 06:48:53.711039 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:53 crc kubenswrapper[4978]: I0225 06:48:53.711777 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:53 crc kubenswrapper[4978]: I0225 06:48:53.712129 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:53 crc kubenswrapper[4978]: I0225 06:48:53.712925 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.058552 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.059230 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.059530 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.059828 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.060117 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.087343 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-var-lock\") pod \"7451e170-1aa9-4967-b33e-b8b42707ecc6\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.087420 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-kubelet-dir\") pod \"7451e170-1aa9-4967-b33e-b8b42707ecc6\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.087468 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7451e170-1aa9-4967-b33e-b8b42707ecc6-kube-api-access\") pod \"7451e170-1aa9-4967-b33e-b8b42707ecc6\" (UID: \"7451e170-1aa9-4967-b33e-b8b42707ecc6\") " Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.087957 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-var-lock" (OuterVolumeSpecName: "var-lock") pod "7451e170-1aa9-4967-b33e-b8b42707ecc6" (UID: "7451e170-1aa9-4967-b33e-b8b42707ecc6"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.088015 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7451e170-1aa9-4967-b33e-b8b42707ecc6" (UID: "7451e170-1aa9-4967-b33e-b8b42707ecc6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.092648 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7451e170-1aa9-4967-b33e-b8b42707ecc6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7451e170-1aa9-4967-b33e-b8b42707ecc6" (UID: "7451e170-1aa9-4967-b33e-b8b42707ecc6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.188807 4978 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-var-lock\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.188858 4978 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7451e170-1aa9-4967-b33e-b8b42707ecc6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.188876 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7451e170-1aa9-4967-b33e-b8b42707ecc6-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.716071 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7451e170-1aa9-4967-b33e-b8b42707ecc6","Type":"ContainerDied","Data":"a90098f92d9735dd7148c0954689bec5ab757b83f9230f759001d9180e3d4c4f"} Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.716443 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a90098f92d9735dd7148c0954689bec5ab757b83f9230f759001d9180e3d4c4f" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.716126 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.728288 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.728797 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.729088 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:54 crc kubenswrapper[4978]: I0225 06:48:54.729404 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.401981 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.403069 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.403712 4978 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.404121 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.404404 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.404688 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.404973 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: E0225 06:48:55.483056 4978 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18976a8d1bedb6c3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:48:52.685076163 +0000 UTC m=+226.124332682,LastTimestamp:2026-02-25 06:48:52.685076163 +0000 UTC m=+226.124332682,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.507288 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.507357 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.507461 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.507544 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.507560 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.507666 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.507787 4978 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.507800 4978 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.507809 4978 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.724158 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.726299 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41" exitCode=0 Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.726393 4978 scope.go:117] "RemoveContainer" containerID="6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.726440 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.743350 4978 scope.go:117] "RemoveContainer" containerID="30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.749005 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.749884 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.750527 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.750781 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.750998 4978 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.759621 4978 scope.go:117] "RemoveContainer" containerID="9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.776698 4978 scope.go:117] "RemoveContainer" containerID="52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.793629 4978 scope.go:117] "RemoveContainer" containerID="02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.810503 4978 scope.go:117] "RemoveContainer" containerID="2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.829454 4978 scope.go:117] "RemoveContainer" containerID="6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19" Feb 25 06:48:55 crc kubenswrapper[4978]: E0225 06:48:55.830575 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19\": container with ID starting with 6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19 not found: ID does not exist" containerID="6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.830627 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19"} err="failed to get container status \"6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19\": rpc error: code = NotFound desc = could not find container \"6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19\": container with ID starting with 6216f3e712a624c1bfcafee6795374fd2519647591654cd77e438517d09eef19 not found: ID does not exist" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.830660 4978 scope.go:117] "RemoveContainer" containerID="30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b" Feb 25 06:48:55 crc kubenswrapper[4978]: E0225 06:48:55.831097 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b\": container with ID starting with 30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b not found: ID does not exist" containerID="30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.831142 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b"} err="failed to get container status \"30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b\": rpc error: code = NotFound desc = could not find container \"30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b\": container with ID starting with 30d2cc46a406630540ceb863d01076df066614e9b130319e78fc6ed454b09e3b not found: ID does not exist" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.831174 4978 scope.go:117] "RemoveContainer" containerID="9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228" Feb 25 06:48:55 crc kubenswrapper[4978]: E0225 06:48:55.831466 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228\": container with ID starting with 9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228 not found: ID does not exist" containerID="9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.831487 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228"} err="failed to get container status \"9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228\": rpc error: code = NotFound desc = could not find container \"9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228\": container with ID starting with 9a827e0d389fe056d2777579cd59eb85b9a8927e427b49a09797a3a58c1ff228 not found: ID does not exist" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.831503 4978 scope.go:117] "RemoveContainer" containerID="52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8" Feb 25 06:48:55 crc kubenswrapper[4978]: E0225 06:48:55.831762 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8\": container with ID starting with 52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8 not found: ID does not exist" containerID="52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.831782 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8"} err="failed to get container status \"52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8\": rpc error: code = NotFound desc = could not find container \"52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8\": container with ID starting with 52dad7ff5be1600693dd1b908b4f4a72a7a0c7a510fea6c6166875523facb0b8 not found: ID does not exist" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.831798 4978 scope.go:117] "RemoveContainer" containerID="02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41" Feb 25 06:48:55 crc kubenswrapper[4978]: E0225 06:48:55.832159 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41\": container with ID starting with 02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41 not found: ID does not exist" containerID="02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.832177 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41"} err="failed to get container status \"02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41\": rpc error: code = NotFound desc = could not find container \"02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41\": container with ID starting with 02562b93b8f6e63ae87f5fb375fe40c4c371ba8d62685265d115d6c41b3e0a41 not found: ID does not exist" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.832192 4978 scope.go:117] "RemoveContainer" containerID="2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289" Feb 25 06:48:55 crc kubenswrapper[4978]: E0225 06:48:55.832457 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289\": container with ID starting with 2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289 not found: ID does not exist" containerID="2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289" Feb 25 06:48:55 crc kubenswrapper[4978]: I0225 06:48:55.832476 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289"} err="failed to get container status \"2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289\": rpc error: code = NotFound desc = could not find container \"2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289\": container with ID starting with 2af12559e84be9c9a323c2b6616c1362a1be8571cf6ba171193f17b627142289 not found: ID does not exist" Feb 25 06:48:57 crc kubenswrapper[4978]: I0225 06:48:57.331798 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:57 crc kubenswrapper[4978]: I0225 06:48:57.332108 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:57 crc kubenswrapper[4978]: I0225 06:48:57.332338 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:57 crc kubenswrapper[4978]: I0225 06:48:57.332633 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:57 crc kubenswrapper[4978]: I0225 06:48:57.332973 4978 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:48:57 crc kubenswrapper[4978]: I0225 06:48:57.348147 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 25 06:48:58 crc kubenswrapper[4978]: E0225 06:48:58.353780 4978 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" volumeName="registry-storage" Feb 25 06:49:01 crc kubenswrapper[4978]: E0225 06:49:01.063725 4978 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:01 crc kubenswrapper[4978]: E0225 06:49:01.064318 4978 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:01 crc kubenswrapper[4978]: E0225 06:49:01.064841 4978 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:01 crc kubenswrapper[4978]: E0225 06:49:01.065201 4978 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:01 crc kubenswrapper[4978]: E0225 06:49:01.065545 4978 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:01 crc kubenswrapper[4978]: I0225 06:49:01.065584 4978 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 25 06:49:01 crc kubenswrapper[4978]: E0225 06:49:01.065891 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="200ms" Feb 25 06:49:01 crc kubenswrapper[4978]: E0225 06:49:01.267361 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="400ms" Feb 25 06:49:01 crc kubenswrapper[4978]: E0225 06:49:01.669199 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="800ms" Feb 25 06:49:02 crc kubenswrapper[4978]: E0225 06:49:02.471120 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="1.6s" Feb 25 06:49:04 crc kubenswrapper[4978]: E0225 06:49:04.072228 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="3.2s" Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.682552 4978 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.682617 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.800686 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.801798 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.801879 4978 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f069fc26e6703ac36de7f4179eb3ccbb60416ff19e187774c1c3ef0906c16c47" exitCode=1 Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.801928 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f069fc26e6703ac36de7f4179eb3ccbb60416ff19e187774c1c3ef0906c16c47"} Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.803117 4978 scope.go:117] "RemoveContainer" containerID="f069fc26e6703ac36de7f4179eb3ccbb60416ff19e187774c1c3ef0906c16c47" Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.803518 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.804141 4978 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.804957 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.805658 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:04 crc kubenswrapper[4978]: I0225 06:49:04.806340 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:05 crc kubenswrapper[4978]: E0225 06:49:05.484030 4978 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.146:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18976a8d1bedb6c3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-25 06:48:52.685076163 +0000 UTC m=+226.124332682,LastTimestamp:2026-02-25 06:48:52.685076163 +0000 UTC m=+226.124332682,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 25 06:49:05 crc kubenswrapper[4978]: I0225 06:49:05.811539 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 25 06:49:05 crc kubenswrapper[4978]: I0225 06:49:05.812065 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 25 06:49:05 crc kubenswrapper[4978]: I0225 06:49:05.812113 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"48815216e50d19fd9595f89b0cdd442b050ee9ae5db52962143ecb07aaae7336"} Feb 25 06:49:05 crc kubenswrapper[4978]: I0225 06:49:05.813307 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:05 crc kubenswrapper[4978]: I0225 06:49:05.814017 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:05 crc kubenswrapper[4978]: I0225 06:49:05.814395 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:05 crc kubenswrapper[4978]: I0225 06:49:05.814773 4978 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:05 crc kubenswrapper[4978]: I0225 06:49:05.815204 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:06 crc kubenswrapper[4978]: I0225 06:49:06.928440 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:49:07 crc kubenswrapper[4978]: E0225 06:49:07.274039 4978 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.146:6443: connect: connection refused" interval="6.4s" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.327645 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.331624 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.332177 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.332565 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.332805 4978 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.333006 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.333356 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.333697 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.334003 4978 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.334239 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.334543 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.352134 4978 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33715162-a697-4e39-ba6d-468d3c5d1d74" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.352469 4978 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33715162-a697-4e39-ba6d-468d3c5d1d74" Feb 25 06:49:07 crc kubenswrapper[4978]: E0225 06:49:07.352934 4978 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.353485 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:07 crc kubenswrapper[4978]: W0225 06:49:07.392329 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-9ac1a8d2578abddf9c91d5f60807588268f9a3c078e1861c14250961febed4ea WatchSource:0}: Error finding container 9ac1a8d2578abddf9c91d5f60807588268f9a3c078e1861c14250961febed4ea: Status 404 returned error can't find the container with id 9ac1a8d2578abddf9c91d5f60807588268f9a3c078e1861c14250961febed4ea Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.831517 4978 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d8755a4841176a6d06e99a8056cf237458847093af22fba864accaaca24369dd" exitCode=0 Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.831624 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d8755a4841176a6d06e99a8056cf237458847093af22fba864accaaca24369dd"} Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.831824 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9ac1a8d2578abddf9c91d5f60807588268f9a3c078e1861c14250961febed4ea"} Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.832091 4978 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33715162-a697-4e39-ba6d-468d3c5d1d74" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.832104 4978 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33715162-a697-4e39-ba6d-468d3c5d1d74" Feb 25 06:49:07 crc kubenswrapper[4978]: E0225 06:49:07.832709 4978 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.832696 4978 status_manager.go:851] "Failed to get status for pod" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.833278 4978 status_manager.go:851] "Failed to get status for pod" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-75df7cc8f4-gtrqz\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.833585 4978 status_manager.go:851] "Failed to get status for pod" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-884dcd8c4-ngxj6\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.833900 4978 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:07 crc kubenswrapper[4978]: I0225 06:49:07.834207 4978 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.146:6443: connect: connection refused" Feb 25 06:49:08 crc kubenswrapper[4978]: I0225 06:49:08.854900 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6edbbae10a0e0e4655fbb58c0c84132a16d1c3bb884210517a285bc19f86d9d6"} Feb 25 06:49:08 crc kubenswrapper[4978]: I0225 06:49:08.855613 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0043525598c2404feffefa64aa17dbc884445360d0608e9a99f96752edf4c292"} Feb 25 06:49:08 crc kubenswrapper[4978]: I0225 06:49:08.855624 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e40d34de089d0454a9e4da4a9de0aede7d04b273844176c8ac43b69154048a46"} Feb 25 06:49:09 crc kubenswrapper[4978]: I0225 06:49:09.863190 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b85167c785354d736f581963336c5a5cdb0a3ebe9e26feaf3c5054444ad898ba"} Feb 25 06:49:09 crc kubenswrapper[4978]: I0225 06:49:09.863240 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"43a69fa96d68dc0c8a3bda5ed3fd90ef24704fd5cc55cc57684f3ccf0fc24786"} Feb 25 06:49:09 crc kubenswrapper[4978]: I0225 06:49:09.863515 4978 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33715162-a697-4e39-ba6d-468d3c5d1d74" Feb 25 06:49:09 crc kubenswrapper[4978]: I0225 06:49:09.863533 4978 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33715162-a697-4e39-ba6d-468d3c5d1d74" Feb 25 06:49:09 crc kubenswrapper[4978]: I0225 06:49:09.863820 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:12 crc kubenswrapper[4978]: I0225 06:49:12.353683 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:12 crc kubenswrapper[4978]: I0225 06:49:12.354051 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:12 crc kubenswrapper[4978]: I0225 06:49:12.365185 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:12 crc kubenswrapper[4978]: I0225 06:49:12.567818 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:49:12 crc kubenswrapper[4978]: I0225 06:49:12.576808 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:49:14 crc kubenswrapper[4978]: I0225 06:49:14.878926 4978 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:14 crc kubenswrapper[4978]: I0225 06:49:14.906067 4978 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33715162-a697-4e39-ba6d-468d3c5d1d74" Feb 25 06:49:14 crc kubenswrapper[4978]: I0225 06:49:14.906112 4978 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33715162-a697-4e39-ba6d-468d3c5d1d74" Feb 25 06:49:14 crc kubenswrapper[4978]: I0225 06:49:14.912237 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:15 crc kubenswrapper[4978]: I0225 06:49:15.913554 4978 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33715162-a697-4e39-ba6d-468d3c5d1d74" Feb 25 06:49:15 crc kubenswrapper[4978]: I0225 06:49:15.913701 4978 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33715162-a697-4e39-ba6d-468d3c5d1d74" Feb 25 06:49:16 crc kubenswrapper[4978]: I0225 06:49:16.513872 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" podUID="3494cb07-ccde-443f-8286-f73a5d9ab58f" containerName="oauth-openshift" containerID="cri-o://f26f144dd90d99146f3bc099e517057b00cf1224b1ac80420db090ef40ebea9d" gracePeriod=15 Feb 25 06:49:16 crc kubenswrapper[4978]: I0225 06:49:16.540212 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:49:16 crc kubenswrapper[4978]: I0225 06:49:16.540336 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:49:16 crc kubenswrapper[4978]: I0225 06:49:16.920855 4978 generic.go:334] "Generic (PLEG): container finished" podID="3494cb07-ccde-443f-8286-f73a5d9ab58f" containerID="f26f144dd90d99146f3bc099e517057b00cf1224b1ac80420db090ef40ebea9d" exitCode=0 Feb 25 06:49:16 crc kubenswrapper[4978]: I0225 06:49:16.920991 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" event={"ID":"3494cb07-ccde-443f-8286-f73a5d9ab58f","Type":"ContainerDied","Data":"f26f144dd90d99146f3bc099e517057b00cf1224b1ac80420db090ef40ebea9d"} Feb 25 06:49:16 crc kubenswrapper[4978]: I0225 06:49:16.932700 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.060518 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.141109 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-ocp-branding-template\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.141218 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lrvb\" (UniqueName: \"kubernetes.io/projected/3494cb07-ccde-443f-8286-f73a5d9ab58f-kube-api-access-8lrvb\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.141274 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-policies\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.141359 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-provider-selection\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.141431 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-idp-0-file-data\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.142505 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-router-certs\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.142582 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-serving-cert\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.142626 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-cliconfig\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.142696 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-trusted-ca-bundle\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.142741 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-dir\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.142933 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.142971 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.142839 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-service-ca\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.143179 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-error\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.143288 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-login\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.143357 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-session\") pod \"3494cb07-ccde-443f-8286-f73a5d9ab58f\" (UID: \"3494cb07-ccde-443f-8286-f73a5d9ab58f\") " Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.143719 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.143859 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.144291 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.144326 4978 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.144351 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.144397 4978 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.144571 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.163099 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.164053 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.164848 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.165068 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3494cb07-ccde-443f-8286-f73a5d9ab58f-kube-api-access-8lrvb" (OuterVolumeSpecName: "kube-api-access-8lrvb") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "kube-api-access-8lrvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.165329 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.165875 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.166150 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.166740 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.167220 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "3494cb07-ccde-443f-8286-f73a5d9ab58f" (UID: "3494cb07-ccde-443f-8286-f73a5d9ab58f"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.246431 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.246500 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.246522 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.246541 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.246560 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.246579 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.246598 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.246619 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.246639 4978 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3494cb07-ccde-443f-8286-f73a5d9ab58f-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.246661 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lrvb\" (UniqueName: \"kubernetes.io/projected/3494cb07-ccde-443f-8286-f73a5d9ab58f-kube-api-access-8lrvb\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.361091 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="dbe7c103-600b-4f2c-a675-36a7175b793c" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.930172 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" event={"ID":"3494cb07-ccde-443f-8286-f73a5d9ab58f","Type":"ContainerDied","Data":"869a93fafe8f80d5d0edca5760b26b81882eee0bd1ecee1359dd401042e75103"} Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.930632 4978 scope.go:117] "RemoveContainer" containerID="f26f144dd90d99146f3bc099e517057b00cf1224b1ac80420db090ef40ebea9d" Feb 25 06:49:17 crc kubenswrapper[4978]: I0225 06:49:17.930243 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bgbqw" Feb 25 06:49:24 crc kubenswrapper[4978]: I0225 06:49:24.612337 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 25 06:49:25 crc kubenswrapper[4978]: I0225 06:49:25.789785 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 25 06:49:25 crc kubenswrapper[4978]: I0225 06:49:25.831681 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 25 06:49:25 crc kubenswrapper[4978]: I0225 06:49:25.907121 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 25 06:49:25 crc kubenswrapper[4978]: I0225 06:49:25.931488 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 25 06:49:25 crc kubenswrapper[4978]: I0225 06:49:25.937176 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 25 06:49:26 crc kubenswrapper[4978]: I0225 06:49:26.186706 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 25 06:49:26 crc kubenswrapper[4978]: I0225 06:49:26.288586 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 25 06:49:26 crc kubenswrapper[4978]: I0225 06:49:26.436819 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 25 06:49:26 crc kubenswrapper[4978]: I0225 06:49:26.563350 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 25 06:49:26 crc kubenswrapper[4978]: I0225 06:49:26.706234 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 25 06:49:26 crc kubenswrapper[4978]: I0225 06:49:26.742067 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 25 06:49:26 crc kubenswrapper[4978]: I0225 06:49:26.765644 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 25 06:49:26 crc kubenswrapper[4978]: I0225 06:49:26.887530 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 25 06:49:27 crc kubenswrapper[4978]: I0225 06:49:27.016174 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 25 06:49:27 crc kubenswrapper[4978]: I0225 06:49:27.032067 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 25 06:49:27 crc kubenswrapper[4978]: I0225 06:49:27.161403 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 25 06:49:27 crc kubenswrapper[4978]: I0225 06:49:27.267380 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 25 06:49:27 crc kubenswrapper[4978]: I0225 06:49:27.392530 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 25 06:49:27 crc kubenswrapper[4978]: I0225 06:49:27.766644 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 25 06:49:27 crc kubenswrapper[4978]: I0225 06:49:27.855865 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 25 06:49:27 crc kubenswrapper[4978]: I0225 06:49:27.990461 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 25 06:49:28 crc kubenswrapper[4978]: I0225 06:49:28.215946 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 25 06:49:28 crc kubenswrapper[4978]: I0225 06:49:28.284046 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 25 06:49:28 crc kubenswrapper[4978]: I0225 06:49:28.352214 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 25 06:49:28 crc kubenswrapper[4978]: I0225 06:49:28.412196 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 25 06:49:28 crc kubenswrapper[4978]: I0225 06:49:28.568133 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 25 06:49:28 crc kubenswrapper[4978]: I0225 06:49:28.638062 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 25 06:49:28 crc kubenswrapper[4978]: I0225 06:49:28.995171 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.085952 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.182228 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.305339 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.392210 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.559253 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.669470 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.670331 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.722632 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.812757 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.826164 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.883867 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 25 06:49:29 crc kubenswrapper[4978]: I0225 06:49:29.971788 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.049736 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.081752 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.120336 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.121740 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.160752 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.179616 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.208985 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.250630 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.263306 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.273995 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.323530 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.331416 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.413219 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.455557 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.508193 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.643165 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.647859 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.770766 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.793251 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.799516 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.846207 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.860261 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.899147 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.918168 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.965288 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 25 06:49:30 crc kubenswrapper[4978]: I0225 06:49:30.966255 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.112806 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.155113 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.174179 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.200109 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.207167 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.280524 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.374444 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.395620 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.409791 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.465501 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.587170 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.685993 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.695948 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.748595 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.829631 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.836076 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.882444 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.939323 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 25 06:49:31 crc kubenswrapper[4978]: I0225 06:49:31.964170 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.000893 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.050131 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.054123 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.105669 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.220531 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.260742 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.360789 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.361625 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.381044 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.485883 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.490169 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.574198 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.671472 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.738318 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.740947 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.769242 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.853078 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.896971 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.916967 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.926000 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.927551 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 25 06:49:32 crc kubenswrapper[4978]: I0225 06:49:32.957596 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.098240 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.143855 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.155398 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.333006 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.365031 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.365340 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.376360 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.513802 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.567818 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.587615 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.632804 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.802844 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.820476 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.957521 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.979098 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.991296 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 25 06:49:33 crc kubenswrapper[4978]: I0225 06:49:33.993713 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.129412 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.177584 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.230133 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.236299 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.270696 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.339942 4978 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.347481 4978 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.347772 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" podStartSLOduration=46.347710806 podStartE2EDuration="46.347710806s" podCreationTimestamp="2026-02-25 06:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:49:14.937596339 +0000 UTC m=+248.376852858" watchObservedRunningTime="2026-02-25 06:49:34.347710806 +0000 UTC m=+267.786967335" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.351018 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.351925 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=42.351907448 podStartE2EDuration="42.351907448s" podCreationTimestamp="2026-02-25 06:48:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:49:14.979458987 +0000 UTC m=+248.418715456" watchObservedRunningTime="2026-02-25 06:49:34.351907448 +0000 UTC m=+267.791163947" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.356329 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-bgbqw"] Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.356441 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.366710 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.387518 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.387490648 podStartE2EDuration="20.387490648s" podCreationTimestamp="2026-02-25 06:49:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:49:34.382950055 +0000 UTC m=+267.822206524" watchObservedRunningTime="2026-02-25 06:49:34.387490648 +0000 UTC m=+267.826747137" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.420485 4978 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.446597 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.453874 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.461797 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.549361 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.603045 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.637204 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.701024 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.704190 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.816152 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.862771 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 25 06:49:34 crc kubenswrapper[4978]: I0225 06:49:34.991038 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.015969 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.058510 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.095260 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.253922 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.322053 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.340817 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3494cb07-ccde-443f-8286-f73a5d9ab58f" path="/var/lib/kubelet/pods/3494cb07-ccde-443f-8286-f73a5d9ab58f/volumes" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.412666 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.417318 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.441457 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.504549 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.529560 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.853630 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 25 06:49:35 crc kubenswrapper[4978]: I0225 06:49:35.879535 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.055784 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.060253 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.101027 4978 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.150326 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.260919 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.282415 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.286629 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.336036 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.379865 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.421168 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.434830 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.486889 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.515901 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.547240 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.556613 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.594686 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.637171 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.703440 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.737471 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.743007 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.789879 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.802262 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.823479 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 25 06:49:36 crc kubenswrapper[4978]: I0225 06:49:36.935275 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.015008 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.067675 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.074316 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.098709 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.150844 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.183051 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.271836 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.336073 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.386563 4978 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.386823 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f" gracePeriod=5 Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.405124 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.413149 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.415070 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.497275 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.508304 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.537196 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.591091 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.646599 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.807314 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 25 06:49:37 crc kubenswrapper[4978]: I0225 06:49:37.885944 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.047116 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.107067 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.213063 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.254309 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.279853 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.352110 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.391702 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.392658 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.430917 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.433479 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.479460 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.517258 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 25 06:49:38 crc kubenswrapper[4978]: I0225 06:49:38.903211 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.012601 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.218326 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.249348 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.311925 4978 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.337735 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.343714 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.352524 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.590427 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.675183 4978 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.739647 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.742162 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.893624 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 25 06:49:39 crc kubenswrapper[4978]: I0225 06:49:39.990872 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 25 06:49:40 crc kubenswrapper[4978]: I0225 06:49:40.277063 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 25 06:49:40 crc kubenswrapper[4978]: I0225 06:49:40.337817 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 06:49:40 crc kubenswrapper[4978]: I0225 06:49:40.426149 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 25 06:49:40 crc kubenswrapper[4978]: I0225 06:49:40.959007 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 25 06:49:41 crc kubenswrapper[4978]: I0225 06:49:41.055318 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 25 06:49:41 crc kubenswrapper[4978]: I0225 06:49:41.255402 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 25 06:49:41 crc kubenswrapper[4978]: I0225 06:49:41.324439 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 25 06:49:41 crc kubenswrapper[4978]: I0225 06:49:41.584423 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 25 06:49:41 crc kubenswrapper[4978]: I0225 06:49:41.670781 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.778132 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7d8984bd85-7rp4m"] Feb 25 06:49:42 crc kubenswrapper[4978]: E0225 06:49:42.780119 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3494cb07-ccde-443f-8286-f73a5d9ab58f" containerName="oauth-openshift" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.780151 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3494cb07-ccde-443f-8286-f73a5d9ab58f" containerName="oauth-openshift" Feb 25 06:49:42 crc kubenswrapper[4978]: E0225 06:49:42.780179 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.780195 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 25 06:49:42 crc kubenswrapper[4978]: E0225 06:49:42.780220 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" containerName="installer" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.780238 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" containerName="installer" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.780501 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7451e170-1aa9-4967-b33e-b8b42707ecc6" containerName="installer" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.780533 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.780554 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3494cb07-ccde-443f-8286-f73a5d9ab58f" containerName="oauth-openshift" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.781328 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.787782 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.787960 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.788283 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.788744 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.788811 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.789129 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.792141 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.792243 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.792629 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.792661 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.793512 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.795064 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.806593 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.812940 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d8984bd85-7rp4m"] Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.816149 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.839547 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841355 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841484 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-template-error\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841533 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-session\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841571 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcxlz\" (UniqueName: \"kubernetes.io/projected/1efecc15-6a97-4bca-ad6d-634f7579fb63-kube-api-access-hcxlz\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841657 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841705 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841747 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-template-login\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841793 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-audit-policies\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841845 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841894 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841936 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.841977 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1efecc15-6a97-4bca-ad6d-634f7579fb63-audit-dir\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.842023 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.842076 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943395 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943461 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943490 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943521 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1efecc15-6a97-4bca-ad6d-634f7579fb63-audit-dir\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943554 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943590 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943634 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943664 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-template-error\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943688 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-session\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943713 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcxlz\" (UniqueName: \"kubernetes.io/projected/1efecc15-6a97-4bca-ad6d-634f7579fb63-kube-api-access-hcxlz\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943752 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943780 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943802 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-template-login\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.943826 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-audit-policies\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.945021 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-audit-policies\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.945083 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1efecc15-6a97-4bca-ad6d-634f7579fb63-audit-dir\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.946266 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.946514 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.947477 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.953182 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.953737 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-template-error\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.954263 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.954363 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.956273 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.958146 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.961713 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-user-template-login\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.961943 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/1efecc15-6a97-4bca-ad6d-634f7579fb63-v4-0-config-system-session\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:42 crc kubenswrapper[4978]: I0225 06:49:42.981860 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcxlz\" (UniqueName: \"kubernetes.io/projected/1efecc15-6a97-4bca-ad6d-634f7579fb63-kube-api-access-hcxlz\") pod \"oauth-openshift-7d8984bd85-7rp4m\" (UID: \"1efecc15-6a97-4bca-ad6d-634f7579fb63\") " pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.018619 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.018771 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.116046 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.125454 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.125615 4978 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f" exitCode=137 Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.125705 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.125719 4978 scope.go:117] "RemoveContainer" containerID="9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.147333 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.147463 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.147684 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.147767 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.147895 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.147562 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.149094 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.148887 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.149573 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.150362 4978 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.150440 4978 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.150458 4978 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.150540 4978 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.154781 4978 scope.go:117] "RemoveContainer" containerID="9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f" Feb 25 06:49:43 crc kubenswrapper[4978]: E0225 06:49:43.155312 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f\": container with ID starting with 9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f not found: ID does not exist" containerID="9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.155406 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f"} err="failed to get container status \"9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f\": rpc error: code = NotFound desc = could not find container \"9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f\": container with ID starting with 9eaf06702d580d931efa9b91103781cd6885fbaadfceb0a4f61188986f62854f not found: ID does not exist" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.159717 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.255122 4978 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.346133 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.346457 4978 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.370449 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.370514 4978 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="0b2c032a-a894-4c02-a45b-91143a86a956" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.375529 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.375571 4978 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="0b2c032a-a894-4c02-a45b-91143a86a956" Feb 25 06:49:43 crc kubenswrapper[4978]: I0225 06:49:43.638946 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d8984bd85-7rp4m"] Feb 25 06:49:43 crc kubenswrapper[4978]: W0225 06:49:43.649705 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1efecc15_6a97_4bca_ad6d_634f7579fb63.slice/crio-d753c58123cd499d25345e76c76c41dc24e9acd698734b967cf733ea27fa5182 WatchSource:0}: Error finding container d753c58123cd499d25345e76c76c41dc24e9acd698734b967cf733ea27fa5182: Status 404 returned error can't find the container with id d753c58123cd499d25345e76c76c41dc24e9acd698734b967cf733ea27fa5182 Feb 25 06:49:44 crc kubenswrapper[4978]: I0225 06:49:44.134892 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" event={"ID":"1efecc15-6a97-4bca-ad6d-634f7579fb63","Type":"ContainerStarted","Data":"02c3a61bace2cebf7b51ba545f7bf1470baddf379506aa5e4fb79992d3d8ce5d"} Feb 25 06:49:44 crc kubenswrapper[4978]: I0225 06:49:44.138038 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:44 crc kubenswrapper[4978]: I0225 06:49:44.138263 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" event={"ID":"1efecc15-6a97-4bca-ad6d-634f7579fb63","Type":"ContainerStarted","Data":"d753c58123cd499d25345e76c76c41dc24e9acd698734b967cf733ea27fa5182"} Feb 25 06:49:44 crc kubenswrapper[4978]: I0225 06:49:44.164698 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" podStartSLOduration=53.164670679 podStartE2EDuration="53.164670679s" podCreationTimestamp="2026-02-25 06:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:49:44.158591166 +0000 UTC m=+277.597847685" watchObservedRunningTime="2026-02-25 06:49:44.164670679 +0000 UTC m=+277.603927178" Feb 25 06:49:44 crc kubenswrapper[4978]: I0225 06:49:44.526466 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7d8984bd85-7rp4m" Feb 25 06:49:46 crc kubenswrapper[4978]: I0225 06:49:46.540735 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:49:46 crc kubenswrapper[4978]: I0225 06:49:46.541015 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:49:46 crc kubenswrapper[4978]: I0225 06:49:46.541090 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:49:46 crc kubenswrapper[4978]: I0225 06:49:46.542106 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1681565645f2fc8aba05522208118ff62097a02395870d10eb9361b066bdb939"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 06:49:46 crc kubenswrapper[4978]: I0225 06:49:46.542218 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://1681565645f2fc8aba05522208118ff62097a02395870d10eb9361b066bdb939" gracePeriod=600 Feb 25 06:49:47 crc kubenswrapper[4978]: I0225 06:49:47.158546 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="1681565645f2fc8aba05522208118ff62097a02395870d10eb9361b066bdb939" exitCode=0 Feb 25 06:49:47 crc kubenswrapper[4978]: I0225 06:49:47.158683 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"1681565645f2fc8aba05522208118ff62097a02395870d10eb9361b066bdb939"} Feb 25 06:49:47 crc kubenswrapper[4978]: I0225 06:49:47.159361 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"cda6e0eb7c768cf92c9013ff2dbcca2a8bfe25b85cfa848c319625de918b2b8e"} Feb 25 06:49:48 crc kubenswrapper[4978]: I0225 06:49:48.483795 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz"] Feb 25 06:49:48 crc kubenswrapper[4978]: I0225 06:49:48.485479 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" containerName="controller-manager" containerID="cri-o://08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8" gracePeriod=30 Feb 25 06:49:48 crc kubenswrapper[4978]: I0225 06:49:48.573739 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6"] Feb 25 06:49:48 crc kubenswrapper[4978]: I0225 06:49:48.573999 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" containerName="route-controller-manager" containerID="cri-o://b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f" gracePeriod=30 Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.046796 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.051741 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.164586 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pdcb\" (UniqueName: \"kubernetes.io/projected/5093e120-5216-40be-8ebb-8d3c5154d98d-kube-api-access-2pdcb\") pod \"5093e120-5216-40be-8ebb-8d3c5154d98d\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165034 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-client-ca\") pod \"8dfc6055-a97f-4a1c-80de-6e93fe154912\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165065 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pv9v8\" (UniqueName: \"kubernetes.io/projected/8dfc6055-a97f-4a1c-80de-6e93fe154912-kube-api-access-pv9v8\") pod \"8dfc6055-a97f-4a1c-80de-6e93fe154912\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165091 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-client-ca\") pod \"5093e120-5216-40be-8ebb-8d3c5154d98d\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165111 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-config\") pod \"8dfc6055-a97f-4a1c-80de-6e93fe154912\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165152 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-proxy-ca-bundles\") pod \"8dfc6055-a97f-4a1c-80de-6e93fe154912\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165171 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-config\") pod \"5093e120-5216-40be-8ebb-8d3c5154d98d\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165199 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5093e120-5216-40be-8ebb-8d3c5154d98d-serving-cert\") pod \"5093e120-5216-40be-8ebb-8d3c5154d98d\" (UID: \"5093e120-5216-40be-8ebb-8d3c5154d98d\") " Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165268 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dfc6055-a97f-4a1c-80de-6e93fe154912-serving-cert\") pod \"8dfc6055-a97f-4a1c-80de-6e93fe154912\" (UID: \"8dfc6055-a97f-4a1c-80de-6e93fe154912\") " Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165776 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-client-ca" (OuterVolumeSpecName: "client-ca") pod "5093e120-5216-40be-8ebb-8d3c5154d98d" (UID: "5093e120-5216-40be-8ebb-8d3c5154d98d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165927 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-client-ca" (OuterVolumeSpecName: "client-ca") pod "8dfc6055-a97f-4a1c-80de-6e93fe154912" (UID: "8dfc6055-a97f-4a1c-80de-6e93fe154912"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.165915 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8dfc6055-a97f-4a1c-80de-6e93fe154912" (UID: "8dfc6055-a97f-4a1c-80de-6e93fe154912"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.166099 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-config" (OuterVolumeSpecName: "config") pod "5093e120-5216-40be-8ebb-8d3c5154d98d" (UID: "5093e120-5216-40be-8ebb-8d3c5154d98d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.167209 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-config" (OuterVolumeSpecName: "config") pod "8dfc6055-a97f-4a1c-80de-6e93fe154912" (UID: "8dfc6055-a97f-4a1c-80de-6e93fe154912"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.171358 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dfc6055-a97f-4a1c-80de-6e93fe154912-kube-api-access-pv9v8" (OuterVolumeSpecName: "kube-api-access-pv9v8") pod "8dfc6055-a97f-4a1c-80de-6e93fe154912" (UID: "8dfc6055-a97f-4a1c-80de-6e93fe154912"). InnerVolumeSpecName "kube-api-access-pv9v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.172174 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5093e120-5216-40be-8ebb-8d3c5154d98d-kube-api-access-2pdcb" (OuterVolumeSpecName: "kube-api-access-2pdcb") pod "5093e120-5216-40be-8ebb-8d3c5154d98d" (UID: "5093e120-5216-40be-8ebb-8d3c5154d98d"). InnerVolumeSpecName "kube-api-access-2pdcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.172248 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5093e120-5216-40be-8ebb-8d3c5154d98d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5093e120-5216-40be-8ebb-8d3c5154d98d" (UID: "5093e120-5216-40be-8ebb-8d3c5154d98d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.172567 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dfc6055-a97f-4a1c-80de-6e93fe154912-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8dfc6055-a97f-4a1c-80de-6e93fe154912" (UID: "8dfc6055-a97f-4a1c-80de-6e93fe154912"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.173658 4978 generic.go:334] "Generic (PLEG): container finished" podID="8dfc6055-a97f-4a1c-80de-6e93fe154912" containerID="08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8" exitCode=0 Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.173786 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" event={"ID":"8dfc6055-a97f-4a1c-80de-6e93fe154912","Type":"ContainerDied","Data":"08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8"} Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.173883 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" event={"ID":"8dfc6055-a97f-4a1c-80de-6e93fe154912","Type":"ContainerDied","Data":"bb6fddcb3a36667a9750f5e992238d373ef10b8e556e10bec70421bbb9e48137"} Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.173957 4978 scope.go:117] "RemoveContainer" containerID="08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.174119 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.176161 4978 generic.go:334] "Generic (PLEG): container finished" podID="5093e120-5216-40be-8ebb-8d3c5154d98d" containerID="b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f" exitCode=0 Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.176185 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" event={"ID":"5093e120-5216-40be-8ebb-8d3c5154d98d","Type":"ContainerDied","Data":"b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f"} Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.176200 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" event={"ID":"5093e120-5216-40be-8ebb-8d3c5154d98d","Type":"ContainerDied","Data":"326b1e7f919bc9482fd5375ed75bd32fb87f5049fbae8ed1784090ca7e1dd1f9"} Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.176241 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.198658 4978 scope.go:117] "RemoveContainer" containerID="08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8" Feb 25 06:49:49 crc kubenswrapper[4978]: E0225 06:49:49.199990 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8\": container with ID starting with 08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8 not found: ID does not exist" containerID="08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.200055 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8"} err="failed to get container status \"08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8\": rpc error: code = NotFound desc = could not find container \"08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8\": container with ID starting with 08946740622d17ce6ab65c7d8b2e3c3e35c8d549a10ff364d0da3b0942cf5ef8 not found: ID does not exist" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.200091 4978 scope.go:117] "RemoveContainer" containerID="b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.218578 4978 scope.go:117] "RemoveContainer" containerID="b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.218872 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz"] Feb 25 06:49:49 crc kubenswrapper[4978]: E0225 06:49:49.219084 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f\": container with ID starting with b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f not found: ID does not exist" containerID="b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.219131 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f"} err="failed to get container status \"b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f\": rpc error: code = NotFound desc = could not find container \"b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f\": container with ID starting with b1404a9b5a3db6a4a8f2d1b7a8352434eef39113cc979fd0d9cda6ec7ca71f9f not found: ID does not exist" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.226305 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-75df7cc8f4-gtrqz"] Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.230475 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6"] Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.233500 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-884dcd8c4-ngxj6"] Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.266818 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.266866 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pv9v8\" (UniqueName: \"kubernetes.io/projected/8dfc6055-a97f-4a1c-80de-6e93fe154912-kube-api-access-pv9v8\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.266878 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.266888 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.266897 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8dfc6055-a97f-4a1c-80de-6e93fe154912-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.266905 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5093e120-5216-40be-8ebb-8d3c5154d98d-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.266913 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5093e120-5216-40be-8ebb-8d3c5154d98d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.266922 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dfc6055-a97f-4a1c-80de-6e93fe154912-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.266930 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pdcb\" (UniqueName: \"kubernetes.io/projected/5093e120-5216-40be-8ebb-8d3c5154d98d-kube-api-access-2pdcb\") on node \"crc\" DevicePath \"\"" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.341168 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" path="/var/lib/kubelet/pods/5093e120-5216-40be-8ebb-8d3c5154d98d/volumes" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.342100 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" path="/var/lib/kubelet/pods/8dfc6055-a97f-4a1c-80de-6e93fe154912/volumes" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.776515 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd"] Feb 25 06:49:49 crc kubenswrapper[4978]: E0225 06:49:49.777010 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" containerName="controller-manager" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.777045 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" containerName="controller-manager" Feb 25 06:49:49 crc kubenswrapper[4978]: E0225 06:49:49.777078 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" containerName="route-controller-manager" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.777096 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" containerName="route-controller-manager" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.777321 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dfc6055-a97f-4a1c-80de-6e93fe154912" containerName="controller-manager" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.777353 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5093e120-5216-40be-8ebb-8d3c5154d98d" containerName="route-controller-manager" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.778233 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.780895 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x"] Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.781713 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.782520 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.787665 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.788305 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.794826 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.795066 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.795207 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.795229 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.795497 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.795667 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.796144 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.796700 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.802518 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x"] Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.802774 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.806289 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.806341 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd"] Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.875351 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-config\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.875457 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-config\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.875484 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-proxy-ca-bundles\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.875523 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2d9l\" (UniqueName: \"kubernetes.io/projected/4d7e65ac-c661-40ee-ab23-4e68476287df-kube-api-access-l2d9l\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.875553 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36d457f1-a6d0-4413-98df-145c5cfac166-serving-cert\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.875577 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-client-ca\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.875605 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-client-ca\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.875624 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d7e65ac-c661-40ee-ab23-4e68476287df-serving-cert\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.875672 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-978gk\" (UniqueName: \"kubernetes.io/projected/36d457f1-a6d0-4413-98df-145c5cfac166-kube-api-access-978gk\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.976740 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-config\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.977109 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-config\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.977253 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-proxy-ca-bundles\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.977398 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36d457f1-a6d0-4413-98df-145c5cfac166-serving-cert\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.978592 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-proxy-ca-bundles\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.978728 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-config\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.978762 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-client-ca\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.978818 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-client-ca\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.978840 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2d9l\" (UniqueName: \"kubernetes.io/projected/4d7e65ac-c661-40ee-ab23-4e68476287df-kube-api-access-l2d9l\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.979307 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-client-ca\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.979330 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d7e65ac-c661-40ee-ab23-4e68476287df-serving-cert\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.979452 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-978gk\" (UniqueName: \"kubernetes.io/projected/36d457f1-a6d0-4413-98df-145c5cfac166-kube-api-access-978gk\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.980749 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-client-ca\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.980791 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-config\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.983981 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36d457f1-a6d0-4413-98df-145c5cfac166-serving-cert\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.995425 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d7e65ac-c661-40ee-ab23-4e68476287df-serving-cert\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.997031 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-978gk\" (UniqueName: \"kubernetes.io/projected/36d457f1-a6d0-4413-98df-145c5cfac166-kube-api-access-978gk\") pod \"controller-manager-6d6fb8f486-6g4cd\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:49 crc kubenswrapper[4978]: I0225 06:49:49.998603 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2d9l\" (UniqueName: \"kubernetes.io/projected/4d7e65ac-c661-40ee-ab23-4e68476287df-kube-api-access-l2d9l\") pod \"route-controller-manager-64b664c989-wxx7x\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:50 crc kubenswrapper[4978]: I0225 06:49:50.155816 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:50 crc kubenswrapper[4978]: I0225 06:49:50.169980 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:50 crc kubenswrapper[4978]: I0225 06:49:50.364294 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x"] Feb 25 06:49:50 crc kubenswrapper[4978]: W0225 06:49:50.370307 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d7e65ac_c661_40ee_ab23_4e68476287df.slice/crio-e21581dca1007be81a7e0fd56b26cf491bef2dd86d094cc1de0052c44ab0c463 WatchSource:0}: Error finding container e21581dca1007be81a7e0fd56b26cf491bef2dd86d094cc1de0052c44ab0c463: Status 404 returned error can't find the container with id e21581dca1007be81a7e0fd56b26cf491bef2dd86d094cc1de0052c44ab0c463 Feb 25 06:49:50 crc kubenswrapper[4978]: I0225 06:49:50.440785 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd"] Feb 25 06:49:51 crc kubenswrapper[4978]: I0225 06:49:51.190931 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" event={"ID":"4d7e65ac-c661-40ee-ab23-4e68476287df","Type":"ContainerStarted","Data":"904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e"} Feb 25 06:49:51 crc kubenswrapper[4978]: I0225 06:49:51.191499 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:51 crc kubenswrapper[4978]: I0225 06:49:51.191518 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" event={"ID":"4d7e65ac-c661-40ee-ab23-4e68476287df","Type":"ContainerStarted","Data":"e21581dca1007be81a7e0fd56b26cf491bef2dd86d094cc1de0052c44ab0c463"} Feb 25 06:49:51 crc kubenswrapper[4978]: I0225 06:49:51.192274 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" event={"ID":"36d457f1-a6d0-4413-98df-145c5cfac166","Type":"ContainerStarted","Data":"249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281"} Feb 25 06:49:51 crc kubenswrapper[4978]: I0225 06:49:51.192317 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" event={"ID":"36d457f1-a6d0-4413-98df-145c5cfac166","Type":"ContainerStarted","Data":"3b194fc601d872a5619e5293172029805122966242ea69da9c978b01cd9bfd3b"} Feb 25 06:49:51 crc kubenswrapper[4978]: I0225 06:49:51.197680 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:49:51 crc kubenswrapper[4978]: I0225 06:49:51.214137 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" podStartSLOduration=3.214109765 podStartE2EDuration="3.214109765s" podCreationTimestamp="2026-02-25 06:49:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:49:51.208561508 +0000 UTC m=+284.647817987" watchObservedRunningTime="2026-02-25 06:49:51.214109765 +0000 UTC m=+284.653366224" Feb 25 06:49:51 crc kubenswrapper[4978]: I0225 06:49:51.232676 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" podStartSLOduration=3.232653993 podStartE2EDuration="3.232653993s" podCreationTimestamp="2026-02-25 06:49:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:49:51.229311118 +0000 UTC m=+284.668567587" watchObservedRunningTime="2026-02-25 06:49:51.232653993 +0000 UTC m=+284.671910452" Feb 25 06:49:52 crc kubenswrapper[4978]: I0225 06:49:52.198989 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:52 crc kubenswrapper[4978]: I0225 06:49:52.208104 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:49:59 crc kubenswrapper[4978]: I0225 06:49:59.169554 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.176868 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533370-4jjws"] Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.177993 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533370-4jjws" Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.182583 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.182727 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.185141 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.186415 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533370-4jjws"] Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.338048 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzhp5\" (UniqueName: \"kubernetes.io/projected/ed03df44-2b1e-4633-9743-79c9a9961ff0-kube-api-access-lzhp5\") pod \"auto-csr-approver-29533370-4jjws\" (UID: \"ed03df44-2b1e-4633-9743-79c9a9961ff0\") " pod="openshift-infra/auto-csr-approver-29533370-4jjws" Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.439401 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzhp5\" (UniqueName: \"kubernetes.io/projected/ed03df44-2b1e-4633-9743-79c9a9961ff0-kube-api-access-lzhp5\") pod \"auto-csr-approver-29533370-4jjws\" (UID: \"ed03df44-2b1e-4633-9743-79c9a9961ff0\") " pod="openshift-infra/auto-csr-approver-29533370-4jjws" Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.462163 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzhp5\" (UniqueName: \"kubernetes.io/projected/ed03df44-2b1e-4633-9743-79c9a9961ff0-kube-api-access-lzhp5\") pod \"auto-csr-approver-29533370-4jjws\" (UID: \"ed03df44-2b1e-4633-9743-79c9a9961ff0\") " pod="openshift-infra/auto-csr-approver-29533370-4jjws" Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.512729 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533370-4jjws" Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.566676 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 25 06:50:00 crc kubenswrapper[4978]: I0225 06:50:00.951651 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533370-4jjws"] Feb 25 06:50:00 crc kubenswrapper[4978]: W0225 06:50:00.963933 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded03df44_2b1e_4633_9743_79c9a9961ff0.slice/crio-46be8f050cec4474177cc57f507db90fea030e9395e9be6c017c5db2d5a6e610 WatchSource:0}: Error finding container 46be8f050cec4474177cc57f507db90fea030e9395e9be6c017c5db2d5a6e610: Status 404 returned error can't find the container with id 46be8f050cec4474177cc57f507db90fea030e9395e9be6c017c5db2d5a6e610 Feb 25 06:50:01 crc kubenswrapper[4978]: I0225 06:50:01.262710 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533370-4jjws" event={"ID":"ed03df44-2b1e-4633-9743-79c9a9961ff0","Type":"ContainerStarted","Data":"46be8f050cec4474177cc57f507db90fea030e9395e9be6c017c5db2d5a6e610"} Feb 25 06:50:03 crc kubenswrapper[4978]: I0225 06:50:03.281306 4978 generic.go:334] "Generic (PLEG): container finished" podID="ed03df44-2b1e-4633-9743-79c9a9961ff0" containerID="5679b9a49e1fbf84e5cbefc047e8cd2f9f7857495901ecc6e7deb6d4640ceea9" exitCode=0 Feb 25 06:50:03 crc kubenswrapper[4978]: I0225 06:50:03.281420 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533370-4jjws" event={"ID":"ed03df44-2b1e-4633-9743-79c9a9961ff0","Type":"ContainerDied","Data":"5679b9a49e1fbf84e5cbefc047e8cd2f9f7857495901ecc6e7deb6d4640ceea9"} Feb 25 06:50:04 crc kubenswrapper[4978]: I0225 06:50:04.733383 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533370-4jjws" Feb 25 06:50:04 crc kubenswrapper[4978]: I0225 06:50:04.910891 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzhp5\" (UniqueName: \"kubernetes.io/projected/ed03df44-2b1e-4633-9743-79c9a9961ff0-kube-api-access-lzhp5\") pod \"ed03df44-2b1e-4633-9743-79c9a9961ff0\" (UID: \"ed03df44-2b1e-4633-9743-79c9a9961ff0\") " Feb 25 06:50:04 crc kubenswrapper[4978]: I0225 06:50:04.920430 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed03df44-2b1e-4633-9743-79c9a9961ff0-kube-api-access-lzhp5" (OuterVolumeSpecName: "kube-api-access-lzhp5") pod "ed03df44-2b1e-4633-9743-79c9a9961ff0" (UID: "ed03df44-2b1e-4633-9743-79c9a9961ff0"). InnerVolumeSpecName "kube-api-access-lzhp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:50:05 crc kubenswrapper[4978]: I0225 06:50:05.012861 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzhp5\" (UniqueName: \"kubernetes.io/projected/ed03df44-2b1e-4633-9743-79c9a9961ff0-kube-api-access-lzhp5\") on node \"crc\" DevicePath \"\"" Feb 25 06:50:05 crc kubenswrapper[4978]: I0225 06:50:05.300734 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533370-4jjws" event={"ID":"ed03df44-2b1e-4633-9743-79c9a9961ff0","Type":"ContainerDied","Data":"46be8f050cec4474177cc57f507db90fea030e9395e9be6c017c5db2d5a6e610"} Feb 25 06:50:05 crc kubenswrapper[4978]: I0225 06:50:05.300798 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46be8f050cec4474177cc57f507db90fea030e9395e9be6c017c5db2d5a6e610" Feb 25 06:50:05 crc kubenswrapper[4978]: I0225 06:50:05.300820 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533370-4jjws" Feb 25 06:50:08 crc kubenswrapper[4978]: I0225 06:50:08.487486 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd"] Feb 25 06:50:08 crc kubenswrapper[4978]: I0225 06:50:08.489725 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" podUID="36d457f1-a6d0-4413-98df-145c5cfac166" containerName="controller-manager" containerID="cri-o://249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281" gracePeriod=30 Feb 25 06:50:08 crc kubenswrapper[4978]: I0225 06:50:08.510233 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x"] Feb 25 06:50:08 crc kubenswrapper[4978]: I0225 06:50:08.510561 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" podUID="4d7e65ac-c661-40ee-ab23-4e68476287df" containerName="route-controller-manager" containerID="cri-o://904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e" gracePeriod=30 Feb 25 06:50:08 crc kubenswrapper[4978]: I0225 06:50:08.614828 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.054496 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.175240 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-client-ca\") pod \"4d7e65ac-c661-40ee-ab23-4e68476287df\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.175317 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d7e65ac-c661-40ee-ab23-4e68476287df-serving-cert\") pod \"4d7e65ac-c661-40ee-ab23-4e68476287df\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.175353 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-config\") pod \"4d7e65ac-c661-40ee-ab23-4e68476287df\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.175502 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2d9l\" (UniqueName: \"kubernetes.io/projected/4d7e65ac-c661-40ee-ab23-4e68476287df-kube-api-access-l2d9l\") pod \"4d7e65ac-c661-40ee-ab23-4e68476287df\" (UID: \"4d7e65ac-c661-40ee-ab23-4e68476287df\") " Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.176727 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-config" (OuterVolumeSpecName: "config") pod "4d7e65ac-c661-40ee-ab23-4e68476287df" (UID: "4d7e65ac-c661-40ee-ab23-4e68476287df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.176802 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-client-ca" (OuterVolumeSpecName: "client-ca") pod "4d7e65ac-c661-40ee-ab23-4e68476287df" (UID: "4d7e65ac-c661-40ee-ab23-4e68476287df"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.183858 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7e65ac-c661-40ee-ab23-4e68476287df-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4d7e65ac-c661-40ee-ab23-4e68476287df" (UID: "4d7e65ac-c661-40ee-ab23-4e68476287df"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.184041 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d7e65ac-c661-40ee-ab23-4e68476287df-kube-api-access-l2d9l" (OuterVolumeSpecName: "kube-api-access-l2d9l") pod "4d7e65ac-c661-40ee-ab23-4e68476287df" (UID: "4d7e65ac-c661-40ee-ab23-4e68476287df"). InnerVolumeSpecName "kube-api-access-l2d9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.228998 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.277220 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d7e65ac-c661-40ee-ab23-4e68476287df-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.277282 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.277322 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2d9l\" (UniqueName: \"kubernetes.io/projected/4d7e65ac-c661-40ee-ab23-4e68476287df-kube-api-access-l2d9l\") on node \"crc\" DevicePath \"\"" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.277346 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d7e65ac-c661-40ee-ab23-4e68476287df-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.334193 4978 generic.go:334] "Generic (PLEG): container finished" podID="36d457f1-a6d0-4413-98df-145c5cfac166" containerID="249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281" exitCode=0 Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.334414 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.337048 4978 generic.go:334] "Generic (PLEG): container finished" podID="4d7e65ac-c661-40ee-ab23-4e68476287df" containerID="904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e" exitCode=0 Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.337152 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.346229 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" event={"ID":"36d457f1-a6d0-4413-98df-145c5cfac166","Type":"ContainerDied","Data":"249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281"} Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.346289 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd" event={"ID":"36d457f1-a6d0-4413-98df-145c5cfac166","Type":"ContainerDied","Data":"3b194fc601d872a5619e5293172029805122966242ea69da9c978b01cd9bfd3b"} Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.346312 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" event={"ID":"4d7e65ac-c661-40ee-ab23-4e68476287df","Type":"ContainerDied","Data":"904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e"} Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.346334 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x" event={"ID":"4d7e65ac-c661-40ee-ab23-4e68476287df","Type":"ContainerDied","Data":"e21581dca1007be81a7e0fd56b26cf491bef2dd86d094cc1de0052c44ab0c463"} Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.346364 4978 scope.go:117] "RemoveContainer" containerID="249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.369671 4978 scope.go:117] "RemoveContainer" containerID="249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281" Feb 25 06:50:09 crc kubenswrapper[4978]: E0225 06:50:09.370697 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281\": container with ID starting with 249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281 not found: ID does not exist" containerID="249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.370774 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281"} err="failed to get container status \"249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281\": rpc error: code = NotFound desc = could not find container \"249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281\": container with ID starting with 249bcad4765c6b1e589da1fb1e431dc657a54112832417cf0457bf8fb42a5281 not found: ID does not exist" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.370818 4978 scope.go:117] "RemoveContainer" containerID="904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.378183 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-978gk\" (UniqueName: \"kubernetes.io/projected/36d457f1-a6d0-4413-98df-145c5cfac166-kube-api-access-978gk\") pod \"36d457f1-a6d0-4413-98df-145c5cfac166\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.378277 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-client-ca\") pod \"36d457f1-a6d0-4413-98df-145c5cfac166\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.378319 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-config\") pod \"36d457f1-a6d0-4413-98df-145c5cfac166\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.378458 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36d457f1-a6d0-4413-98df-145c5cfac166-serving-cert\") pod \"36d457f1-a6d0-4413-98df-145c5cfac166\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.378493 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-proxy-ca-bundles\") pod \"36d457f1-a6d0-4413-98df-145c5cfac166\" (UID: \"36d457f1-a6d0-4413-98df-145c5cfac166\") " Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.380772 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-client-ca" (OuterVolumeSpecName: "client-ca") pod "36d457f1-a6d0-4413-98df-145c5cfac166" (UID: "36d457f1-a6d0-4413-98df-145c5cfac166"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.380881 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-config" (OuterVolumeSpecName: "config") pod "36d457f1-a6d0-4413-98df-145c5cfac166" (UID: "36d457f1-a6d0-4413-98df-145c5cfac166"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.383540 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "36d457f1-a6d0-4413-98df-145c5cfac166" (UID: "36d457f1-a6d0-4413-98df-145c5cfac166"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.385396 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36d457f1-a6d0-4413-98df-145c5cfac166-kube-api-access-978gk" (OuterVolumeSpecName: "kube-api-access-978gk") pod "36d457f1-a6d0-4413-98df-145c5cfac166" (UID: "36d457f1-a6d0-4413-98df-145c5cfac166"). InnerVolumeSpecName "kube-api-access-978gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.385805 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36d457f1-a6d0-4413-98df-145c5cfac166-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "36d457f1-a6d0-4413-98df-145c5cfac166" (UID: "36d457f1-a6d0-4413-98df-145c5cfac166"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.390249 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x"] Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.392704 4978 scope.go:117] "RemoveContainer" containerID="904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e" Feb 25 06:50:09 crc kubenswrapper[4978]: E0225 06:50:09.393291 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e\": container with ID starting with 904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e not found: ID does not exist" containerID="904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.393357 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e"} err="failed to get container status \"904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e\": rpc error: code = NotFound desc = could not find container \"904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e\": container with ID starting with 904e3b00b3f71ef4569f69fed0288f96f622515f1dcab62953961d87da9ad58e not found: ID does not exist" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.395357 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64b664c989-wxx7x"] Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.481222 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-978gk\" (UniqueName: \"kubernetes.io/projected/36d457f1-a6d0-4413-98df-145c5cfac166-kube-api-access-978gk\") on node \"crc\" DevicePath \"\"" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.481309 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.481336 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.481359 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36d457f1-a6d0-4413-98df-145c5cfac166-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.481417 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36d457f1-a6d0-4413-98df-145c5cfac166-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.671843 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd"] Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.691251 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6d6fb8f486-6g4cd"] Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.793355 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7cc687dd9c-vnswp"] Feb 25 06:50:09 crc kubenswrapper[4978]: E0225 06:50:09.793760 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7e65ac-c661-40ee-ab23-4e68476287df" containerName="route-controller-manager" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.793784 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7e65ac-c661-40ee-ab23-4e68476287df" containerName="route-controller-manager" Feb 25 06:50:09 crc kubenswrapper[4978]: E0225 06:50:09.793805 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed03df44-2b1e-4633-9743-79c9a9961ff0" containerName="oc" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.793815 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed03df44-2b1e-4633-9743-79c9a9961ff0" containerName="oc" Feb 25 06:50:09 crc kubenswrapper[4978]: E0225 06:50:09.793834 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d457f1-a6d0-4413-98df-145c5cfac166" containerName="controller-manager" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.793845 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d457f1-a6d0-4413-98df-145c5cfac166" containerName="controller-manager" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.794010 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7e65ac-c661-40ee-ab23-4e68476287df" containerName="route-controller-manager" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.794032 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed03df44-2b1e-4633-9743-79c9a9961ff0" containerName="oc" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.794049 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d457f1-a6d0-4413-98df-145c5cfac166" containerName="controller-manager" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.794875 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.797799 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.797838 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.797905 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.798473 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.798765 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.803219 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb"] Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.804202 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.804700 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.808574 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.809192 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.809661 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.813639 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.813862 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.813972 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.818438 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.820028 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cc687dd9c-vnswp"] Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.825595 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb"] Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.988420 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-client-ca\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.988502 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n85nx\" (UniqueName: \"kubernetes.io/projected/4e8765fb-fa99-4b4f-82e5-cee23cd14381-kube-api-access-n85nx\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.988543 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-client-ca\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.988596 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-config\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.988619 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-proxy-ca-bundles\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.988755 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eca509d4-31d3-43da-8ad3-2932d080ef40-serving-cert\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.988996 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7m4n\" (UniqueName: \"kubernetes.io/projected/eca509d4-31d3-43da-8ad3-2932d080ef40-kube-api-access-p7m4n\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.989049 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-config\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:09 crc kubenswrapper[4978]: I0225 06:50:09.989203 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e8765fb-fa99-4b4f-82e5-cee23cd14381-serving-cert\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.091033 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-client-ca\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.091092 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n85nx\" (UniqueName: \"kubernetes.io/projected/4e8765fb-fa99-4b4f-82e5-cee23cd14381-kube-api-access-n85nx\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.091120 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-client-ca\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.091160 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-config\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.091177 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-proxy-ca-bundles\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.091193 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eca509d4-31d3-43da-8ad3-2932d080ef40-serving-cert\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.091215 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-config\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.091232 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7m4n\" (UniqueName: \"kubernetes.io/projected/eca509d4-31d3-43da-8ad3-2932d080ef40-kube-api-access-p7m4n\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.091261 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e8765fb-fa99-4b4f-82e5-cee23cd14381-serving-cert\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.092106 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-client-ca\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.092484 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-client-ca\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.092782 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-config\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.093510 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-config\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.093580 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-proxy-ca-bundles\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.097762 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e8765fb-fa99-4b4f-82e5-cee23cd14381-serving-cert\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.099059 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eca509d4-31d3-43da-8ad3-2932d080ef40-serving-cert\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.115065 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n85nx\" (UniqueName: \"kubernetes.io/projected/4e8765fb-fa99-4b4f-82e5-cee23cd14381-kube-api-access-n85nx\") pod \"controller-manager-7cc687dd9c-vnswp\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.117520 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.118005 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7m4n\" (UniqueName: \"kubernetes.io/projected/eca509d4-31d3-43da-8ad3-2932d080ef40-kube-api-access-p7m4n\") pod \"route-controller-manager-c669855c-w7rbb\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.134606 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.353114 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cc687dd9c-vnswp"] Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.361215 4978 generic.go:334] "Generic (PLEG): container finished" podID="9b843a14-9586-4f29-bfba-8413f4e52104" containerID="c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c" exitCode=0 Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.361321 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" event={"ID":"9b843a14-9586-4f29-bfba-8413f4e52104","Type":"ContainerDied","Data":"c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c"} Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.362125 4978 scope.go:117] "RemoveContainer" containerID="c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c" Feb 25 06:50:10 crc kubenswrapper[4978]: I0225 06:50:10.420710 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb"] Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.339208 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36d457f1-a6d0-4413-98df-145c5cfac166" path="/var/lib/kubelet/pods/36d457f1-a6d0-4413-98df-145c5cfac166/volumes" Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.340983 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d7e65ac-c661-40ee-ab23-4e68476287df" path="/var/lib/kubelet/pods/4d7e65ac-c661-40ee-ab23-4e68476287df/volumes" Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.384818 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" event={"ID":"9b843a14-9586-4f29-bfba-8413f4e52104","Type":"ContainerStarted","Data":"55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa"} Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.385326 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.386263 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" event={"ID":"eca509d4-31d3-43da-8ad3-2932d080ef40","Type":"ContainerStarted","Data":"0716b405cdda02473d3998dff6b22beb74fb86a438118ee30c6b9bfe167bec2c"} Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.386324 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" event={"ID":"eca509d4-31d3-43da-8ad3-2932d080ef40","Type":"ContainerStarted","Data":"7a81798e8d4a3141ba239756bb1a6212f9304afa489264346ee440cce76438ce"} Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.386879 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.387359 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.389502 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" event={"ID":"4e8765fb-fa99-4b4f-82e5-cee23cd14381","Type":"ContainerStarted","Data":"ab72fd886e2913173239aa9b6ea47f1c1ac143da7f0ff1cb54705651fe504177"} Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.390419 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.390514 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" event={"ID":"4e8765fb-fa99-4b4f-82e5-cee23cd14381","Type":"ContainerStarted","Data":"3a01dead6b3e615bb05a9455ff0089624a3ad3c4b85720ac06d5bde17c8ff270"} Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.393620 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.394120 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.466208 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" podStartSLOduration=3.466187776 podStartE2EDuration="3.466187776s" podCreationTimestamp="2026-02-25 06:50:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:50:11.463421548 +0000 UTC m=+304.902678027" watchObservedRunningTime="2026-02-25 06:50:11.466187776 +0000 UTC m=+304.905444245" Feb 25 06:50:11 crc kubenswrapper[4978]: I0225 06:50:11.491326 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" podStartSLOduration=3.491303824 podStartE2EDuration="3.491303824s" podCreationTimestamp="2026-02-25 06:50:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:50:11.487048259 +0000 UTC m=+304.926304758" watchObservedRunningTime="2026-02-25 06:50:11.491303824 +0000 UTC m=+304.930560293" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.524131 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-plhd2"] Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.526139 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.540759 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-plhd2"] Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.676918 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4d0208e6-fd09-4e33-8317-a0d486220f44-installation-pull-secrets\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.676982 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4d0208e6-fd09-4e33-8317-a0d486220f44-bound-sa-token\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.677018 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4d0208e6-fd09-4e33-8317-a0d486220f44-registry-tls\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.677267 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4d0208e6-fd09-4e33-8317-a0d486220f44-registry-certificates\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.677488 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.677558 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4d0208e6-fd09-4e33-8317-a0d486220f44-ca-trust-extracted\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.677602 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d0208e6-fd09-4e33-8317-a0d486220f44-trusted-ca\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.677633 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69xrn\" (UniqueName: \"kubernetes.io/projected/4d0208e6-fd09-4e33-8317-a0d486220f44-kube-api-access-69xrn\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.703864 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.779046 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4d0208e6-fd09-4e33-8317-a0d486220f44-registry-certificates\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.779148 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4d0208e6-fd09-4e33-8317-a0d486220f44-ca-trust-extracted\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.779186 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d0208e6-fd09-4e33-8317-a0d486220f44-trusted-ca\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.779219 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69xrn\" (UniqueName: \"kubernetes.io/projected/4d0208e6-fd09-4e33-8317-a0d486220f44-kube-api-access-69xrn\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.779251 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4d0208e6-fd09-4e33-8317-a0d486220f44-installation-pull-secrets\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.779283 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4d0208e6-fd09-4e33-8317-a0d486220f44-bound-sa-token\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.779312 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4d0208e6-fd09-4e33-8317-a0d486220f44-registry-tls\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.779803 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4d0208e6-fd09-4e33-8317-a0d486220f44-ca-trust-extracted\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.780888 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4d0208e6-fd09-4e33-8317-a0d486220f44-registry-certificates\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.782061 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4d0208e6-fd09-4e33-8317-a0d486220f44-trusted-ca\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.788923 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4d0208e6-fd09-4e33-8317-a0d486220f44-registry-tls\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.789341 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4d0208e6-fd09-4e33-8317-a0d486220f44-installation-pull-secrets\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.798576 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69xrn\" (UniqueName: \"kubernetes.io/projected/4d0208e6-fd09-4e33-8317-a0d486220f44-kube-api-access-69xrn\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.812428 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4d0208e6-fd09-4e33-8317-a0d486220f44-bound-sa-token\") pod \"image-registry-66df7c8f76-plhd2\" (UID: \"4d0208e6-fd09-4e33-8317-a0d486220f44\") " pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:02 crc kubenswrapper[4978]: I0225 06:51:02.856243 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:03 crc kubenswrapper[4978]: I0225 06:51:03.287946 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-plhd2"] Feb 25 06:51:03 crc kubenswrapper[4978]: I0225 06:51:03.736674 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" event={"ID":"4d0208e6-fd09-4e33-8317-a0d486220f44","Type":"ContainerStarted","Data":"e3d9783ccaa1a63cfd6a9cfd89a35565ad48f0a376cb5fec11e8cf24ee13c786"} Feb 25 06:51:03 crc kubenswrapper[4978]: I0225 06:51:03.736727 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" event={"ID":"4d0208e6-fd09-4e33-8317-a0d486220f44","Type":"ContainerStarted","Data":"095019aa4849588997217e3ba1dd6a47ba984c7274efa35fcc4c88eabdc4af6f"} Feb 25 06:51:03 crc kubenswrapper[4978]: I0225 06:51:03.736892 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:03 crc kubenswrapper[4978]: I0225 06:51:03.767687 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" podStartSLOduration=1.767665249 podStartE2EDuration="1.767665249s" podCreationTimestamp="2026-02-25 06:51:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:51:03.763529766 +0000 UTC m=+357.202786285" watchObservedRunningTime="2026-02-25 06:51:03.767665249 +0000 UTC m=+357.206921738" Feb 25 06:51:08 crc kubenswrapper[4978]: I0225 06:51:08.476890 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cc687dd9c-vnswp"] Feb 25 06:51:08 crc kubenswrapper[4978]: I0225 06:51:08.477812 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" podUID="4e8765fb-fa99-4b4f-82e5-cee23cd14381" containerName="controller-manager" containerID="cri-o://ab72fd886e2913173239aa9b6ea47f1c1ac143da7f0ff1cb54705651fe504177" gracePeriod=30 Feb 25 06:51:08 crc kubenswrapper[4978]: I0225 06:51:08.503349 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb"] Feb 25 06:51:08 crc kubenswrapper[4978]: I0225 06:51:08.503594 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" podUID="eca509d4-31d3-43da-8ad3-2932d080ef40" containerName="route-controller-manager" containerID="cri-o://0716b405cdda02473d3998dff6b22beb74fb86a438118ee30c6b9bfe167bec2c" gracePeriod=30 Feb 25 06:51:08 crc kubenswrapper[4978]: I0225 06:51:08.769700 4978 generic.go:334] "Generic (PLEG): container finished" podID="eca509d4-31d3-43da-8ad3-2932d080ef40" containerID="0716b405cdda02473d3998dff6b22beb74fb86a438118ee30c6b9bfe167bec2c" exitCode=0 Feb 25 06:51:08 crc kubenswrapper[4978]: I0225 06:51:08.769815 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" event={"ID":"eca509d4-31d3-43da-8ad3-2932d080ef40","Type":"ContainerDied","Data":"0716b405cdda02473d3998dff6b22beb74fb86a438118ee30c6b9bfe167bec2c"} Feb 25 06:51:08 crc kubenswrapper[4978]: I0225 06:51:08.771649 4978 generic.go:334] "Generic (PLEG): container finished" podID="4e8765fb-fa99-4b4f-82e5-cee23cd14381" containerID="ab72fd886e2913173239aa9b6ea47f1c1ac143da7f0ff1cb54705651fe504177" exitCode=0 Feb 25 06:51:08 crc kubenswrapper[4978]: I0225 06:51:08.771695 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" event={"ID":"4e8765fb-fa99-4b4f-82e5-cee23cd14381","Type":"ContainerDied","Data":"ab72fd886e2913173239aa9b6ea47f1c1ac143da7f0ff1cb54705651fe504177"} Feb 25 06:51:08 crc kubenswrapper[4978]: I0225 06:51:08.933036 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:51:08 crc kubenswrapper[4978]: I0225 06:51:08.937495 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.068706 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7m4n\" (UniqueName: \"kubernetes.io/projected/eca509d4-31d3-43da-8ad3-2932d080ef40-kube-api-access-p7m4n\") pod \"eca509d4-31d3-43da-8ad3-2932d080ef40\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.068812 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e8765fb-fa99-4b4f-82e5-cee23cd14381-serving-cert\") pod \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.068841 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-client-ca\") pod \"eca509d4-31d3-43da-8ad3-2932d080ef40\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.068888 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eca509d4-31d3-43da-8ad3-2932d080ef40-serving-cert\") pod \"eca509d4-31d3-43da-8ad3-2932d080ef40\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.068907 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-config\") pod \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.068921 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-config\") pod \"eca509d4-31d3-43da-8ad3-2932d080ef40\" (UID: \"eca509d4-31d3-43da-8ad3-2932d080ef40\") " Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.068935 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-client-ca\") pod \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.068980 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n85nx\" (UniqueName: \"kubernetes.io/projected/4e8765fb-fa99-4b4f-82e5-cee23cd14381-kube-api-access-n85nx\") pod \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.069005 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-proxy-ca-bundles\") pod \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\" (UID: \"4e8765fb-fa99-4b4f-82e5-cee23cd14381\") " Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.069774 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4e8765fb-fa99-4b4f-82e5-cee23cd14381" (UID: "4e8765fb-fa99-4b4f-82e5-cee23cd14381"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.069855 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-config" (OuterVolumeSpecName: "config") pod "eca509d4-31d3-43da-8ad3-2932d080ef40" (UID: "eca509d4-31d3-43da-8ad3-2932d080ef40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.069922 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-client-ca" (OuterVolumeSpecName: "client-ca") pod "eca509d4-31d3-43da-8ad3-2932d080ef40" (UID: "eca509d4-31d3-43da-8ad3-2932d080ef40"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.070133 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-client-ca" (OuterVolumeSpecName: "client-ca") pod "4e8765fb-fa99-4b4f-82e5-cee23cd14381" (UID: "4e8765fb-fa99-4b4f-82e5-cee23cd14381"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.070351 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-config" (OuterVolumeSpecName: "config") pod "4e8765fb-fa99-4b4f-82e5-cee23cd14381" (UID: "4e8765fb-fa99-4b4f-82e5-cee23cd14381"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.074503 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e8765fb-fa99-4b4f-82e5-cee23cd14381-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4e8765fb-fa99-4b4f-82e5-cee23cd14381" (UID: "4e8765fb-fa99-4b4f-82e5-cee23cd14381"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.074660 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e8765fb-fa99-4b4f-82e5-cee23cd14381-kube-api-access-n85nx" (OuterVolumeSpecName: "kube-api-access-n85nx") pod "4e8765fb-fa99-4b4f-82e5-cee23cd14381" (UID: "4e8765fb-fa99-4b4f-82e5-cee23cd14381"). InnerVolumeSpecName "kube-api-access-n85nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.075031 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eca509d4-31d3-43da-8ad3-2932d080ef40-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "eca509d4-31d3-43da-8ad3-2932d080ef40" (UID: "eca509d4-31d3-43da-8ad3-2932d080ef40"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.075762 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eca509d4-31d3-43da-8ad3-2932d080ef40-kube-api-access-p7m4n" (OuterVolumeSpecName: "kube-api-access-p7m4n") pod "eca509d4-31d3-43da-8ad3-2932d080ef40" (UID: "eca509d4-31d3-43da-8ad3-2932d080ef40"). InnerVolumeSpecName "kube-api-access-p7m4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.170199 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eca509d4-31d3-43da-8ad3-2932d080ef40-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.170240 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.170250 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.170260 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.170272 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n85nx\" (UniqueName: \"kubernetes.io/projected/4e8765fb-fa99-4b4f-82e5-cee23cd14381-kube-api-access-n85nx\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.170286 4978 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4e8765fb-fa99-4b4f-82e5-cee23cd14381-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.170298 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7m4n\" (UniqueName: \"kubernetes.io/projected/eca509d4-31d3-43da-8ad3-2932d080ef40-kube-api-access-p7m4n\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.170309 4978 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4e8765fb-fa99-4b4f-82e5-cee23cd14381-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.170320 4978 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/eca509d4-31d3-43da-8ad3-2932d080ef40-client-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.779856 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.780609 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb" event={"ID":"eca509d4-31d3-43da-8ad3-2932d080ef40","Type":"ContainerDied","Data":"7a81798e8d4a3141ba239756bb1a6212f9304afa489264346ee440cce76438ce"} Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.780695 4978 scope.go:117] "RemoveContainer" containerID="0716b405cdda02473d3998dff6b22beb74fb86a438118ee30c6b9bfe167bec2c" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.782272 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" event={"ID":"4e8765fb-fa99-4b4f-82e5-cee23cd14381","Type":"ContainerDied","Data":"3a01dead6b3e615bb05a9455ff0089624a3ad3c4b85720ac06d5bde17c8ff270"} Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.782650 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cc687dd9c-vnswp" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.797566 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cc687dd9c-vnswp"] Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.804881 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7cc687dd9c-vnswp"] Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.812110 4978 scope.go:117] "RemoveContainer" containerID="ab72fd886e2913173239aa9b6ea47f1c1ac143da7f0ff1cb54705651fe504177" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.816639 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb"] Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.820427 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c669855c-w7rbb"] Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.832728 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z"] Feb 25 06:51:09 crc kubenswrapper[4978]: E0225 06:51:09.832999 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eca509d4-31d3-43da-8ad3-2932d080ef40" containerName="route-controller-manager" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.833018 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="eca509d4-31d3-43da-8ad3-2932d080ef40" containerName="route-controller-manager" Feb 25 06:51:09 crc kubenswrapper[4978]: E0225 06:51:09.833035 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8765fb-fa99-4b4f-82e5-cee23cd14381" containerName="controller-manager" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.833043 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8765fb-fa99-4b4f-82e5-cee23cd14381" containerName="controller-manager" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.833150 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="eca509d4-31d3-43da-8ad3-2932d080ef40" containerName="route-controller-manager" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.833171 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e8765fb-fa99-4b4f-82e5-cee23cd14381" containerName="controller-manager" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.833638 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.836219 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.840391 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.840417 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.840551 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.840675 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.841229 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.845258 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d6fb8f486-d42z9"] Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.846249 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.851558 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z"] Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.855550 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d6fb8f486-d42z9"] Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.862460 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.862832 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.863046 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.864237 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.864699 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.865583 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.867321 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.992361 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8qtx\" (UniqueName: \"kubernetes.io/projected/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-kube-api-access-g8qtx\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.992478 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-proxy-ca-bundles\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.992539 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-config\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.992573 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-serving-cert\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.992607 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-client-ca\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.992733 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q8j6\" (UniqueName: \"kubernetes.io/projected/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-kube-api-access-7q8j6\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.992802 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-client-ca\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.992955 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-config\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:09 crc kubenswrapper[4978]: I0225 06:51:09.993038 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-serving-cert\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.094257 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-client-ca\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.094309 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q8j6\" (UniqueName: \"kubernetes.io/projected/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-kube-api-access-7q8j6\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.094337 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-client-ca\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.094395 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-config\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.094415 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-serving-cert\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.094454 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8qtx\" (UniqueName: \"kubernetes.io/projected/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-kube-api-access-g8qtx\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.094471 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-proxy-ca-bundles\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.094499 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-config\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.094519 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-serving-cert\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.095534 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-client-ca\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.095838 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-config\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.096151 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-config\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.096564 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-proxy-ca-bundles\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.096676 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-client-ca\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.105094 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-serving-cert\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.110888 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q8j6\" (UniqueName: \"kubernetes.io/projected/3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3-kube-api-access-7q8j6\") pod \"controller-manager-6d6fb8f486-d42z9\" (UID: \"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3\") " pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.114974 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-serving-cert\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.115960 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8qtx\" (UniqueName: \"kubernetes.io/projected/d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11-kube-api-access-g8qtx\") pod \"route-controller-manager-64b664c989-mnm8z\" (UID: \"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11\") " pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.153355 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.173183 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.361485 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d6fb8f486-d42z9"] Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.624340 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z"] Feb 25 06:51:10 crc kubenswrapper[4978]: W0225 06:51:10.637835 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8e2f38f_b3bb_4ee8_ab4c_6aaeff667a11.slice/crio-a6a18a0004bdeaf7a30dbf9611b2aeab37c96638128588c1f59dca1381617262 WatchSource:0}: Error finding container a6a18a0004bdeaf7a30dbf9611b2aeab37c96638128588c1f59dca1381617262: Status 404 returned error can't find the container with id a6a18a0004bdeaf7a30dbf9611b2aeab37c96638128588c1f59dca1381617262 Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.790269 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" event={"ID":"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11","Type":"ContainerStarted","Data":"7c4aa76e74d4a5cec109dba11e81cb7cb6b81de0b8d4e269c4c1d5bffecf9194"} Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.790326 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" event={"ID":"d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11","Type":"ContainerStarted","Data":"a6a18a0004bdeaf7a30dbf9611b2aeab37c96638128588c1f59dca1381617262"} Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.790352 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.792942 4978 patch_prober.go:28] interesting pod/route-controller-manager-64b664c989-mnm8z container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.72:8443/healthz\": dial tcp 10.217.0.72:8443: connect: connection refused" start-of-body= Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.793006 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" event={"ID":"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3","Type":"ContainerStarted","Data":"f8accefef02d71ee8323ca897fa60a5e0669dccf3ad1b6ddd8aa235afe600e2f"} Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.793015 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" podUID="d8e2f38f-b3bb-4ee8-ab4c-6aaeff667a11" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.72:8443/healthz\": dial tcp 10.217.0.72:8443: connect: connection refused" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.793066 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" event={"ID":"3feb25fb-5d3c-496d-9c6c-51c1b05eb1b3","Type":"ContainerStarted","Data":"b450b156b272b82075e37a3531277ab9f5dee2a67c7dfba4ac8ee0f557256ac6"} Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.793313 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.802518 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.809333 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" podStartSLOduration=2.809315302 podStartE2EDuration="2.809315302s" podCreationTimestamp="2026-02-25 06:51:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:51:10.808589259 +0000 UTC m=+364.247845728" watchObservedRunningTime="2026-02-25 06:51:10.809315302 +0000 UTC m=+364.248571761" Feb 25 06:51:10 crc kubenswrapper[4978]: I0225 06:51:10.830047 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d6fb8f486-d42z9" podStartSLOduration=2.830026232 podStartE2EDuration="2.830026232s" podCreationTimestamp="2026-02-25 06:51:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:51:10.82871762 +0000 UTC m=+364.267974089" watchObservedRunningTime="2026-02-25 06:51:10.830026232 +0000 UTC m=+364.269282701" Feb 25 06:51:11 crc kubenswrapper[4978]: I0225 06:51:11.336225 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e8765fb-fa99-4b4f-82e5-cee23cd14381" path="/var/lib/kubelet/pods/4e8765fb-fa99-4b4f-82e5-cee23cd14381/volumes" Feb 25 06:51:11 crc kubenswrapper[4978]: I0225 06:51:11.337285 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eca509d4-31d3-43da-8ad3-2932d080ef40" path="/var/lib/kubelet/pods/eca509d4-31d3-43da-8ad3-2932d080ef40/volumes" Feb 25 06:51:11 crc kubenswrapper[4978]: I0225 06:51:11.810058 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-64b664c989-mnm8z" Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.835787 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6fbkc"] Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.839163 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6fbkc" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" containerName="registry-server" containerID="cri-o://ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b" gracePeriod=30 Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.864240 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h9wx2"] Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.868689 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h9wx2" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerName="registry-server" containerID="cri-o://84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31" gracePeriod=30 Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.871255 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qf9f"] Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.871620 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" podUID="9b843a14-9586-4f29-bfba-8413f4e52104" containerName="marketplace-operator" containerID="cri-o://55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa" gracePeriod=30 Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.889721 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9fsk"] Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.890138 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v9fsk" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerName="registry-server" containerID="cri-o://f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8" gracePeriod=30 Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.911495 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cghmv"] Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.911840 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cghmv" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerName="registry-server" containerID="cri-o://157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9" gracePeriod=30 Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.915936 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hpjsn"] Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.917291 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:17 crc kubenswrapper[4978]: I0225 06:51:17.928636 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hpjsn"] Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.013384 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ecf97051-f58e-4888-8e7e-f03481e89bea-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hpjsn\" (UID: \"ecf97051-f58e-4888-8e7e-f03481e89bea\") " pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.013495 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ecf97051-f58e-4888-8e7e-f03481e89bea-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hpjsn\" (UID: \"ecf97051-f58e-4888-8e7e-f03481e89bea\") " pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.013539 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b75m\" (UniqueName: \"kubernetes.io/projected/ecf97051-f58e-4888-8e7e-f03481e89bea-kube-api-access-9b75m\") pod \"marketplace-operator-79b997595-hpjsn\" (UID: \"ecf97051-f58e-4888-8e7e-f03481e89bea\") " pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.114721 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ecf97051-f58e-4888-8e7e-f03481e89bea-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hpjsn\" (UID: \"ecf97051-f58e-4888-8e7e-f03481e89bea\") " pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.114801 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ecf97051-f58e-4888-8e7e-f03481e89bea-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hpjsn\" (UID: \"ecf97051-f58e-4888-8e7e-f03481e89bea\") " pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.114830 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b75m\" (UniqueName: \"kubernetes.io/projected/ecf97051-f58e-4888-8e7e-f03481e89bea-kube-api-access-9b75m\") pod \"marketplace-operator-79b997595-hpjsn\" (UID: \"ecf97051-f58e-4888-8e7e-f03481e89bea\") " pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.116072 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ecf97051-f58e-4888-8e7e-f03481e89bea-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hpjsn\" (UID: \"ecf97051-f58e-4888-8e7e-f03481e89bea\") " pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.121282 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ecf97051-f58e-4888-8e7e-f03481e89bea-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hpjsn\" (UID: \"ecf97051-f58e-4888-8e7e-f03481e89bea\") " pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.135742 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b75m\" (UniqueName: \"kubernetes.io/projected/ecf97051-f58e-4888-8e7e-f03481e89bea-kube-api-access-9b75m\") pod \"marketplace-operator-79b997595-hpjsn\" (UID: \"ecf97051-f58e-4888-8e7e-f03481e89bea\") " pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.366925 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.452819 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.623453 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvlxt\" (UniqueName: \"kubernetes.io/projected/3054616f-87b0-467e-b5a7-499e3addf68a-kube-api-access-gvlxt\") pod \"3054616f-87b0-467e-b5a7-499e3addf68a\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.623576 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-utilities\") pod \"3054616f-87b0-467e-b5a7-499e3addf68a\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.623631 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-catalog-content\") pod \"3054616f-87b0-467e-b5a7-499e3addf68a\" (UID: \"3054616f-87b0-467e-b5a7-499e3addf68a\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.631047 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-utilities" (OuterVolumeSpecName: "utilities") pod "3054616f-87b0-467e-b5a7-499e3addf68a" (UID: "3054616f-87b0-467e-b5a7-499e3addf68a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.641609 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3054616f-87b0-467e-b5a7-499e3addf68a-kube-api-access-gvlxt" (OuterVolumeSpecName: "kube-api-access-gvlxt") pod "3054616f-87b0-467e-b5a7-499e3addf68a" (UID: "3054616f-87b0-467e-b5a7-499e3addf68a"). InnerVolumeSpecName "kube-api-access-gvlxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.643994 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.652054 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.666770 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.681716 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3054616f-87b0-467e-b5a7-499e3addf68a" (UID: "3054616f-87b0-467e-b5a7-499e3addf68a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.690240 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.725291 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.725334 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3054616f-87b0-467e-b5a7-499e3addf68a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.725350 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvlxt\" (UniqueName: \"kubernetes.io/projected/3054616f-87b0-467e-b5a7-499e3addf68a-kube-api-access-gvlxt\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.826800 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-catalog-content\") pod \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.826850 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-catalog-content\") pod \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.826876 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-catalog-content\") pod \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.826914 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-utilities\") pod \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.826972 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-trusted-ca\") pod \"9b843a14-9586-4f29-bfba-8413f4e52104\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.826990 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-utilities\") pod \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.827016 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w2qq\" (UniqueName: \"kubernetes.io/projected/8ae5f061-fdff-4fc5-b252-4734b7ad892a-kube-api-access-8w2qq\") pod \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.827049 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtlhx\" (UniqueName: \"kubernetes.io/projected/9b843a14-9586-4f29-bfba-8413f4e52104-kube-api-access-rtlhx\") pod \"9b843a14-9586-4f29-bfba-8413f4e52104\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.827074 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctx5l\" (UniqueName: \"kubernetes.io/projected/6f6ed9c0-cb59-4aac-af72-2633f18378c9-kube-api-access-ctx5l\") pod \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\" (UID: \"6f6ed9c0-cb59-4aac-af72-2633f18378c9\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.827112 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvwwx\" (UniqueName: \"kubernetes.io/projected/9283fdd2-ab8a-48ca-a326-1a0b107671a0-kube-api-access-gvwwx\") pod \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\" (UID: \"9283fdd2-ab8a-48ca-a326-1a0b107671a0\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.827133 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-utilities\") pod \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\" (UID: \"8ae5f061-fdff-4fc5-b252-4734b7ad892a\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.827163 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-operator-metrics\") pod \"9b843a14-9586-4f29-bfba-8413f4e52104\" (UID: \"9b843a14-9586-4f29-bfba-8413f4e52104\") " Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.831280 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "9b843a14-9586-4f29-bfba-8413f4e52104" (UID: "9b843a14-9586-4f29-bfba-8413f4e52104"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.832681 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ae5f061-fdff-4fc5-b252-4734b7ad892a-kube-api-access-8w2qq" (OuterVolumeSpecName: "kube-api-access-8w2qq") pod "8ae5f061-fdff-4fc5-b252-4734b7ad892a" (UID: "8ae5f061-fdff-4fc5-b252-4734b7ad892a"). InnerVolumeSpecName "kube-api-access-8w2qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.834034 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-utilities" (OuterVolumeSpecName: "utilities") pod "6f6ed9c0-cb59-4aac-af72-2633f18378c9" (UID: "6f6ed9c0-cb59-4aac-af72-2633f18378c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.834991 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "9b843a14-9586-4f29-bfba-8413f4e52104" (UID: "9b843a14-9586-4f29-bfba-8413f4e52104"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.837063 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b843a14-9586-4f29-bfba-8413f4e52104-kube-api-access-rtlhx" (OuterVolumeSpecName: "kube-api-access-rtlhx") pod "9b843a14-9586-4f29-bfba-8413f4e52104" (UID: "9b843a14-9586-4f29-bfba-8413f4e52104"). InnerVolumeSpecName "kube-api-access-rtlhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.837155 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-utilities" (OuterVolumeSpecName: "utilities") pod "8ae5f061-fdff-4fc5-b252-4734b7ad892a" (UID: "8ae5f061-fdff-4fc5-b252-4734b7ad892a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.838558 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-utilities" (OuterVolumeSpecName: "utilities") pod "9283fdd2-ab8a-48ca-a326-1a0b107671a0" (UID: "9283fdd2-ab8a-48ca-a326-1a0b107671a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.838623 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f6ed9c0-cb59-4aac-af72-2633f18378c9-kube-api-access-ctx5l" (OuterVolumeSpecName: "kube-api-access-ctx5l") pod "6f6ed9c0-cb59-4aac-af72-2633f18378c9" (UID: "6f6ed9c0-cb59-4aac-af72-2633f18378c9"). InnerVolumeSpecName "kube-api-access-ctx5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.846217 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9283fdd2-ab8a-48ca-a326-1a0b107671a0-kube-api-access-gvwwx" (OuterVolumeSpecName: "kube-api-access-gvwwx") pod "9283fdd2-ab8a-48ca-a326-1a0b107671a0" (UID: "9283fdd2-ab8a-48ca-a326-1a0b107671a0"). InnerVolumeSpecName "kube-api-access-gvwwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.860582 4978 generic.go:334] "Generic (PLEG): container finished" podID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerID="f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8" exitCode=0 Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.860683 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v9fsk" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.860691 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9fsk" event={"ID":"8ae5f061-fdff-4fc5-b252-4734b7ad892a","Type":"ContainerDied","Data":"f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8"} Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.861478 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v9fsk" event={"ID":"8ae5f061-fdff-4fc5-b252-4734b7ad892a","Type":"ContainerDied","Data":"007e2ab9c79f9fab639dfdf6d2fedcdb60166a5acdb28a35d1aa9d22a3c18ec7"} Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.861525 4978 scope.go:117] "RemoveContainer" containerID="f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.865467 4978 generic.go:334] "Generic (PLEG): container finished" podID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerID="84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31" exitCode=0 Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.865536 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9wx2" event={"ID":"9283fdd2-ab8a-48ca-a326-1a0b107671a0","Type":"ContainerDied","Data":"84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31"} Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.865567 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h9wx2" event={"ID":"9283fdd2-ab8a-48ca-a326-1a0b107671a0","Type":"ContainerDied","Data":"f4455be87d9d95d12c5a65362832348affe098310322398912a17232c2146f10"} Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.865642 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h9wx2" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.866057 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ae5f061-fdff-4fc5-b252-4734b7ad892a" (UID: "8ae5f061-fdff-4fc5-b252-4734b7ad892a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.870158 4978 generic.go:334] "Generic (PLEG): container finished" podID="3054616f-87b0-467e-b5a7-499e3addf68a" containerID="ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b" exitCode=0 Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.870211 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fbkc" event={"ID":"3054616f-87b0-467e-b5a7-499e3addf68a","Type":"ContainerDied","Data":"ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b"} Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.870234 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fbkc" event={"ID":"3054616f-87b0-467e-b5a7-499e3addf68a","Type":"ContainerDied","Data":"e23e2633f2ce8205b0d62b2109ade68af4e8bdc5c8f8f8d6f5c0a62134ea0a8e"} Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.870216 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6fbkc" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.874834 4978 generic.go:334] "Generic (PLEG): container finished" podID="9b843a14-9586-4f29-bfba-8413f4e52104" containerID="55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa" exitCode=0 Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.874874 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" event={"ID":"9b843a14-9586-4f29-bfba-8413f4e52104","Type":"ContainerDied","Data":"55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa"} Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.874915 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" event={"ID":"9b843a14-9586-4f29-bfba-8413f4e52104","Type":"ContainerDied","Data":"826f5f8329047bb1483d89d2a9a156c5e63fb37b32383340c4781bae3e34d2a8"} Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.874854 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9qf9f" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.877084 4978 generic.go:334] "Generic (PLEG): container finished" podID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerID="157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9" exitCode=0 Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.877121 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cghmv" event={"ID":"6f6ed9c0-cb59-4aac-af72-2633f18378c9","Type":"ContainerDied","Data":"157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9"} Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.877169 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cghmv" event={"ID":"6f6ed9c0-cb59-4aac-af72-2633f18378c9","Type":"ContainerDied","Data":"671793f646d39d34d80a7eaf6dff6f51fd8cf43ef70f7de1cb983ba7d9261b52"} Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.877273 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cghmv" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.886232 4978 scope.go:117] "RemoveContainer" containerID="271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.892803 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9283fdd2-ab8a-48ca-a326-1a0b107671a0" (UID: "9283fdd2-ab8a-48ca-a326-1a0b107671a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.911475 4978 scope.go:117] "RemoveContainer" containerID="33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.919744 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6fbkc"] Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.923650 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6fbkc"] Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.926926 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hpjsn"] Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929225 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929284 4978 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929299 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929310 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w2qq\" (UniqueName: \"kubernetes.io/projected/8ae5f061-fdff-4fc5-b252-4734b7ad892a-kube-api-access-8w2qq\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929346 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtlhx\" (UniqueName: \"kubernetes.io/projected/9b843a14-9586-4f29-bfba-8413f4e52104-kube-api-access-rtlhx\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929360 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctx5l\" (UniqueName: \"kubernetes.io/projected/6f6ed9c0-cb59-4aac-af72-2633f18378c9-kube-api-access-ctx5l\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929401 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvwwx\" (UniqueName: \"kubernetes.io/projected/9283fdd2-ab8a-48ca-a326-1a0b107671a0-kube-api-access-gvwwx\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929413 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929425 4978 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9b843a14-9586-4f29-bfba-8413f4e52104-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929435 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9283fdd2-ab8a-48ca-a326-1a0b107671a0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.929446 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ae5f061-fdff-4fc5-b252-4734b7ad892a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.936858 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qf9f"] Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.949948 4978 scope.go:117] "RemoveContainer" containerID="f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.951677 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9qf9f"] Feb 25 06:51:18 crc kubenswrapper[4978]: E0225 06:51:18.952501 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8\": container with ID starting with f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8 not found: ID does not exist" containerID="f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.952641 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8"} err="failed to get container status \"f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8\": rpc error: code = NotFound desc = could not find container \"f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8\": container with ID starting with f9a221fd1176ed7d0595fd74b2d5e7633a867dbef716cecfb31aa37d52767da8 not found: ID does not exist" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.952761 4978 scope.go:117] "RemoveContainer" containerID="271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c" Feb 25 06:51:18 crc kubenswrapper[4978]: E0225 06:51:18.953393 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c\": container with ID starting with 271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c not found: ID does not exist" containerID="271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.953453 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c"} err="failed to get container status \"271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c\": rpc error: code = NotFound desc = could not find container \"271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c\": container with ID starting with 271ccb26634b131b4955ad174d5be54b5540540d9d213eadb6035024a3a5e29c not found: ID does not exist" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.953493 4978 scope.go:117] "RemoveContainer" containerID="33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5" Feb 25 06:51:18 crc kubenswrapper[4978]: E0225 06:51:18.953819 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5\": container with ID starting with 33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5 not found: ID does not exist" containerID="33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.953860 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5"} err="failed to get container status \"33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5\": rpc error: code = NotFound desc = could not find container \"33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5\": container with ID starting with 33aea8073a0986de7c671c7d809a49161b534b106d73ba77d388f7569d6cbad5 not found: ID does not exist" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.953886 4978 scope.go:117] "RemoveContainer" containerID="84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31" Feb 25 06:51:18 crc kubenswrapper[4978]: I0225 06:51:18.979141 4978 scope.go:117] "RemoveContainer" containerID="cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.068598 4978 scope.go:117] "RemoveContainer" containerID="b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.072253 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f6ed9c0-cb59-4aac-af72-2633f18378c9" (UID: "6f6ed9c0-cb59-4aac-af72-2633f18378c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.098401 4978 scope.go:117] "RemoveContainer" containerID="84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.107510 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31\": container with ID starting with 84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31 not found: ID does not exist" containerID="84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.107571 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31"} err="failed to get container status \"84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31\": rpc error: code = NotFound desc = could not find container \"84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31\": container with ID starting with 84a59cfdcb05ba2a790be6073aec1dd1197350a73e1b45dff8f971af6dbe1d31 not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.107610 4978 scope.go:117] "RemoveContainer" containerID="cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.114481 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd\": container with ID starting with cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd not found: ID does not exist" containerID="cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.114538 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd"} err="failed to get container status \"cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd\": rpc error: code = NotFound desc = could not find container \"cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd\": container with ID starting with cb08740c0e3b2e9c336803686e4018806bfda18b55c4a0a4b83028f910274fdd not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.114568 4978 scope.go:117] "RemoveContainer" containerID="b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.119526 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a\": container with ID starting with b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a not found: ID does not exist" containerID="b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.119580 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a"} err="failed to get container status \"b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a\": rpc error: code = NotFound desc = could not find container \"b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a\": container with ID starting with b095dc728d8a15968562d42155603fce4dc11109f1507c75aec8f3088b0e368a not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.119616 4978 scope.go:117] "RemoveContainer" containerID="ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.139146 4978 scope.go:117] "RemoveContainer" containerID="0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.143795 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f6ed9c0-cb59-4aac-af72-2633f18378c9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.155138 4978 scope.go:117] "RemoveContainer" containerID="c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.172341 4978 scope.go:117] "RemoveContainer" containerID="ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.179189 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b\": container with ID starting with ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b not found: ID does not exist" containerID="ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.179218 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b"} err="failed to get container status \"ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b\": rpc error: code = NotFound desc = could not find container \"ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b\": container with ID starting with ef40afd85fe5438156dbd7daa95f382debe82c207ee08f4daeb148e30bbcb19b not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.179246 4978 scope.go:117] "RemoveContainer" containerID="0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.179933 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a\": container with ID starting with 0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a not found: ID does not exist" containerID="0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.179959 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a"} err="failed to get container status \"0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a\": rpc error: code = NotFound desc = could not find container \"0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a\": container with ID starting with 0225f9ce755a160e70eda6763242b66a2769f6c49a084f63d3efe8a950829b8a not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.179974 4978 scope.go:117] "RemoveContainer" containerID="c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.180172 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76\": container with ID starting with c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76 not found: ID does not exist" containerID="c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.180210 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76"} err="failed to get container status \"c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76\": rpc error: code = NotFound desc = could not find container \"c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76\": container with ID starting with c03a1ec8dcfad96a40100248c57422b766f17bc958ed0000bede0dc57b41cb76 not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.180226 4978 scope.go:117] "RemoveContainer" containerID="55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.194584 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9fsk"] Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.200501 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v9fsk"] Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.207390 4978 scope.go:117] "RemoveContainer" containerID="c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.217514 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h9wx2"] Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.234783 4978 scope.go:117] "RemoveContainer" containerID="55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.235408 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa\": container with ID starting with 55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa not found: ID does not exist" containerID="55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.235492 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa"} err="failed to get container status \"55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa\": rpc error: code = NotFound desc = could not find container \"55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa\": container with ID starting with 55ffb41ad40006ce75a19f7418d417b160d78d3dae9626df62c76f7d98b2e0aa not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.235529 4978 scope.go:117] "RemoveContainer" containerID="c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.236151 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c\": container with ID starting with c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c not found: ID does not exist" containerID="c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.236221 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c"} err="failed to get container status \"c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c\": rpc error: code = NotFound desc = could not find container \"c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c\": container with ID starting with c2b044a8879c9b3c8fd9050484154568e2b6a7c42633c3c61508643b8f0aec1c not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.236274 4978 scope.go:117] "RemoveContainer" containerID="157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.240949 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h9wx2"] Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.245305 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cghmv"] Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.248805 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cghmv"] Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.251756 4978 scope.go:117] "RemoveContainer" containerID="6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.268735 4978 scope.go:117] "RemoveContainer" containerID="5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.284141 4978 scope.go:117] "RemoveContainer" containerID="157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.284674 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9\": container with ID starting with 157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9 not found: ID does not exist" containerID="157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.284713 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9"} err="failed to get container status \"157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9\": rpc error: code = NotFound desc = could not find container \"157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9\": container with ID starting with 157c5d351e191fa1be4d1f9858f399b9df5d44d2b5f56e3f81210b5a74889aa9 not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.284735 4978 scope.go:117] "RemoveContainer" containerID="6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.284933 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09\": container with ID starting with 6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09 not found: ID does not exist" containerID="6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.284958 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09"} err="failed to get container status \"6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09\": rpc error: code = NotFound desc = could not find container \"6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09\": container with ID starting with 6db1860a33f44a0189955b725ede27f3af280ab84b82ff17488c03bcf3476e09 not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.284977 4978 scope.go:117] "RemoveContainer" containerID="5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.285169 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab\": container with ID starting with 5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab not found: ID does not exist" containerID="5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.285187 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab"} err="failed to get container status \"5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab\": rpc error: code = NotFound desc = could not find container \"5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab\": container with ID starting with 5c227b570faf633768ea94991ea8325737db6d695cc6c79195c2cd8103e651ab not found: ID does not exist" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.341709 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" path="/var/lib/kubelet/pods/3054616f-87b0-467e-b5a7-499e3addf68a/volumes" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.342641 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" path="/var/lib/kubelet/pods/6f6ed9c0-cb59-4aac-af72-2633f18378c9/volumes" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.343490 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" path="/var/lib/kubelet/pods/8ae5f061-fdff-4fc5-b252-4734b7ad892a/volumes" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.344944 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" path="/var/lib/kubelet/pods/9283fdd2-ab8a-48ca-a326-1a0b107671a0/volumes" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.345791 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b843a14-9586-4f29-bfba-8413f4e52104" path="/var/lib/kubelet/pods/9b843a14-9586-4f29-bfba-8413f4e52104/volumes" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.452740 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p9gbz"] Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453019 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerName="extract-content" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453040 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerName="extract-content" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453060 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" containerName="extract-content" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453072 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" containerName="extract-content" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453091 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453104 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453122 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" containerName="extract-utilities" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453134 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" containerName="extract-utilities" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453154 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerName="extract-content" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453165 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerName="extract-content" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453185 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerName="extract-utilities" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453198 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerName="extract-utilities" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453225 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerName="extract-utilities" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453237 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerName="extract-utilities" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453253 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerName="extract-utilities" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453266 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerName="extract-utilities" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453280 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453292 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453307 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerName="extract-content" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453319 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerName="extract-content" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453335 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453348 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453416 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b843a14-9586-4f29-bfba-8413f4e52104" containerName="marketplace-operator" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453431 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b843a14-9586-4f29-bfba-8413f4e52104" containerName="marketplace-operator" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453448 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b843a14-9586-4f29-bfba-8413f4e52104" containerName="marketplace-operator" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453460 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b843a14-9586-4f29-bfba-8413f4e52104" containerName="marketplace-operator" Feb 25 06:51:19 crc kubenswrapper[4978]: E0225 06:51:19.453473 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453486 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453635 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b843a14-9586-4f29-bfba-8413f4e52104" containerName="marketplace-operator" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453659 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9283fdd2-ab8a-48ca-a326-1a0b107671a0" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453676 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f6ed9c0-cb59-4aac-af72-2633f18378c9" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453696 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ae5f061-fdff-4fc5-b252-4734b7ad892a" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453715 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b843a14-9586-4f29-bfba-8413f4e52104" containerName="marketplace-operator" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.453735 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3054616f-87b0-467e-b5a7-499e3addf68a" containerName="registry-server" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.454874 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.457258 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.463945 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9gbz"] Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.650570 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-utilities\") pod \"certified-operators-p9gbz\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.650634 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxpz4\" (UniqueName: \"kubernetes.io/projected/376ae4f8-8441-46a8-ab12-07d3f9c729d9-kube-api-access-wxpz4\") pod \"certified-operators-p9gbz\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.650734 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-catalog-content\") pod \"certified-operators-p9gbz\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.752000 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-catalog-content\") pod \"certified-operators-p9gbz\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.752163 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-utilities\") pod \"certified-operators-p9gbz\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.752252 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxpz4\" (UniqueName: \"kubernetes.io/projected/376ae4f8-8441-46a8-ab12-07d3f9c729d9-kube-api-access-wxpz4\") pod \"certified-operators-p9gbz\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.752759 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-catalog-content\") pod \"certified-operators-p9gbz\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.752797 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-utilities\") pod \"certified-operators-p9gbz\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.776635 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxpz4\" (UniqueName: \"kubernetes.io/projected/376ae4f8-8441-46a8-ab12-07d3f9c729d9-kube-api-access-wxpz4\") pod \"certified-operators-p9gbz\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.892055 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" event={"ID":"ecf97051-f58e-4888-8e7e-f03481e89bea","Type":"ContainerStarted","Data":"7afc5a28aa1bfea1d8f3a0df6026fece80d32a641d8a922f21b8aeb5331a99a8"} Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.892134 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" event={"ID":"ecf97051-f58e-4888-8e7e-f03481e89bea","Type":"ContainerStarted","Data":"a202bb7cc991f8237efd7be5ab252faf436ddb2b57b7040aa2a9de4c4a1c0f1d"} Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.892464 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.898410 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" Feb 25 06:51:19 crc kubenswrapper[4978]: I0225 06:51:19.917831 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hpjsn" podStartSLOduration=2.917805954 podStartE2EDuration="2.917805954s" podCreationTimestamp="2026-02-25 06:51:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:51:19.912264035 +0000 UTC m=+373.351520494" watchObservedRunningTime="2026-02-25 06:51:19.917805954 +0000 UTC m=+373.357062423" Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.074113 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.463736 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p9gbz"] Feb 25 06:51:20 crc kubenswrapper[4978]: W0225 06:51:20.471293 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod376ae4f8_8441_46a8_ab12_07d3f9c729d9.slice/crio-62a9b13fbcfb83dd4141d27fb137a93e25a27aec9e0388d2a9c7d562be295a5a WatchSource:0}: Error finding container 62a9b13fbcfb83dd4141d27fb137a93e25a27aec9e0388d2a9c7d562be295a5a: Status 404 returned error can't find the container with id 62a9b13fbcfb83dd4141d27fb137a93e25a27aec9e0388d2a9c7d562be295a5a Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.857276 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dgg7r"] Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.858260 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.862819 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.865359 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgg7r"] Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.903776 4978 generic.go:334] "Generic (PLEG): container finished" podID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerID="651089a42eddc2c0ef7fb45e03a8c62470d00477605851084e76764c43ac5288" exitCode=0 Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.903881 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9gbz" event={"ID":"376ae4f8-8441-46a8-ab12-07d3f9c729d9","Type":"ContainerDied","Data":"651089a42eddc2c0ef7fb45e03a8c62470d00477605851084e76764c43ac5288"} Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.903920 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9gbz" event={"ID":"376ae4f8-8441-46a8-ab12-07d3f9c729d9","Type":"ContainerStarted","Data":"62a9b13fbcfb83dd4141d27fb137a93e25a27aec9e0388d2a9c7d562be295a5a"} Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.968993 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-utilities\") pod \"redhat-marketplace-dgg7r\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.969046 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn94s\" (UniqueName: \"kubernetes.io/projected/e4a76118-7308-4b0f-aae5-41cd6888a845-kube-api-access-sn94s\") pod \"redhat-marketplace-dgg7r\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:20 crc kubenswrapper[4978]: I0225 06:51:20.969579 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-catalog-content\") pod \"redhat-marketplace-dgg7r\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.077071 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-utilities\") pod \"redhat-marketplace-dgg7r\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.077130 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn94s\" (UniqueName: \"kubernetes.io/projected/e4a76118-7308-4b0f-aae5-41cd6888a845-kube-api-access-sn94s\") pod \"redhat-marketplace-dgg7r\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.077198 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-catalog-content\") pod \"redhat-marketplace-dgg7r\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.077635 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-catalog-content\") pod \"redhat-marketplace-dgg7r\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.080106 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-utilities\") pod \"redhat-marketplace-dgg7r\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.102581 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn94s\" (UniqueName: \"kubernetes.io/projected/e4a76118-7308-4b0f-aae5-41cd6888a845-kube-api-access-sn94s\") pod \"redhat-marketplace-dgg7r\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.172902 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.577530 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgg7r"] Feb 25 06:51:21 crc kubenswrapper[4978]: W0225 06:51:21.585996 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4a76118_7308_4b0f_aae5_41cd6888a845.slice/crio-bea11dacbf3072b566b229e9a09f5d37cd7f3247015fff92cec2410b75e42c3e WatchSource:0}: Error finding container bea11dacbf3072b566b229e9a09f5d37cd7f3247015fff92cec2410b75e42c3e: Status 404 returned error can't find the container with id bea11dacbf3072b566b229e9a09f5d37cd7f3247015fff92cec2410b75e42c3e Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.847822 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-krctw"] Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.849480 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.851741 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.859427 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krctw"] Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.887040 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-catalog-content\") pod \"redhat-operators-krctw\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.887125 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg6kp\" (UniqueName: \"kubernetes.io/projected/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-kube-api-access-qg6kp\") pod \"redhat-operators-krctw\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.887181 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-utilities\") pod \"redhat-operators-krctw\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.910979 4978 generic.go:334] "Generic (PLEG): container finished" podID="e4a76118-7308-4b0f-aae5-41cd6888a845" containerID="0920ec34fa9aa9b1576d58fbbdfa0613a2278d444564b9929aa479fb7fec7bfa" exitCode=0 Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.911231 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgg7r" event={"ID":"e4a76118-7308-4b0f-aae5-41cd6888a845","Type":"ContainerDied","Data":"0920ec34fa9aa9b1576d58fbbdfa0613a2278d444564b9929aa479fb7fec7bfa"} Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.911310 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgg7r" event={"ID":"e4a76118-7308-4b0f-aae5-41cd6888a845","Type":"ContainerStarted","Data":"bea11dacbf3072b566b229e9a09f5d37cd7f3247015fff92cec2410b75e42c3e"} Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.988581 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-utilities\") pod \"redhat-operators-krctw\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.988760 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-catalog-content\") pod \"redhat-operators-krctw\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.988841 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg6kp\" (UniqueName: \"kubernetes.io/projected/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-kube-api-access-qg6kp\") pod \"redhat-operators-krctw\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.989521 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-utilities\") pod \"redhat-operators-krctw\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:21 crc kubenswrapper[4978]: I0225 06:51:21.989531 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-catalog-content\") pod \"redhat-operators-krctw\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:22 crc kubenswrapper[4978]: I0225 06:51:22.012642 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg6kp\" (UniqueName: \"kubernetes.io/projected/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-kube-api-access-qg6kp\") pod \"redhat-operators-krctw\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:22 crc kubenswrapper[4978]: I0225 06:51:22.179821 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:22 crc kubenswrapper[4978]: I0225 06:51:22.596719 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krctw"] Feb 25 06:51:22 crc kubenswrapper[4978]: I0225 06:51:22.862848 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-plhd2" Feb 25 06:51:22 crc kubenswrapper[4978]: I0225 06:51:22.911763 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w6cv7"] Feb 25 06:51:22 crc kubenswrapper[4978]: I0225 06:51:22.928553 4978 generic.go:334] "Generic (PLEG): container finished" podID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerID="de7ecaac1189f75bd90f33d54a8d89e205e6520a488f4046c8e09e86bbcd6e57" exitCode=0 Feb 25 06:51:22 crc kubenswrapper[4978]: I0225 06:51:22.928630 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krctw" event={"ID":"8520c8a9-3af3-465a-8c1d-aa9d10c36c40","Type":"ContainerDied","Data":"de7ecaac1189f75bd90f33d54a8d89e205e6520a488f4046c8e09e86bbcd6e57"} Feb 25 06:51:22 crc kubenswrapper[4978]: I0225 06:51:22.928655 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krctw" event={"ID":"8520c8a9-3af3-465a-8c1d-aa9d10c36c40","Type":"ContainerStarted","Data":"0976048df8fa87410d01567784b369b848bf3e8022d1889e49a008d23f0235ee"} Feb 25 06:51:22 crc kubenswrapper[4978]: I0225 06:51:22.933199 4978 generic.go:334] "Generic (PLEG): container finished" podID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerID="483743f33f036db29860af77324099774dafa4d24b6780d28576261c2a1843ed" exitCode=0 Feb 25 06:51:22 crc kubenswrapper[4978]: I0225 06:51:22.933526 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9gbz" event={"ID":"376ae4f8-8441-46a8-ab12-07d3f9c729d9","Type":"ContainerDied","Data":"483743f33f036db29860af77324099774dafa4d24b6780d28576261c2a1843ed"} Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.248782 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9qw7l"] Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.252858 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.255926 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.270943 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9qw7l"] Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.317256 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4ff8\" (UniqueName: \"kubernetes.io/projected/75fbd0de-dd88-4079-a5c9-72aff9976339-kube-api-access-g4ff8\") pod \"community-operators-9qw7l\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.317340 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-catalog-content\") pod \"community-operators-9qw7l\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.317411 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-utilities\") pod \"community-operators-9qw7l\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.418102 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-utilities\") pod \"community-operators-9qw7l\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.418202 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4ff8\" (UniqueName: \"kubernetes.io/projected/75fbd0de-dd88-4079-a5c9-72aff9976339-kube-api-access-g4ff8\") pod \"community-operators-9qw7l\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.418249 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-catalog-content\") pod \"community-operators-9qw7l\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.419068 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-catalog-content\") pod \"community-operators-9qw7l\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.419074 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-utilities\") pod \"community-operators-9qw7l\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.440695 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4ff8\" (UniqueName: \"kubernetes.io/projected/75fbd0de-dd88-4079-a5c9-72aff9976339-kube-api-access-g4ff8\") pod \"community-operators-9qw7l\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.619608 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.944747 4978 generic.go:334] "Generic (PLEG): container finished" podID="e4a76118-7308-4b0f-aae5-41cd6888a845" containerID="64c441d2520e340c4c91ad918adeec0be7d87420a2b916096c443dd6208f6e81" exitCode=0 Feb 25 06:51:23 crc kubenswrapper[4978]: I0225 06:51:23.944811 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgg7r" event={"ID":"e4a76118-7308-4b0f-aae5-41cd6888a845","Type":"ContainerDied","Data":"64c441d2520e340c4c91ad918adeec0be7d87420a2b916096c443dd6208f6e81"} Feb 25 06:51:24 crc kubenswrapper[4978]: I0225 06:51:24.039085 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9qw7l"] Feb 25 06:51:24 crc kubenswrapper[4978]: W0225 06:51:24.048974 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75fbd0de_dd88_4079_a5c9_72aff9976339.slice/crio-7248abf77ec7dfb8cbdfad57f49f813fe18f666ce7003c67ab00beb7cebb16b2 WatchSource:0}: Error finding container 7248abf77ec7dfb8cbdfad57f49f813fe18f666ce7003c67ab00beb7cebb16b2: Status 404 returned error can't find the container with id 7248abf77ec7dfb8cbdfad57f49f813fe18f666ce7003c67ab00beb7cebb16b2 Feb 25 06:51:24 crc kubenswrapper[4978]: I0225 06:51:24.951249 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9gbz" event={"ID":"376ae4f8-8441-46a8-ab12-07d3f9c729d9","Type":"ContainerStarted","Data":"a5824fbad2e9141759c8ce6c7565783807ce9f28c844443d3d3fe1b52533423e"} Feb 25 06:51:24 crc kubenswrapper[4978]: I0225 06:51:24.952731 4978 generic.go:334] "Generic (PLEG): container finished" podID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerID="c4fe110e6b4c66227e91f464c9a2f21257b98defd9e1b377a4183e36582da88d" exitCode=0 Feb 25 06:51:24 crc kubenswrapper[4978]: I0225 06:51:24.952770 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qw7l" event={"ID":"75fbd0de-dd88-4079-a5c9-72aff9976339","Type":"ContainerDied","Data":"c4fe110e6b4c66227e91f464c9a2f21257b98defd9e1b377a4183e36582da88d"} Feb 25 06:51:24 crc kubenswrapper[4978]: I0225 06:51:24.952794 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qw7l" event={"ID":"75fbd0de-dd88-4079-a5c9-72aff9976339","Type":"ContainerStarted","Data":"7248abf77ec7dfb8cbdfad57f49f813fe18f666ce7003c67ab00beb7cebb16b2"} Feb 25 06:51:24 crc kubenswrapper[4978]: I0225 06:51:24.974194 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p9gbz" podStartSLOduration=3.238740642 podStartE2EDuration="5.974173204s" podCreationTimestamp="2026-02-25 06:51:19 +0000 UTC" firstStartedPulling="2026-02-25 06:51:20.90625485 +0000 UTC m=+374.345511329" lastFinishedPulling="2026-02-25 06:51:23.641687432 +0000 UTC m=+377.080943891" observedRunningTime="2026-02-25 06:51:24.972268363 +0000 UTC m=+378.411524852" watchObservedRunningTime="2026-02-25 06:51:24.974173204 +0000 UTC m=+378.413429673" Feb 25 06:51:25 crc kubenswrapper[4978]: I0225 06:51:25.959441 4978 generic.go:334] "Generic (PLEG): container finished" podID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerID="0f5bccda1e84ea1f67c9b0c2e1a920ebe0c9151d2e7c2daf427dacad83d535c2" exitCode=0 Feb 25 06:51:25 crc kubenswrapper[4978]: I0225 06:51:25.959744 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krctw" event={"ID":"8520c8a9-3af3-465a-8c1d-aa9d10c36c40","Type":"ContainerDied","Data":"0f5bccda1e84ea1f67c9b0c2e1a920ebe0c9151d2e7c2daf427dacad83d535c2"} Feb 25 06:51:25 crc kubenswrapper[4978]: I0225 06:51:25.964633 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgg7r" event={"ID":"e4a76118-7308-4b0f-aae5-41cd6888a845","Type":"ContainerStarted","Data":"5cbfd50bb21d245634688454de6092339a95aed5c8edfda3ec1af5e802077baf"} Feb 25 06:51:25 crc kubenswrapper[4978]: I0225 06:51:25.966405 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qw7l" event={"ID":"75fbd0de-dd88-4079-a5c9-72aff9976339","Type":"ContainerStarted","Data":"99ea167ab72e2e78b8cfd016f128caf8117a21c93283c67947ca6f4c74868140"} Feb 25 06:51:26 crc kubenswrapper[4978]: I0225 06:51:26.006731 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dgg7r" podStartSLOduration=2.960461041 podStartE2EDuration="6.006715366s" podCreationTimestamp="2026-02-25 06:51:20 +0000 UTC" firstStartedPulling="2026-02-25 06:51:21.912495251 +0000 UTC m=+375.351751710" lastFinishedPulling="2026-02-25 06:51:24.958749536 +0000 UTC m=+378.398006035" observedRunningTime="2026-02-25 06:51:26.002726398 +0000 UTC m=+379.441982877" watchObservedRunningTime="2026-02-25 06:51:26.006715366 +0000 UTC m=+379.445971825" Feb 25 06:51:26 crc kubenswrapper[4978]: I0225 06:51:26.973564 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krctw" event={"ID":"8520c8a9-3af3-465a-8c1d-aa9d10c36c40","Type":"ContainerStarted","Data":"b177e600244116e90d5fab43280a48653d1fb3d11c36074d427767becec43ed2"} Feb 25 06:51:26 crc kubenswrapper[4978]: I0225 06:51:26.976553 4978 generic.go:334] "Generic (PLEG): container finished" podID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerID="99ea167ab72e2e78b8cfd016f128caf8117a21c93283c67947ca6f4c74868140" exitCode=0 Feb 25 06:51:26 crc kubenswrapper[4978]: I0225 06:51:26.976602 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qw7l" event={"ID":"75fbd0de-dd88-4079-a5c9-72aff9976339","Type":"ContainerDied","Data":"99ea167ab72e2e78b8cfd016f128caf8117a21c93283c67947ca6f4c74868140"} Feb 25 06:51:27 crc kubenswrapper[4978]: I0225 06:51:27.026338 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-krctw" podStartSLOduration=2.528593734 podStartE2EDuration="6.026316739s" podCreationTimestamp="2026-02-25 06:51:21 +0000 UTC" firstStartedPulling="2026-02-25 06:51:22.930913886 +0000 UTC m=+376.370170345" lastFinishedPulling="2026-02-25 06:51:26.428636871 +0000 UTC m=+379.867893350" observedRunningTime="2026-02-25 06:51:27.003822842 +0000 UTC m=+380.443079341" watchObservedRunningTime="2026-02-25 06:51:27.026316739 +0000 UTC m=+380.465573208" Feb 25 06:51:27 crc kubenswrapper[4978]: I0225 06:51:27.984177 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qw7l" event={"ID":"75fbd0de-dd88-4079-a5c9-72aff9976339","Type":"ContainerStarted","Data":"f7d98c0a14d86a3e620f5bc3b576d8cecca71fb39836d9b77252c4c09e6ca202"} Feb 25 06:51:28 crc kubenswrapper[4978]: I0225 06:51:28.004981 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9qw7l" podStartSLOduration=2.514251581 podStartE2EDuration="5.004965019s" podCreationTimestamp="2026-02-25 06:51:23 +0000 UTC" firstStartedPulling="2026-02-25 06:51:24.956663938 +0000 UTC m=+378.395920397" lastFinishedPulling="2026-02-25 06:51:27.447377366 +0000 UTC m=+380.886633835" observedRunningTime="2026-02-25 06:51:28.002740706 +0000 UTC m=+381.441997205" watchObservedRunningTime="2026-02-25 06:51:28.004965019 +0000 UTC m=+381.444221478" Feb 25 06:51:30 crc kubenswrapper[4978]: I0225 06:51:30.075203 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:30 crc kubenswrapper[4978]: I0225 06:51:30.075260 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:30 crc kubenswrapper[4978]: I0225 06:51:30.140075 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:31 crc kubenswrapper[4978]: I0225 06:51:31.058009 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 06:51:31 crc kubenswrapper[4978]: I0225 06:51:31.173342 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:31 crc kubenswrapper[4978]: I0225 06:51:31.173439 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:31 crc kubenswrapper[4978]: I0225 06:51:31.210998 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:32 crc kubenswrapper[4978]: I0225 06:51:32.047598 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 06:51:32 crc kubenswrapper[4978]: I0225 06:51:32.181149 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:32 crc kubenswrapper[4978]: I0225 06:51:32.181217 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:33 crc kubenswrapper[4978]: I0225 06:51:33.228066 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-krctw" podUID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerName="registry-server" probeResult="failure" output=< Feb 25 06:51:33 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 06:51:33 crc kubenswrapper[4978]: > Feb 25 06:51:33 crc kubenswrapper[4978]: I0225 06:51:33.620208 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:33 crc kubenswrapper[4978]: I0225 06:51:33.620277 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:33 crc kubenswrapper[4978]: I0225 06:51:33.678835 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:34 crc kubenswrapper[4978]: I0225 06:51:34.065183 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9qw7l" Feb 25 06:51:42 crc kubenswrapper[4978]: I0225 06:51:42.240325 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:42 crc kubenswrapper[4978]: I0225 06:51:42.293726 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-krctw" Feb 25 06:51:46 crc kubenswrapper[4978]: I0225 06:51:46.541004 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:51:46 crc kubenswrapper[4978]: I0225 06:51:46.541648 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:51:47 crc kubenswrapper[4978]: I0225 06:51:47.974240 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" podUID="8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" containerName="registry" containerID="cri-o://496811eccf93eaa4d501eb231825144e91b438ae0f66a11b53f747acee76bb2a" gracePeriod=30 Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.111229 4978 generic.go:334] "Generic (PLEG): container finished" podID="8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" containerID="496811eccf93eaa4d501eb231825144e91b438ae0f66a11b53f747acee76bb2a" exitCode=0 Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.111402 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" event={"ID":"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b","Type":"ContainerDied","Data":"496811eccf93eaa4d501eb231825144e91b438ae0f66a11b53f747acee76bb2a"} Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.479657 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.589516 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-trusted-ca\") pod \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.590795 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.590998 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-certificates\") pod \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.591966 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.592075 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-ca-trust-extracted\") pod \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.592182 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-installation-pull-secrets\") pod \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.592337 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.592453 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-bound-sa-token\") pod \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.592492 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df24m\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-kube-api-access-df24m\") pod \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.592536 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-tls\") pod \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\" (UID: \"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b\") " Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.592917 4978 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.592946 4978 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.603148 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-kube-api-access-df24m" (OuterVolumeSpecName: "kube-api-access-df24m") pod "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b"). InnerVolumeSpecName "kube-api-access-df24m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.603482 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.604106 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.607891 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.623816 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.625720 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" (UID: "8cbd051d-53bc-4bd1-9c06-3c71859e7a7b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.693908 4978 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.693980 4978 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.694009 4978 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.694034 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df24m\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-kube-api-access-df24m\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:48 crc kubenswrapper[4978]: I0225 06:51:48.694059 4978 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 25 06:51:49 crc kubenswrapper[4978]: I0225 06:51:49.121351 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" event={"ID":"8cbd051d-53bc-4bd1-9c06-3c71859e7a7b","Type":"ContainerDied","Data":"a98d99ded2c7c406b81f98a34c22fd475387a93da080fed28f540903efc001e0"} Feb 25 06:51:49 crc kubenswrapper[4978]: I0225 06:51:49.122720 4978 scope.go:117] "RemoveContainer" containerID="496811eccf93eaa4d501eb231825144e91b438ae0f66a11b53f747acee76bb2a" Feb 25 06:51:49 crc kubenswrapper[4978]: I0225 06:51:49.121494 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-w6cv7" Feb 25 06:51:49 crc kubenswrapper[4978]: I0225 06:51:49.163433 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w6cv7"] Feb 25 06:51:49 crc kubenswrapper[4978]: I0225 06:51:49.167589 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-w6cv7"] Feb 25 06:51:49 crc kubenswrapper[4978]: I0225 06:51:49.339881 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" path="/var/lib/kubelet/pods/8cbd051d-53bc-4bd1-9c06-3c71859e7a7b/volumes" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.145748 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533372-rftts"] Feb 25 06:52:00 crc kubenswrapper[4978]: E0225 06:52:00.147041 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" containerName="registry" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.147072 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" containerName="registry" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.147315 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbd051d-53bc-4bd1-9c06-3c71859e7a7b" containerName="registry" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.148106 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533372-rftts" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.150432 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.150755 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.151196 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.154115 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533372-rftts"] Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.251669 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jttfd\" (UniqueName: \"kubernetes.io/projected/0b8e1547-6a49-400f-b8e1-b64a35c33a9a-kube-api-access-jttfd\") pod \"auto-csr-approver-29533372-rftts\" (UID: \"0b8e1547-6a49-400f-b8e1-b64a35c33a9a\") " pod="openshift-infra/auto-csr-approver-29533372-rftts" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.353128 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jttfd\" (UniqueName: \"kubernetes.io/projected/0b8e1547-6a49-400f-b8e1-b64a35c33a9a-kube-api-access-jttfd\") pod \"auto-csr-approver-29533372-rftts\" (UID: \"0b8e1547-6a49-400f-b8e1-b64a35c33a9a\") " pod="openshift-infra/auto-csr-approver-29533372-rftts" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.386885 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jttfd\" (UniqueName: \"kubernetes.io/projected/0b8e1547-6a49-400f-b8e1-b64a35c33a9a-kube-api-access-jttfd\") pod \"auto-csr-approver-29533372-rftts\" (UID: \"0b8e1547-6a49-400f-b8e1-b64a35c33a9a\") " pod="openshift-infra/auto-csr-approver-29533372-rftts" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.474000 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533372-rftts" Feb 25 06:52:00 crc kubenswrapper[4978]: I0225 06:52:00.708961 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533372-rftts"] Feb 25 06:52:01 crc kubenswrapper[4978]: I0225 06:52:01.209713 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533372-rftts" event={"ID":"0b8e1547-6a49-400f-b8e1-b64a35c33a9a","Type":"ContainerStarted","Data":"08320cde1b67c39bd05014507948749e3cceeec66ef869f291fb8563b9cbb2b0"} Feb 25 06:52:02 crc kubenswrapper[4978]: I0225 06:52:02.218468 4978 generic.go:334] "Generic (PLEG): container finished" podID="0b8e1547-6a49-400f-b8e1-b64a35c33a9a" containerID="7f20e32af178d49752be54c12becf8e0217975964cf4eb1a98f4182765e14e83" exitCode=0 Feb 25 06:52:02 crc kubenswrapper[4978]: I0225 06:52:02.218513 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533372-rftts" event={"ID":"0b8e1547-6a49-400f-b8e1-b64a35c33a9a","Type":"ContainerDied","Data":"7f20e32af178d49752be54c12becf8e0217975964cf4eb1a98f4182765e14e83"} Feb 25 06:52:03 crc kubenswrapper[4978]: I0225 06:52:03.501759 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533372-rftts" Feb 25 06:52:03 crc kubenswrapper[4978]: I0225 06:52:03.595619 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jttfd\" (UniqueName: \"kubernetes.io/projected/0b8e1547-6a49-400f-b8e1-b64a35c33a9a-kube-api-access-jttfd\") pod \"0b8e1547-6a49-400f-b8e1-b64a35c33a9a\" (UID: \"0b8e1547-6a49-400f-b8e1-b64a35c33a9a\") " Feb 25 06:52:03 crc kubenswrapper[4978]: I0225 06:52:03.605251 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b8e1547-6a49-400f-b8e1-b64a35c33a9a-kube-api-access-jttfd" (OuterVolumeSpecName: "kube-api-access-jttfd") pod "0b8e1547-6a49-400f-b8e1-b64a35c33a9a" (UID: "0b8e1547-6a49-400f-b8e1-b64a35c33a9a"). InnerVolumeSpecName "kube-api-access-jttfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:52:03 crc kubenswrapper[4978]: I0225 06:52:03.697641 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jttfd\" (UniqueName: \"kubernetes.io/projected/0b8e1547-6a49-400f-b8e1-b64a35c33a9a-kube-api-access-jttfd\") on node \"crc\" DevicePath \"\"" Feb 25 06:52:04 crc kubenswrapper[4978]: I0225 06:52:04.231933 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533372-rftts" event={"ID":"0b8e1547-6a49-400f-b8e1-b64a35c33a9a","Type":"ContainerDied","Data":"08320cde1b67c39bd05014507948749e3cceeec66ef869f291fb8563b9cbb2b0"} Feb 25 06:52:04 crc kubenswrapper[4978]: I0225 06:52:04.231977 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08320cde1b67c39bd05014507948749e3cceeec66ef869f291fb8563b9cbb2b0" Feb 25 06:52:04 crc kubenswrapper[4978]: I0225 06:52:04.232025 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533372-rftts" Feb 25 06:52:16 crc kubenswrapper[4978]: I0225 06:52:16.540860 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:52:16 crc kubenswrapper[4978]: I0225 06:52:16.541593 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:52:46 crc kubenswrapper[4978]: I0225 06:52:46.541319 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:52:46 crc kubenswrapper[4978]: I0225 06:52:46.542099 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:52:46 crc kubenswrapper[4978]: I0225 06:52:46.542183 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:52:46 crc kubenswrapper[4978]: I0225 06:52:46.543103 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cda6e0eb7c768cf92c9013ff2dbcca2a8bfe25b85cfa848c319625de918b2b8e"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 06:52:46 crc kubenswrapper[4978]: I0225 06:52:46.543197 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://cda6e0eb7c768cf92c9013ff2dbcca2a8bfe25b85cfa848c319625de918b2b8e" gracePeriod=600 Feb 25 06:52:46 crc kubenswrapper[4978]: I0225 06:52:46.787596 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="cda6e0eb7c768cf92c9013ff2dbcca2a8bfe25b85cfa848c319625de918b2b8e" exitCode=0 Feb 25 06:52:46 crc kubenswrapper[4978]: I0225 06:52:46.787770 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"cda6e0eb7c768cf92c9013ff2dbcca2a8bfe25b85cfa848c319625de918b2b8e"} Feb 25 06:52:46 crc kubenswrapper[4978]: I0225 06:52:46.788086 4978 scope.go:117] "RemoveContainer" containerID="1681565645f2fc8aba05522208118ff62097a02395870d10eb9361b066bdb939" Feb 25 06:52:47 crc kubenswrapper[4978]: I0225 06:52:47.797643 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"4f1b5b9f287453a0bb1864000bcfb5ad43747062c90588cbb9fbe231a11317b3"} Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.154527 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533374-fdvtk"] Feb 25 06:54:00 crc kubenswrapper[4978]: E0225 06:54:00.158975 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b8e1547-6a49-400f-b8e1-b64a35c33a9a" containerName="oc" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.159010 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b8e1547-6a49-400f-b8e1-b64a35c33a9a" containerName="oc" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.159274 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b8e1547-6a49-400f-b8e1-b64a35c33a9a" containerName="oc" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.160129 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533374-fdvtk" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.162349 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.163057 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.164331 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.169252 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533374-fdvtk"] Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.324076 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5lmn\" (UniqueName: \"kubernetes.io/projected/e559d93f-7e03-4ee7-850c-7a305a76e3cd-kube-api-access-s5lmn\") pod \"auto-csr-approver-29533374-fdvtk\" (UID: \"e559d93f-7e03-4ee7-850c-7a305a76e3cd\") " pod="openshift-infra/auto-csr-approver-29533374-fdvtk" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.425840 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5lmn\" (UniqueName: \"kubernetes.io/projected/e559d93f-7e03-4ee7-850c-7a305a76e3cd-kube-api-access-s5lmn\") pod \"auto-csr-approver-29533374-fdvtk\" (UID: \"e559d93f-7e03-4ee7-850c-7a305a76e3cd\") " pod="openshift-infra/auto-csr-approver-29533374-fdvtk" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.453662 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5lmn\" (UniqueName: \"kubernetes.io/projected/e559d93f-7e03-4ee7-850c-7a305a76e3cd-kube-api-access-s5lmn\") pod \"auto-csr-approver-29533374-fdvtk\" (UID: \"e559d93f-7e03-4ee7-850c-7a305a76e3cd\") " pod="openshift-infra/auto-csr-approver-29533374-fdvtk" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.495940 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533374-fdvtk" Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.793583 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533374-fdvtk"] Feb 25 06:54:00 crc kubenswrapper[4978]: I0225 06:54:00.810574 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 06:54:01 crc kubenswrapper[4978]: I0225 06:54:01.309606 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533374-fdvtk" event={"ID":"e559d93f-7e03-4ee7-850c-7a305a76e3cd","Type":"ContainerStarted","Data":"05b38442bb27ea818ca6e7799e2e1b98ccf7a6f6f7b15370bb2029e1d9b2eefd"} Feb 25 06:54:02 crc kubenswrapper[4978]: I0225 06:54:02.317489 4978 generic.go:334] "Generic (PLEG): container finished" podID="e559d93f-7e03-4ee7-850c-7a305a76e3cd" containerID="929f4573b7b94c9d2dd4dbe58627b2bd0a3d624d3974ea2369e01a289d5ba5b2" exitCode=0 Feb 25 06:54:02 crc kubenswrapper[4978]: I0225 06:54:02.317538 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533374-fdvtk" event={"ID":"e559d93f-7e03-4ee7-850c-7a305a76e3cd","Type":"ContainerDied","Data":"929f4573b7b94c9d2dd4dbe58627b2bd0a3d624d3974ea2369e01a289d5ba5b2"} Feb 25 06:54:03 crc kubenswrapper[4978]: I0225 06:54:03.601457 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533374-fdvtk" Feb 25 06:54:03 crc kubenswrapper[4978]: I0225 06:54:03.775232 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5lmn\" (UniqueName: \"kubernetes.io/projected/e559d93f-7e03-4ee7-850c-7a305a76e3cd-kube-api-access-s5lmn\") pod \"e559d93f-7e03-4ee7-850c-7a305a76e3cd\" (UID: \"e559d93f-7e03-4ee7-850c-7a305a76e3cd\") " Feb 25 06:54:03 crc kubenswrapper[4978]: I0225 06:54:03.783966 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e559d93f-7e03-4ee7-850c-7a305a76e3cd-kube-api-access-s5lmn" (OuterVolumeSpecName: "kube-api-access-s5lmn") pod "e559d93f-7e03-4ee7-850c-7a305a76e3cd" (UID: "e559d93f-7e03-4ee7-850c-7a305a76e3cd"). InnerVolumeSpecName "kube-api-access-s5lmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:54:03 crc kubenswrapper[4978]: I0225 06:54:03.877160 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5lmn\" (UniqueName: \"kubernetes.io/projected/e559d93f-7e03-4ee7-850c-7a305a76e3cd-kube-api-access-s5lmn\") on node \"crc\" DevicePath \"\"" Feb 25 06:54:04 crc kubenswrapper[4978]: I0225 06:54:04.333463 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533374-fdvtk" event={"ID":"e559d93f-7e03-4ee7-850c-7a305a76e3cd","Type":"ContainerDied","Data":"05b38442bb27ea818ca6e7799e2e1b98ccf7a6f6f7b15370bb2029e1d9b2eefd"} Feb 25 06:54:04 crc kubenswrapper[4978]: I0225 06:54:04.333869 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05b38442bb27ea818ca6e7799e2e1b98ccf7a6f6f7b15370bb2029e1d9b2eefd" Feb 25 06:54:04 crc kubenswrapper[4978]: I0225 06:54:04.333564 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533374-fdvtk" Feb 25 06:54:04 crc kubenswrapper[4978]: I0225 06:54:04.667886 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533368-gn4dn"] Feb 25 06:54:04 crc kubenswrapper[4978]: I0225 06:54:04.670998 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533368-gn4dn"] Feb 25 06:54:05 crc kubenswrapper[4978]: I0225 06:54:05.337225 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dec7193e-a91c-45b7-a8b1-236d626f0540" path="/var/lib/kubelet/pods/dec7193e-a91c-45b7-a8b1-236d626f0540/volumes" Feb 25 06:54:46 crc kubenswrapper[4978]: I0225 06:54:46.540809 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:54:46 crc kubenswrapper[4978]: I0225 06:54:46.542565 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:55:14 crc kubenswrapper[4978]: I0225 06:55:14.314000 4978 scope.go:117] "RemoveContainer" containerID="b4d31aab45082a45f8ae53c881d441c93ca3611ac7427910de1c0f13ced0d0d1" Feb 25 06:55:14 crc kubenswrapper[4978]: I0225 06:55:14.371217 4978 scope.go:117] "RemoveContainer" containerID="81fcefb665a61468a87a293b4733f7a494926b838956fa87fc95b6892b281f99" Feb 25 06:55:16 crc kubenswrapper[4978]: I0225 06:55:16.540756 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:55:16 crc kubenswrapper[4978]: I0225 06:55:16.540836 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:55:46 crc kubenswrapper[4978]: I0225 06:55:46.540850 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:55:46 crc kubenswrapper[4978]: I0225 06:55:46.542553 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:55:46 crc kubenswrapper[4978]: I0225 06:55:46.542651 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:55:46 crc kubenswrapper[4978]: I0225 06:55:46.543649 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4f1b5b9f287453a0bb1864000bcfb5ad43747062c90588cbb9fbe231a11317b3"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 06:55:46 crc kubenswrapper[4978]: I0225 06:55:46.543785 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://4f1b5b9f287453a0bb1864000bcfb5ad43747062c90588cbb9fbe231a11317b3" gracePeriod=600 Feb 25 06:55:47 crc kubenswrapper[4978]: I0225 06:55:47.098187 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="4f1b5b9f287453a0bb1864000bcfb5ad43747062c90588cbb9fbe231a11317b3" exitCode=0 Feb 25 06:55:47 crc kubenswrapper[4978]: I0225 06:55:47.098300 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"4f1b5b9f287453a0bb1864000bcfb5ad43747062c90588cbb9fbe231a11317b3"} Feb 25 06:55:47 crc kubenswrapper[4978]: I0225 06:55:47.098630 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"d6649b854ec6cf99def7e6418068fe617fbe3b2759595971796ed8d3fffb9745"} Feb 25 06:55:47 crc kubenswrapper[4978]: I0225 06:55:47.098665 4978 scope.go:117] "RemoveContainer" containerID="cda6e0eb7c768cf92c9013ff2dbcca2a8bfe25b85cfa848c319625de918b2b8e" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.157650 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533376-xrqml"] Feb 25 06:56:00 crc kubenswrapper[4978]: E0225 06:56:00.164144 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e559d93f-7e03-4ee7-850c-7a305a76e3cd" containerName="oc" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.164493 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e559d93f-7e03-4ee7-850c-7a305a76e3cd" containerName="oc" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.164841 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e559d93f-7e03-4ee7-850c-7a305a76e3cd" containerName="oc" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.165802 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533376-xrqml" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.168900 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533376-xrqml"] Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.169947 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.170319 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.170331 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.249709 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl5nn\" (UniqueName: \"kubernetes.io/projected/7dc77d60-3127-4d1d-acfa-87d1a8822d8a-kube-api-access-fl5nn\") pod \"auto-csr-approver-29533376-xrqml\" (UID: \"7dc77d60-3127-4d1d-acfa-87d1a8822d8a\") " pod="openshift-infra/auto-csr-approver-29533376-xrqml" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.350807 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl5nn\" (UniqueName: \"kubernetes.io/projected/7dc77d60-3127-4d1d-acfa-87d1a8822d8a-kube-api-access-fl5nn\") pod \"auto-csr-approver-29533376-xrqml\" (UID: \"7dc77d60-3127-4d1d-acfa-87d1a8822d8a\") " pod="openshift-infra/auto-csr-approver-29533376-xrqml" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.385345 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl5nn\" (UniqueName: \"kubernetes.io/projected/7dc77d60-3127-4d1d-acfa-87d1a8822d8a-kube-api-access-fl5nn\") pod \"auto-csr-approver-29533376-xrqml\" (UID: \"7dc77d60-3127-4d1d-acfa-87d1a8822d8a\") " pod="openshift-infra/auto-csr-approver-29533376-xrqml" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.509426 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533376-xrqml" Feb 25 06:56:00 crc kubenswrapper[4978]: I0225 06:56:00.813593 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533376-xrqml"] Feb 25 06:56:01 crc kubenswrapper[4978]: I0225 06:56:01.184918 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533376-xrqml" event={"ID":"7dc77d60-3127-4d1d-acfa-87d1a8822d8a","Type":"ContainerStarted","Data":"fa2808fcebb8612aa221a48f06b7fe25ba12174722cf926dc5851a3512d81632"} Feb 25 06:56:02 crc kubenswrapper[4978]: I0225 06:56:02.201742 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533376-xrqml" event={"ID":"7dc77d60-3127-4d1d-acfa-87d1a8822d8a","Type":"ContainerStarted","Data":"767d35074d443f0c5027d813c91e7792fdc019d8ca89a519280927f80bb82cf8"} Feb 25 06:56:02 crc kubenswrapper[4978]: I0225 06:56:02.230657 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533376-xrqml" podStartSLOduration=1.445185639 podStartE2EDuration="2.230631509s" podCreationTimestamp="2026-02-25 06:56:00 +0000 UTC" firstStartedPulling="2026-02-25 06:56:00.825993693 +0000 UTC m=+654.265250162" lastFinishedPulling="2026-02-25 06:56:01.611439563 +0000 UTC m=+655.050696032" observedRunningTime="2026-02-25 06:56:02.229772832 +0000 UTC m=+655.669029321" watchObservedRunningTime="2026-02-25 06:56:02.230631509 +0000 UTC m=+655.669888048" Feb 25 06:56:03 crc kubenswrapper[4978]: I0225 06:56:03.211039 4978 generic.go:334] "Generic (PLEG): container finished" podID="7dc77d60-3127-4d1d-acfa-87d1a8822d8a" containerID="767d35074d443f0c5027d813c91e7792fdc019d8ca89a519280927f80bb82cf8" exitCode=0 Feb 25 06:56:03 crc kubenswrapper[4978]: I0225 06:56:03.211118 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533376-xrqml" event={"ID":"7dc77d60-3127-4d1d-acfa-87d1a8822d8a","Type":"ContainerDied","Data":"767d35074d443f0c5027d813c91e7792fdc019d8ca89a519280927f80bb82cf8"} Feb 25 06:56:04 crc kubenswrapper[4978]: I0225 06:56:04.536776 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533376-xrqml" Feb 25 06:56:04 crc kubenswrapper[4978]: I0225 06:56:04.620400 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl5nn\" (UniqueName: \"kubernetes.io/projected/7dc77d60-3127-4d1d-acfa-87d1a8822d8a-kube-api-access-fl5nn\") pod \"7dc77d60-3127-4d1d-acfa-87d1a8822d8a\" (UID: \"7dc77d60-3127-4d1d-acfa-87d1a8822d8a\") " Feb 25 06:56:04 crc kubenswrapper[4978]: I0225 06:56:04.629601 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dc77d60-3127-4d1d-acfa-87d1a8822d8a-kube-api-access-fl5nn" (OuterVolumeSpecName: "kube-api-access-fl5nn") pod "7dc77d60-3127-4d1d-acfa-87d1a8822d8a" (UID: "7dc77d60-3127-4d1d-acfa-87d1a8822d8a"). InnerVolumeSpecName "kube-api-access-fl5nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:56:04 crc kubenswrapper[4978]: I0225 06:56:04.722810 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl5nn\" (UniqueName: \"kubernetes.io/projected/7dc77d60-3127-4d1d-acfa-87d1a8822d8a-kube-api-access-fl5nn\") on node \"crc\" DevicePath \"\"" Feb 25 06:56:05 crc kubenswrapper[4978]: I0225 06:56:05.227360 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533376-xrqml" event={"ID":"7dc77d60-3127-4d1d-acfa-87d1a8822d8a","Type":"ContainerDied","Data":"fa2808fcebb8612aa221a48f06b7fe25ba12174722cf926dc5851a3512d81632"} Feb 25 06:56:05 crc kubenswrapper[4978]: I0225 06:56:05.227800 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa2808fcebb8612aa221a48f06b7fe25ba12174722cf926dc5851a3512d81632" Feb 25 06:56:05 crc kubenswrapper[4978]: I0225 06:56:05.227521 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533376-xrqml" Feb 25 06:56:05 crc kubenswrapper[4978]: I0225 06:56:05.308323 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533370-4jjws"] Feb 25 06:56:05 crc kubenswrapper[4978]: I0225 06:56:05.314987 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533370-4jjws"] Feb 25 06:56:05 crc kubenswrapper[4978]: I0225 06:56:05.340283 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed03df44-2b1e-4633-9743-79c9a9961ff0" path="/var/lib/kubelet/pods/ed03df44-2b1e-4633-9743-79c9a9961ff0/volumes" Feb 25 06:56:14 crc kubenswrapper[4978]: I0225 06:56:14.424210 4978 scope.go:117] "RemoveContainer" containerID="5679b9a49e1fbf84e5cbefc047e8cd2f9f7857495901ecc6e7deb6d4640ceea9" Feb 25 06:57:17 crc kubenswrapper[4978]: I0225 06:57:17.080885 4978 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 25 06:57:46 crc kubenswrapper[4978]: I0225 06:57:46.541054 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:57:46 crc kubenswrapper[4978]: I0225 06:57:46.542681 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.150196 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533378-wmlm2"] Feb 25 06:58:00 crc kubenswrapper[4978]: E0225 06:58:00.151299 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc77d60-3127-4d1d-acfa-87d1a8822d8a" containerName="oc" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.151324 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc77d60-3127-4d1d-acfa-87d1a8822d8a" containerName="oc" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.151575 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc77d60-3127-4d1d-acfa-87d1a8822d8a" containerName="oc" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.152163 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533378-wmlm2" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.159287 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.159301 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.159540 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.162159 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533378-wmlm2"] Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.261862 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdwlm\" (UniqueName: \"kubernetes.io/projected/09a0d1ce-3f33-46a6-98a9-a0a2337e7a33-kube-api-access-wdwlm\") pod \"auto-csr-approver-29533378-wmlm2\" (UID: \"09a0d1ce-3f33-46a6-98a9-a0a2337e7a33\") " pod="openshift-infra/auto-csr-approver-29533378-wmlm2" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.363295 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdwlm\" (UniqueName: \"kubernetes.io/projected/09a0d1ce-3f33-46a6-98a9-a0a2337e7a33-kube-api-access-wdwlm\") pod \"auto-csr-approver-29533378-wmlm2\" (UID: \"09a0d1ce-3f33-46a6-98a9-a0a2337e7a33\") " pod="openshift-infra/auto-csr-approver-29533378-wmlm2" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.399579 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdwlm\" (UniqueName: \"kubernetes.io/projected/09a0d1ce-3f33-46a6-98a9-a0a2337e7a33-kube-api-access-wdwlm\") pod \"auto-csr-approver-29533378-wmlm2\" (UID: \"09a0d1ce-3f33-46a6-98a9-a0a2337e7a33\") " pod="openshift-infra/auto-csr-approver-29533378-wmlm2" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.484398 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533378-wmlm2" Feb 25 06:58:00 crc kubenswrapper[4978]: I0225 06:58:00.965797 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533378-wmlm2"] Feb 25 06:58:01 crc kubenswrapper[4978]: I0225 06:58:01.122267 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533378-wmlm2" event={"ID":"09a0d1ce-3f33-46a6-98a9-a0a2337e7a33","Type":"ContainerStarted","Data":"c163d02f59eb41c938426571447dfcbbc4ce459741b75d57524f442b913a4501"} Feb 25 06:58:03 crc kubenswrapper[4978]: I0225 06:58:03.141249 4978 generic.go:334] "Generic (PLEG): container finished" podID="09a0d1ce-3f33-46a6-98a9-a0a2337e7a33" containerID="d5c59d5bffe4a5b678647ead046fd39dc8f8128c1eb6627de9e03bd857a35439" exitCode=0 Feb 25 06:58:03 crc kubenswrapper[4978]: I0225 06:58:03.141328 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533378-wmlm2" event={"ID":"09a0d1ce-3f33-46a6-98a9-a0a2337e7a33","Type":"ContainerDied","Data":"d5c59d5bffe4a5b678647ead046fd39dc8f8128c1eb6627de9e03bd857a35439"} Feb 25 06:58:04 crc kubenswrapper[4978]: I0225 06:58:04.460852 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533378-wmlm2" Feb 25 06:58:04 crc kubenswrapper[4978]: I0225 06:58:04.522631 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdwlm\" (UniqueName: \"kubernetes.io/projected/09a0d1ce-3f33-46a6-98a9-a0a2337e7a33-kube-api-access-wdwlm\") pod \"09a0d1ce-3f33-46a6-98a9-a0a2337e7a33\" (UID: \"09a0d1ce-3f33-46a6-98a9-a0a2337e7a33\") " Feb 25 06:58:04 crc kubenswrapper[4978]: I0225 06:58:04.531341 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09a0d1ce-3f33-46a6-98a9-a0a2337e7a33-kube-api-access-wdwlm" (OuterVolumeSpecName: "kube-api-access-wdwlm") pod "09a0d1ce-3f33-46a6-98a9-a0a2337e7a33" (UID: "09a0d1ce-3f33-46a6-98a9-a0a2337e7a33"). InnerVolumeSpecName "kube-api-access-wdwlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:58:04 crc kubenswrapper[4978]: I0225 06:58:04.624319 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdwlm\" (UniqueName: \"kubernetes.io/projected/09a0d1ce-3f33-46a6-98a9-a0a2337e7a33-kube-api-access-wdwlm\") on node \"crc\" DevicePath \"\"" Feb 25 06:58:05 crc kubenswrapper[4978]: I0225 06:58:05.159262 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533378-wmlm2" event={"ID":"09a0d1ce-3f33-46a6-98a9-a0a2337e7a33","Type":"ContainerDied","Data":"c163d02f59eb41c938426571447dfcbbc4ce459741b75d57524f442b913a4501"} Feb 25 06:58:05 crc kubenswrapper[4978]: I0225 06:58:05.159331 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c163d02f59eb41c938426571447dfcbbc4ce459741b75d57524f442b913a4501" Feb 25 06:58:05 crc kubenswrapper[4978]: I0225 06:58:05.159358 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533378-wmlm2" Feb 25 06:58:05 crc kubenswrapper[4978]: I0225 06:58:05.547665 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533372-rftts"] Feb 25 06:58:05 crc kubenswrapper[4978]: I0225 06:58:05.555261 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533372-rftts"] Feb 25 06:58:07 crc kubenswrapper[4978]: I0225 06:58:07.350665 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b8e1547-6a49-400f-b8e1-b64a35c33a9a" path="/var/lib/kubelet/pods/0b8e1547-6a49-400f-b8e1-b64a35c33a9a/volumes" Feb 25 06:58:14 crc kubenswrapper[4978]: I0225 06:58:14.514406 4978 scope.go:117] "RemoveContainer" containerID="7f20e32af178d49752be54c12becf8e0217975964cf4eb1a98f4182765e14e83" Feb 25 06:58:16 crc kubenswrapper[4978]: I0225 06:58:16.541273 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:58:16 crc kubenswrapper[4978]: I0225 06:58:16.541435 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:58:46 crc kubenswrapper[4978]: I0225 06:58:46.540819 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 06:58:46 crc kubenswrapper[4978]: I0225 06:58:46.541756 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 06:58:46 crc kubenswrapper[4978]: I0225 06:58:46.541829 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 06:58:46 crc kubenswrapper[4978]: I0225 06:58:46.542599 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d6649b854ec6cf99def7e6418068fe617fbe3b2759595971796ed8d3fffb9745"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 06:58:46 crc kubenswrapper[4978]: I0225 06:58:46.542699 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://d6649b854ec6cf99def7e6418068fe617fbe3b2759595971796ed8d3fffb9745" gracePeriod=600 Feb 25 06:58:47 crc kubenswrapper[4978]: I0225 06:58:47.462914 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="d6649b854ec6cf99def7e6418068fe617fbe3b2759595971796ed8d3fffb9745" exitCode=0 Feb 25 06:58:47 crc kubenswrapper[4978]: I0225 06:58:47.462988 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"d6649b854ec6cf99def7e6418068fe617fbe3b2759595971796ed8d3fffb9745"} Feb 25 06:58:47 crc kubenswrapper[4978]: I0225 06:58:47.463824 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"ea6d2ab6806f2c754878e481d2101f535bcfdd355af23df81d835564bbd06358"} Feb 25 06:58:47 crc kubenswrapper[4978]: I0225 06:58:47.463866 4978 scope.go:117] "RemoveContainer" containerID="4f1b5b9f287453a0bb1864000bcfb5ad43747062c90588cbb9fbe231a11317b3" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.436563 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2gb8f"] Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.437799 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovn-controller" containerID="cri-o://166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671" gracePeriod=30 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.437966 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="kube-rbac-proxy-node" containerID="cri-o://585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de" gracePeriod=30 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.437933 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171" gracePeriod=30 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.438060 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovn-acl-logging" containerID="cri-o://81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124" gracePeriod=30 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.437933 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="nbdb" containerID="cri-o://078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf" gracePeriod=30 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.438176 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="northd" containerID="cri-o://2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3" gracePeriod=30 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.438437 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="sbdb" containerID="cri-o://7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22" gracePeriod=30 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.510564 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovnkube-controller" containerID="cri-o://9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768" gracePeriod=30 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.613463 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lrt6k_b3724722-3928-4d0b-8ef3-f81f91753a30/kube-multus/0.log" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.613861 4978 generic.go:334] "Generic (PLEG): container finished" podID="b3724722-3928-4d0b-8ef3-f81f91753a30" containerID="091bbfdc48a42fe53c5a7cb75c83171476fe2a6a650e2eaa350688796dac097b" exitCode=2 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.613918 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lrt6k" event={"ID":"b3724722-3928-4d0b-8ef3-f81f91753a30","Type":"ContainerDied","Data":"091bbfdc48a42fe53c5a7cb75c83171476fe2a6a650e2eaa350688796dac097b"} Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.614578 4978 scope.go:117] "RemoveContainer" containerID="091bbfdc48a42fe53c5a7cb75c83171476fe2a6a650e2eaa350688796dac097b" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.624860 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gb8f_b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5/ovn-acl-logging/0.log" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.625430 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gb8f_b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5/ovn-controller/0.log" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.625796 4978 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerID="ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171" exitCode=0 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.625824 4978 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerID="585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de" exitCode=0 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.625834 4978 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerID="81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124" exitCode=143 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.625846 4978 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerID="166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671" exitCode=143 Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.625884 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerDied","Data":"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171"} Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.626172 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerDied","Data":"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de"} Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.626228 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerDied","Data":"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124"} Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.626241 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerDied","Data":"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671"} Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.848863 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gb8f_b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5/ovn-acl-logging/0.log" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.849849 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gb8f_b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5/ovn-controller/0.log" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.850323 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.910669 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-sn6n7"] Feb 25 06:59:04 crc kubenswrapper[4978]: E0225 06:59:04.910919 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="kube-rbac-proxy-ovn-metrics" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.910942 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="kube-rbac-proxy-ovn-metrics" Feb 25 06:59:04 crc kubenswrapper[4978]: E0225 06:59:04.910955 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovn-acl-logging" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.910963 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovn-acl-logging" Feb 25 06:59:04 crc kubenswrapper[4978]: E0225 06:59:04.910978 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="nbdb" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.910986 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="nbdb" Feb 25 06:59:04 crc kubenswrapper[4978]: E0225 06:59:04.910998 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="northd" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911005 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="northd" Feb 25 06:59:04 crc kubenswrapper[4978]: E0225 06:59:04.911014 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a0d1ce-3f33-46a6-98a9-a0a2337e7a33" containerName="oc" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911021 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a0d1ce-3f33-46a6-98a9-a0a2337e7a33" containerName="oc" Feb 25 06:59:04 crc kubenswrapper[4978]: E0225 06:59:04.911031 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="kube-rbac-proxy-node" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911038 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="kube-rbac-proxy-node" Feb 25 06:59:04 crc kubenswrapper[4978]: E0225 06:59:04.911046 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="sbdb" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911053 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="sbdb" Feb 25 06:59:04 crc kubenswrapper[4978]: E0225 06:59:04.911061 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovnkube-controller" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911068 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovnkube-controller" Feb 25 06:59:04 crc kubenswrapper[4978]: E0225 06:59:04.911077 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="kubecfg-setup" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911084 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="kubecfg-setup" Feb 25 06:59:04 crc kubenswrapper[4978]: E0225 06:59:04.911100 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovn-controller" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911107 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovn-controller" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911223 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovn-acl-logging" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911234 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="kube-rbac-proxy-node" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911243 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="northd" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911254 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="kube-rbac-proxy-ovn-metrics" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911263 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="nbdb" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911273 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovn-controller" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911282 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="sbdb" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911291 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerName="ovnkube-controller" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.911303 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="09a0d1ce-3f33-46a6-98a9-a0a2337e7a33" containerName="oc" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.913173 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917019 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-var-lib-openvswitch\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917056 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/963d73c4-5aa9-481c-a32e-80fd85bf6093-ovn-node-metrics-cert\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917082 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/963d73c4-5aa9-481c-a32e-80fd85bf6093-ovnkube-script-lib\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917216 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-run-netns\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917286 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-systemd-units\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917333 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917412 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-cni-bin\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917466 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-cni-netd\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917530 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-run-ovn-kubernetes\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917569 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-run-ovn\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917595 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-kubelet\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917612 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-run-systemd\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917630 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-etc-openvswitch\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917759 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/963d73c4-5aa9-481c-a32e-80fd85bf6093-env-overrides\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917918 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-run-openvswitch\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.917979 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-node-log\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.918017 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2b47\" (UniqueName: \"kubernetes.io/projected/963d73c4-5aa9-481c-a32e-80fd85bf6093-kube-api-access-j2b47\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.918050 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/963d73c4-5aa9-481c-a32e-80fd85bf6093-ovnkube-config\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.918137 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-log-socket\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:04 crc kubenswrapper[4978]: I0225 06:59:04.918172 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-slash\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.023941 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-log-socket\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024013 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024025 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-log-socket" (OuterVolumeSpecName: "log-socket") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024046 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-systemd\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024130 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-ovn-kubernetes\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024233 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-openvswitch\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024282 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-slash\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024261 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024326 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-config\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024339 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024353 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024394 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-script-lib\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024408 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-slash" (OuterVolumeSpecName: "host-slash") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024426 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-ovn\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024464 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-netns\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024492 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-node-log\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024534 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdcj7\" (UniqueName: \"kubernetes.io/projected/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-kube-api-access-zdcj7\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024564 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-systemd-units\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024595 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovn-node-metrics-cert\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024626 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-kubelet\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024658 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-env-overrides\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024692 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-etc-openvswitch\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024719 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-var-lib-openvswitch\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024765 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-netd\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024807 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-bin\") pod \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\" (UID: \"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5\") " Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024948 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-slash\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024996 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-var-lib-openvswitch\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025034 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/963d73c4-5aa9-481c-a32e-80fd85bf6093-ovn-node-metrics-cert\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025064 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/963d73c4-5aa9-481c-a32e-80fd85bf6093-ovnkube-script-lib\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025108 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-run-netns\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025138 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025171 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-systemd-units\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025212 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-cni-bin\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025243 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-cni-netd\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025256 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-slash\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024764 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024775 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024792 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024804 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024811 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-node-log" (OuterVolumeSpecName: "node-log") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024827 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025322 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-run-ovn-kubernetes\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.024853 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025165 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025353 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025192 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025209 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025427 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-cni-bin\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025421 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-run-netns\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025471 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-cni-netd\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025398 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-systemd-units\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025279 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-run-ovn-kubernetes\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025439 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025476 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-var-lib-openvswitch\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025609 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-run-ovn\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025661 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-kubelet\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025695 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-run-systemd\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025726 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-etc-openvswitch\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025768 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/963d73c4-5aa9-481c-a32e-80fd85bf6093-env-overrides\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025809 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-run-openvswitch\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025843 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-node-log\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025848 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-host-kubelet\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025766 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-run-ovn\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025881 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2b47\" (UniqueName: \"kubernetes.io/projected/963d73c4-5aa9-481c-a32e-80fd85bf6093-kube-api-access-j2b47\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025945 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/963d73c4-5aa9-481c-a32e-80fd85bf6093-ovnkube-config\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026003 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-log-socket\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026118 4978 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-log-socket\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026139 4978 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026165 4978 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026185 4978 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026208 4978 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026227 4978 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-slash\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026244 4978 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026260 4978 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026276 4978 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026295 4978 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-node-log\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026312 4978 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026329 4978 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026322 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-run-openvswitch\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026347 4978 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026446 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-etc-openvswitch\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026451 4978 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026488 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-node-log\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.025809 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-run-systemd\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026493 4978 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026546 4978 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026411 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/963d73c4-5aa9-481c-a32e-80fd85bf6093-log-socket\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026677 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.026799 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/963d73c4-5aa9-481c-a32e-80fd85bf6093-ovnkube-script-lib\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.027358 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/963d73c4-5aa9-481c-a32e-80fd85bf6093-env-overrides\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.027648 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/963d73c4-5aa9-481c-a32e-80fd85bf6093-ovnkube-config\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.031707 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/963d73c4-5aa9-481c-a32e-80fd85bf6093-ovn-node-metrics-cert\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.032529 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-kube-api-access-zdcj7" (OuterVolumeSpecName: "kube-api-access-zdcj7") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "kube-api-access-zdcj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.034463 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.046206 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" (UID: "b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.054479 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2b47\" (UniqueName: \"kubernetes.io/projected/963d73c4-5aa9-481c-a32e-80fd85bf6093-kube-api-access-j2b47\") pod \"ovnkube-node-sn6n7\" (UID: \"963d73c4-5aa9-481c-a32e-80fd85bf6093\") " pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.127355 4978 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.127402 4978 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.127416 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdcj7\" (UniqueName: \"kubernetes.io/projected/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-kube-api-access-zdcj7\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.127430 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.237842 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.635465 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-lrt6k_b3724722-3928-4d0b-8ef3-f81f91753a30/kube-multus/0.log" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.636212 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lrt6k" event={"ID":"b3724722-3928-4d0b-8ef3-f81f91753a30","Type":"ContainerStarted","Data":"df444783075911acce9cc5c821c027ea4971dfb934fe011e09aaf9dc526c1371"} Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.640360 4978 generic.go:334] "Generic (PLEG): container finished" podID="963d73c4-5aa9-481c-a32e-80fd85bf6093" containerID="bdc4722d67d87e311308c6808dd353b772b8446618710b4184dff047ae87a163" exitCode=0 Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.640436 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" event={"ID":"963d73c4-5aa9-481c-a32e-80fd85bf6093","Type":"ContainerDied","Data":"bdc4722d67d87e311308c6808dd353b772b8446618710b4184dff047ae87a163"} Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.640492 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" event={"ID":"963d73c4-5aa9-481c-a32e-80fd85bf6093","Type":"ContainerStarted","Data":"04cc31961fe4ccea854e431b654e4b010dc9c41ad42d6527ddc0d8ef55e7e117"} Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.652027 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gb8f_b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5/ovn-acl-logging/0.log" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.652982 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2gb8f_b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5/ovn-controller/0.log" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654257 4978 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerID="9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768" exitCode=0 Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654291 4978 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerID="7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22" exitCode=0 Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654307 4978 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerID="078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf" exitCode=0 Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654326 4978 generic.go:334] "Generic (PLEG): container finished" podID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" containerID="2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3" exitCode=0 Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654355 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerDied","Data":"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768"} Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654427 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerDied","Data":"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22"} Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654452 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerDied","Data":"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf"} Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654459 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654489 4978 scope.go:117] "RemoveContainer" containerID="9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654472 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerDied","Data":"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3"} Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.654651 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2gb8f" event={"ID":"b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5","Type":"ContainerDied","Data":"6ccf95dfe30f0cf5ba6d44f75b9b4ee67f798431bc25d006a16c8aca57789666"} Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.697434 4978 scope.go:117] "RemoveContainer" containerID="7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.720953 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2gb8f"] Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.725591 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2gb8f"] Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.756562 4978 scope.go:117] "RemoveContainer" containerID="078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.780441 4978 scope.go:117] "RemoveContainer" containerID="2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.813836 4978 scope.go:117] "RemoveContainer" containerID="ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.835967 4978 scope.go:117] "RemoveContainer" containerID="585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.864592 4978 scope.go:117] "RemoveContainer" containerID="81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.879334 4978 scope.go:117] "RemoveContainer" containerID="166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.910259 4978 scope.go:117] "RemoveContainer" containerID="e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.930683 4978 scope.go:117] "RemoveContainer" containerID="9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768" Feb 25 06:59:05 crc kubenswrapper[4978]: E0225 06:59:05.931181 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768\": container with ID starting with 9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768 not found: ID does not exist" containerID="9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.931218 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768"} err="failed to get container status \"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768\": rpc error: code = NotFound desc = could not find container \"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768\": container with ID starting with 9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.931243 4978 scope.go:117] "RemoveContainer" containerID="7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22" Feb 25 06:59:05 crc kubenswrapper[4978]: E0225 06:59:05.931575 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22\": container with ID starting with 7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22 not found: ID does not exist" containerID="7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.931599 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22"} err="failed to get container status \"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22\": rpc error: code = NotFound desc = could not find container \"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22\": container with ID starting with 7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.932394 4978 scope.go:117] "RemoveContainer" containerID="078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf" Feb 25 06:59:05 crc kubenswrapper[4978]: E0225 06:59:05.932919 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf\": container with ID starting with 078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf not found: ID does not exist" containerID="078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.932945 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf"} err="failed to get container status \"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf\": rpc error: code = NotFound desc = could not find container \"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf\": container with ID starting with 078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.932962 4978 scope.go:117] "RemoveContainer" containerID="2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3" Feb 25 06:59:05 crc kubenswrapper[4978]: E0225 06:59:05.933330 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3\": container with ID starting with 2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3 not found: ID does not exist" containerID="2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.933354 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3"} err="failed to get container status \"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3\": rpc error: code = NotFound desc = could not find container \"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3\": container with ID starting with 2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.933397 4978 scope.go:117] "RemoveContainer" containerID="ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171" Feb 25 06:59:05 crc kubenswrapper[4978]: E0225 06:59:05.934026 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171\": container with ID starting with ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171 not found: ID does not exist" containerID="ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.934053 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171"} err="failed to get container status \"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171\": rpc error: code = NotFound desc = could not find container \"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171\": container with ID starting with ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.934264 4978 scope.go:117] "RemoveContainer" containerID="585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de" Feb 25 06:59:05 crc kubenswrapper[4978]: E0225 06:59:05.935532 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de\": container with ID starting with 585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de not found: ID does not exist" containerID="585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.935571 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de"} err="failed to get container status \"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de\": rpc error: code = NotFound desc = could not find container \"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de\": container with ID starting with 585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.935588 4978 scope.go:117] "RemoveContainer" containerID="81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124" Feb 25 06:59:05 crc kubenswrapper[4978]: E0225 06:59:05.936228 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124\": container with ID starting with 81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124 not found: ID does not exist" containerID="81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.936252 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124"} err="failed to get container status \"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124\": rpc error: code = NotFound desc = could not find container \"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124\": container with ID starting with 81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.936269 4978 scope.go:117] "RemoveContainer" containerID="166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671" Feb 25 06:59:05 crc kubenswrapper[4978]: E0225 06:59:05.936617 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671\": container with ID starting with 166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671 not found: ID does not exist" containerID="166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.936678 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671"} err="failed to get container status \"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671\": rpc error: code = NotFound desc = could not find container \"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671\": container with ID starting with 166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.936706 4978 scope.go:117] "RemoveContainer" containerID="e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574" Feb 25 06:59:05 crc kubenswrapper[4978]: E0225 06:59:05.937003 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574\": container with ID starting with e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574 not found: ID does not exist" containerID="e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.937024 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574"} err="failed to get container status \"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574\": rpc error: code = NotFound desc = could not find container \"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574\": container with ID starting with e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.937055 4978 scope.go:117] "RemoveContainer" containerID="9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.937273 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768"} err="failed to get container status \"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768\": rpc error: code = NotFound desc = could not find container \"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768\": container with ID starting with 9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.937308 4978 scope.go:117] "RemoveContainer" containerID="7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.937826 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22"} err="failed to get container status \"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22\": rpc error: code = NotFound desc = could not find container \"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22\": container with ID starting with 7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.937849 4978 scope.go:117] "RemoveContainer" containerID="078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.938214 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf"} err="failed to get container status \"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf\": rpc error: code = NotFound desc = could not find container \"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf\": container with ID starting with 078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.938237 4978 scope.go:117] "RemoveContainer" containerID="2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.938740 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3"} err="failed to get container status \"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3\": rpc error: code = NotFound desc = could not find container \"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3\": container with ID starting with 2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.938763 4978 scope.go:117] "RemoveContainer" containerID="ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.939153 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171"} err="failed to get container status \"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171\": rpc error: code = NotFound desc = could not find container \"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171\": container with ID starting with ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.939174 4978 scope.go:117] "RemoveContainer" containerID="585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.939704 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de"} err="failed to get container status \"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de\": rpc error: code = NotFound desc = could not find container \"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de\": container with ID starting with 585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.939756 4978 scope.go:117] "RemoveContainer" containerID="81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.940178 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124"} err="failed to get container status \"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124\": rpc error: code = NotFound desc = could not find container \"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124\": container with ID starting with 81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.940198 4978 scope.go:117] "RemoveContainer" containerID="166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.940822 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671"} err="failed to get container status \"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671\": rpc error: code = NotFound desc = could not find container \"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671\": container with ID starting with 166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.940846 4978 scope.go:117] "RemoveContainer" containerID="e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.941191 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574"} err="failed to get container status \"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574\": rpc error: code = NotFound desc = could not find container \"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574\": container with ID starting with e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.941214 4978 scope.go:117] "RemoveContainer" containerID="9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.941593 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768"} err="failed to get container status \"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768\": rpc error: code = NotFound desc = could not find container \"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768\": container with ID starting with 9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.941630 4978 scope.go:117] "RemoveContainer" containerID="7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.941928 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22"} err="failed to get container status \"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22\": rpc error: code = NotFound desc = could not find container \"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22\": container with ID starting with 7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.941952 4978 scope.go:117] "RemoveContainer" containerID="078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.942334 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf"} err="failed to get container status \"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf\": rpc error: code = NotFound desc = could not find container \"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf\": container with ID starting with 078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.942356 4978 scope.go:117] "RemoveContainer" containerID="2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.942613 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3"} err="failed to get container status \"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3\": rpc error: code = NotFound desc = could not find container \"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3\": container with ID starting with 2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.942737 4978 scope.go:117] "RemoveContainer" containerID="ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.943159 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171"} err="failed to get container status \"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171\": rpc error: code = NotFound desc = could not find container \"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171\": container with ID starting with ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.943248 4978 scope.go:117] "RemoveContainer" containerID="585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.943647 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de"} err="failed to get container status \"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de\": rpc error: code = NotFound desc = could not find container \"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de\": container with ID starting with 585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.943687 4978 scope.go:117] "RemoveContainer" containerID="81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.944307 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124"} err="failed to get container status \"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124\": rpc error: code = NotFound desc = could not find container \"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124\": container with ID starting with 81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.944422 4978 scope.go:117] "RemoveContainer" containerID="166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.944862 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671"} err="failed to get container status \"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671\": rpc error: code = NotFound desc = could not find container \"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671\": container with ID starting with 166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.944997 4978 scope.go:117] "RemoveContainer" containerID="e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.945420 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574"} err="failed to get container status \"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574\": rpc error: code = NotFound desc = could not find container \"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574\": container with ID starting with e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.945546 4978 scope.go:117] "RemoveContainer" containerID="9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.945825 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768"} err="failed to get container status \"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768\": rpc error: code = NotFound desc = could not find container \"9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768\": container with ID starting with 9057859fc9e7029fb43a66a3b1e5fb82775914bcf60b9372f3a6369d98e0d768 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.945915 4978 scope.go:117] "RemoveContainer" containerID="7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.946354 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22"} err="failed to get container status \"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22\": rpc error: code = NotFound desc = could not find container \"7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22\": container with ID starting with 7e545d6236fb12427d9903663e572dd4b9cefa6b3fc78cfc545ebc1ecbc77b22 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.946472 4978 scope.go:117] "RemoveContainer" containerID="078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.946804 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf"} err="failed to get container status \"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf\": rpc error: code = NotFound desc = could not find container \"078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf\": container with ID starting with 078625738cfa5da560f8b33c712266ae1fcf3d99c349b19792457b9d17e4dbbf not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.946845 4978 scope.go:117] "RemoveContainer" containerID="2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.947223 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3"} err="failed to get container status \"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3\": rpc error: code = NotFound desc = could not find container \"2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3\": container with ID starting with 2be4d20503111d219afa4f87b2200344265d579553fdc960fdfb6453029ab8b3 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.947259 4978 scope.go:117] "RemoveContainer" containerID="ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.947687 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171"} err="failed to get container status \"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171\": rpc error: code = NotFound desc = could not find container \"ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171\": container with ID starting with ea4de1dd840f41de1bdeb8960fe580067506ff228df61c821568f82b16070171 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.947754 4978 scope.go:117] "RemoveContainer" containerID="585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.948419 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de"} err="failed to get container status \"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de\": rpc error: code = NotFound desc = could not find container \"585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de\": container with ID starting with 585bb8c85d6c7241b54753b3dc22195745e22876473693c3b2ece3d6746612de not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.948466 4978 scope.go:117] "RemoveContainer" containerID="81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.949117 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124"} err="failed to get container status \"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124\": rpc error: code = NotFound desc = could not find container \"81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124\": container with ID starting with 81640dd16c7191a9934586734c0aa526bcea43bc8609eedfeb078a097c513124 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.949217 4978 scope.go:117] "RemoveContainer" containerID="166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.949681 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671"} err="failed to get container status \"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671\": rpc error: code = NotFound desc = could not find container \"166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671\": container with ID starting with 166361b933f1d3f1d18dc3efdbaa57e76b51f283b75f58f7f7326298df0a3671 not found: ID does not exist" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.949724 4978 scope.go:117] "RemoveContainer" containerID="e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574" Feb 25 06:59:05 crc kubenswrapper[4978]: I0225 06:59:05.950125 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574"} err="failed to get container status \"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574\": rpc error: code = NotFound desc = could not find container \"e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574\": container with ID starting with e24cccd48bc62df6c4bb093a7c57ee7e1b1b60be963a09f4ef1b9186aea97574 not found: ID does not exist" Feb 25 06:59:06 crc kubenswrapper[4978]: I0225 06:59:06.665893 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" event={"ID":"963d73c4-5aa9-481c-a32e-80fd85bf6093","Type":"ContainerStarted","Data":"da493d29df85ec99bf272ca2cb669502de44826a80bfb02a455d51982e9794a3"} Feb 25 06:59:06 crc kubenswrapper[4978]: I0225 06:59:06.666346 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" event={"ID":"963d73c4-5aa9-481c-a32e-80fd85bf6093","Type":"ContainerStarted","Data":"d64cad8f1fe4c0766de4b648c1373e97cdcc8350df0b61e47324fe5989576f1d"} Feb 25 06:59:06 crc kubenswrapper[4978]: I0225 06:59:06.666393 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" event={"ID":"963d73c4-5aa9-481c-a32e-80fd85bf6093","Type":"ContainerStarted","Data":"2daf5b83161474697a33eaf42140edbcd6f53889539e48dda75fe1f780046cae"} Feb 25 06:59:06 crc kubenswrapper[4978]: I0225 06:59:06.666410 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" event={"ID":"963d73c4-5aa9-481c-a32e-80fd85bf6093","Type":"ContainerStarted","Data":"3910bd1b7e09fd9350ee1704688d1fc04c916a74e68fdc9ae83e967aa8f5c9b6"} Feb 25 06:59:06 crc kubenswrapper[4978]: I0225 06:59:06.666423 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" event={"ID":"963d73c4-5aa9-481c-a32e-80fd85bf6093","Type":"ContainerStarted","Data":"21d7883ad44980e998d8912546788263fe5b06176dd54a4adc115b14469b6436"} Feb 25 06:59:06 crc kubenswrapper[4978]: I0225 06:59:06.666435 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" event={"ID":"963d73c4-5aa9-481c-a32e-80fd85bf6093","Type":"ContainerStarted","Data":"50a0e00a1d5a4a0458ac89a9cde8fa2c4fcb6c9e6f43e671cbf171ef4052426d"} Feb 25 06:59:07 crc kubenswrapper[4978]: I0225 06:59:07.346958 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5" path="/var/lib/kubelet/pods/b9a7a2cb-b9af-48db-aa1b-18beab1fe0b5/volumes" Feb 25 06:59:09 crc kubenswrapper[4978]: I0225 06:59:09.696815 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" event={"ID":"963d73c4-5aa9-481c-a32e-80fd85bf6093","Type":"ContainerStarted","Data":"c48c8fa005e7b430d41f03a586c768183c3bec233d1d8705295aabe98022c695"} Feb 25 06:59:11 crc kubenswrapper[4978]: I0225 06:59:11.714284 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" event={"ID":"963d73c4-5aa9-481c-a32e-80fd85bf6093","Type":"ContainerStarted","Data":"701285033e9ec44c187ab5ed784d921c01ec13ccd2445d58c1ec4bf7e9a4313f"} Feb 25 06:59:12 crc kubenswrapper[4978]: I0225 06:59:12.720618 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:12 crc kubenswrapper[4978]: I0225 06:59:12.720961 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:12 crc kubenswrapper[4978]: I0225 06:59:12.720981 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:12 crc kubenswrapper[4978]: I0225 06:59:12.758800 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:12 crc kubenswrapper[4978]: I0225 06:59:12.759904 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:12 crc kubenswrapper[4978]: I0225 06:59:12.768252 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" podStartSLOduration=8.76821986 podStartE2EDuration="8.76821986s" podCreationTimestamp="2026-02-25 06:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 06:59:12.762290421 +0000 UTC m=+846.201546930" watchObservedRunningTime="2026-02-25 06:59:12.76821986 +0000 UTC m=+846.207476379" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.171108 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-5nhcj"] Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.172436 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.175399 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.175537 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.176080 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.177314 4978 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-7m2v4" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.203111 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-5nhcj"] Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.273748 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/09ef89fc-bdd8-4513-b546-214c92481e2b-crc-storage\") pod \"crc-storage-crc-5nhcj\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.273812 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/09ef89fc-bdd8-4513-b546-214c92481e2b-node-mnt\") pod \"crc-storage-crc-5nhcj\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.273938 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq8gg\" (UniqueName: \"kubernetes.io/projected/09ef89fc-bdd8-4513-b546-214c92481e2b-kube-api-access-lq8gg\") pod \"crc-storage-crc-5nhcj\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.375435 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/09ef89fc-bdd8-4513-b546-214c92481e2b-node-mnt\") pod \"crc-storage-crc-5nhcj\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.375595 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq8gg\" (UniqueName: \"kubernetes.io/projected/09ef89fc-bdd8-4513-b546-214c92481e2b-kube-api-access-lq8gg\") pod \"crc-storage-crc-5nhcj\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.375694 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/09ef89fc-bdd8-4513-b546-214c92481e2b-crc-storage\") pod \"crc-storage-crc-5nhcj\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.375763 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/09ef89fc-bdd8-4513-b546-214c92481e2b-node-mnt\") pod \"crc-storage-crc-5nhcj\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.376948 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/09ef89fc-bdd8-4513-b546-214c92481e2b-crc-storage\") pod \"crc-storage-crc-5nhcj\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.398330 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq8gg\" (UniqueName: \"kubernetes.io/projected/09ef89fc-bdd8-4513-b546-214c92481e2b-kube-api-access-lq8gg\") pod \"crc-storage-crc-5nhcj\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.507618 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: E0225 06:59:14.537405 4978 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5nhcj_crc-storage_09ef89fc-bdd8-4513-b546-214c92481e2b_0(e739477131a96bbd800e9866bdc123480b0005610c7f90d7c94b52b6dd011591): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 25 06:59:14 crc kubenswrapper[4978]: E0225 06:59:14.537516 4978 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5nhcj_crc-storage_09ef89fc-bdd8-4513-b546-214c92481e2b_0(e739477131a96bbd800e9866bdc123480b0005610c7f90d7c94b52b6dd011591): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: E0225 06:59:14.537554 4978 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5nhcj_crc-storage_09ef89fc-bdd8-4513-b546-214c92481e2b_0(e739477131a96bbd800e9866bdc123480b0005610c7f90d7c94b52b6dd011591): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: E0225 06:59:14.537631 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-5nhcj_crc-storage(09ef89fc-bdd8-4513-b546-214c92481e2b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-5nhcj_crc-storage(09ef89fc-bdd8-4513-b546-214c92481e2b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5nhcj_crc-storage_09ef89fc-bdd8-4513-b546-214c92481e2b_0(e739477131a96bbd800e9866bdc123480b0005610c7f90d7c94b52b6dd011591): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-5nhcj" podUID="09ef89fc-bdd8-4513-b546-214c92481e2b" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.737231 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: I0225 06:59:14.738261 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: E0225 06:59:14.786406 4978 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5nhcj_crc-storage_09ef89fc-bdd8-4513-b546-214c92481e2b_0(234a65db4190628df5dd5a1a435196162e74fbedb101b3935ff6f5a03dfc1179): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 25 06:59:14 crc kubenswrapper[4978]: E0225 06:59:14.786495 4978 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5nhcj_crc-storage_09ef89fc-bdd8-4513-b546-214c92481e2b_0(234a65db4190628df5dd5a1a435196162e74fbedb101b3935ff6f5a03dfc1179): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: E0225 06:59:14.786532 4978 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5nhcj_crc-storage_09ef89fc-bdd8-4513-b546-214c92481e2b_0(234a65db4190628df5dd5a1a435196162e74fbedb101b3935ff6f5a03dfc1179): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:14 crc kubenswrapper[4978]: E0225 06:59:14.786602 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-5nhcj_crc-storage(09ef89fc-bdd8-4513-b546-214c92481e2b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-5nhcj_crc-storage(09ef89fc-bdd8-4513-b546-214c92481e2b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-5nhcj_crc-storage_09ef89fc-bdd8-4513-b546-214c92481e2b_0(234a65db4190628df5dd5a1a435196162e74fbedb101b3935ff6f5a03dfc1179): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-5nhcj" podUID="09ef89fc-bdd8-4513-b546-214c92481e2b" Feb 25 06:59:28 crc kubenswrapper[4978]: I0225 06:59:28.326688 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:28 crc kubenswrapper[4978]: I0225 06:59:28.328204 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:28 crc kubenswrapper[4978]: I0225 06:59:28.620966 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-5nhcj"] Feb 25 06:59:28 crc kubenswrapper[4978]: W0225 06:59:28.629711 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09ef89fc_bdd8_4513_b546_214c92481e2b.slice/crio-810d7f96d99262f10f64c82a98ae7cdc5e80e616a4d68a6123f9e2c00fe1b20b WatchSource:0}: Error finding container 810d7f96d99262f10f64c82a98ae7cdc5e80e616a4d68a6123f9e2c00fe1b20b: Status 404 returned error can't find the container with id 810d7f96d99262f10f64c82a98ae7cdc5e80e616a4d68a6123f9e2c00fe1b20b Feb 25 06:59:28 crc kubenswrapper[4978]: I0225 06:59:28.633794 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 06:59:28 crc kubenswrapper[4978]: I0225 06:59:28.842174 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5nhcj" event={"ID":"09ef89fc-bdd8-4513-b546-214c92481e2b","Type":"ContainerStarted","Data":"810d7f96d99262f10f64c82a98ae7cdc5e80e616a4d68a6123f9e2c00fe1b20b"} Feb 25 06:59:30 crc kubenswrapper[4978]: I0225 06:59:30.862959 4978 generic.go:334] "Generic (PLEG): container finished" podID="09ef89fc-bdd8-4513-b546-214c92481e2b" containerID="c0ec7df1bcc042b4a48d1cf800f5c8bb0371f7303e7803ef3bfd7cfcedcea1e5" exitCode=0 Feb 25 06:59:30 crc kubenswrapper[4978]: I0225 06:59:30.863186 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5nhcj" event={"ID":"09ef89fc-bdd8-4513-b546-214c92481e2b","Type":"ContainerDied","Data":"c0ec7df1bcc042b4a48d1cf800f5c8bb0371f7303e7803ef3bfd7cfcedcea1e5"} Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.231511 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.376563 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/09ef89fc-bdd8-4513-b546-214c92481e2b-crc-storage\") pod \"09ef89fc-bdd8-4513-b546-214c92481e2b\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.377179 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/09ef89fc-bdd8-4513-b546-214c92481e2b-node-mnt\") pod \"09ef89fc-bdd8-4513-b546-214c92481e2b\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.377285 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq8gg\" (UniqueName: \"kubernetes.io/projected/09ef89fc-bdd8-4513-b546-214c92481e2b-kube-api-access-lq8gg\") pod \"09ef89fc-bdd8-4513-b546-214c92481e2b\" (UID: \"09ef89fc-bdd8-4513-b546-214c92481e2b\") " Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.377518 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09ef89fc-bdd8-4513-b546-214c92481e2b-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "09ef89fc-bdd8-4513-b546-214c92481e2b" (UID: "09ef89fc-bdd8-4513-b546-214c92481e2b"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.377841 4978 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/09ef89fc-bdd8-4513-b546-214c92481e2b-node-mnt\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.387666 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ef89fc-bdd8-4513-b546-214c92481e2b-kube-api-access-lq8gg" (OuterVolumeSpecName: "kube-api-access-lq8gg") pod "09ef89fc-bdd8-4513-b546-214c92481e2b" (UID: "09ef89fc-bdd8-4513-b546-214c92481e2b"). InnerVolumeSpecName "kube-api-access-lq8gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.402839 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ef89fc-bdd8-4513-b546-214c92481e2b-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "09ef89fc-bdd8-4513-b546-214c92481e2b" (UID: "09ef89fc-bdd8-4513-b546-214c92481e2b"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.479007 4978 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/09ef89fc-bdd8-4513-b546-214c92481e2b-crc-storage\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.479923 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq8gg\" (UniqueName: \"kubernetes.io/projected/09ef89fc-bdd8-4513-b546-214c92481e2b-kube-api-access-lq8gg\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.886950 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-5nhcj" event={"ID":"09ef89fc-bdd8-4513-b546-214c92481e2b","Type":"ContainerDied","Data":"810d7f96d99262f10f64c82a98ae7cdc5e80e616a4d68a6123f9e2c00fe1b20b"} Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.887019 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="810d7f96d99262f10f64c82a98ae7cdc5e80e616a4d68a6123f9e2c00fe1b20b" Feb 25 06:59:32 crc kubenswrapper[4978]: I0225 06:59:32.887108 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-5nhcj" Feb 25 06:59:35 crc kubenswrapper[4978]: I0225 06:59:35.278322 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-sn6n7" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.527788 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl"] Feb 25 06:59:40 crc kubenswrapper[4978]: E0225 06:59:40.528574 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ef89fc-bdd8-4513-b546-214c92481e2b" containerName="storage" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.528590 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ef89fc-bdd8-4513-b546-214c92481e2b" containerName="storage" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.528699 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="09ef89fc-bdd8-4513-b546-214c92481e2b" containerName="storage" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.529607 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.532477 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.540015 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl"] Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.653683 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.653793 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlmzt\" (UniqueName: \"kubernetes.io/projected/107428a7-7b98-43a1-9bd1-a6c08c914569-kube-api-access-xlmzt\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.653875 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.755791 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlmzt\" (UniqueName: \"kubernetes.io/projected/107428a7-7b98-43a1-9bd1-a6c08c914569-kube-api-access-xlmzt\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.755920 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.756016 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.756919 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.757002 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.786876 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlmzt\" (UniqueName: \"kubernetes.io/projected/107428a7-7b98-43a1-9bd1-a6c08c914569-kube-api-access-xlmzt\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:40 crc kubenswrapper[4978]: I0225 06:59:40.849662 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:41 crc kubenswrapper[4978]: I0225 06:59:41.095451 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl"] Feb 25 06:59:41 crc kubenswrapper[4978]: I0225 06:59:41.960846 4978 generic.go:334] "Generic (PLEG): container finished" podID="107428a7-7b98-43a1-9bd1-a6c08c914569" containerID="f94d92ceafe3e06ddd1020d2e274e836590e2a3e6ffb6bde28573d589c4e09e3" exitCode=0 Feb 25 06:59:41 crc kubenswrapper[4978]: I0225 06:59:41.960975 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" event={"ID":"107428a7-7b98-43a1-9bd1-a6c08c914569","Type":"ContainerDied","Data":"f94d92ceafe3e06ddd1020d2e274e836590e2a3e6ffb6bde28573d589c4e09e3"} Feb 25 06:59:41 crc kubenswrapper[4978]: I0225 06:59:41.963172 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" event={"ID":"107428a7-7b98-43a1-9bd1-a6c08c914569","Type":"ContainerStarted","Data":"5031f8dfd274dd755644e9317f97da394e3d745d592e119636f964381630f490"} Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.601065 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5v29t"] Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.603343 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.620922 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5v29t"] Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.689212 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-catalog-content\") pod \"redhat-operators-5v29t\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.689306 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5zcj\" (UniqueName: \"kubernetes.io/projected/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-kube-api-access-j5zcj\") pod \"redhat-operators-5v29t\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.689650 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-utilities\") pod \"redhat-operators-5v29t\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.791183 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-catalog-content\") pod \"redhat-operators-5v29t\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.791355 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5zcj\" (UniqueName: \"kubernetes.io/projected/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-kube-api-access-j5zcj\") pod \"redhat-operators-5v29t\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.791562 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-utilities\") pod \"redhat-operators-5v29t\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.792524 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-utilities\") pod \"redhat-operators-5v29t\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.792545 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-catalog-content\") pod \"redhat-operators-5v29t\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.829074 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5zcj\" (UniqueName: \"kubernetes.io/projected/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-kube-api-access-j5zcj\") pod \"redhat-operators-5v29t\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:42 crc kubenswrapper[4978]: I0225 06:59:42.932403 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:43 crc kubenswrapper[4978]: I0225 06:59:43.198916 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5v29t"] Feb 25 06:59:43 crc kubenswrapper[4978]: I0225 06:59:43.976317 4978 generic.go:334] "Generic (PLEG): container finished" podID="107428a7-7b98-43a1-9bd1-a6c08c914569" containerID="38b04e444a813826d1fc2d747d86d99e040809df550fede28f917751124077bf" exitCode=0 Feb 25 06:59:43 crc kubenswrapper[4978]: I0225 06:59:43.976422 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" event={"ID":"107428a7-7b98-43a1-9bd1-a6c08c914569","Type":"ContainerDied","Data":"38b04e444a813826d1fc2d747d86d99e040809df550fede28f917751124077bf"} Feb 25 06:59:43 crc kubenswrapper[4978]: I0225 06:59:43.978083 4978 generic.go:334] "Generic (PLEG): container finished" podID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerID="f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836" exitCode=0 Feb 25 06:59:43 crc kubenswrapper[4978]: I0225 06:59:43.978135 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v29t" event={"ID":"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db","Type":"ContainerDied","Data":"f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836"} Feb 25 06:59:43 crc kubenswrapper[4978]: I0225 06:59:43.978183 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v29t" event={"ID":"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db","Type":"ContainerStarted","Data":"79d8af0785eb7fcdcd63f1486ced61ab609f438943f00652463fc3ac64f99837"} Feb 25 06:59:44 crc kubenswrapper[4978]: I0225 06:59:44.989176 4978 generic.go:334] "Generic (PLEG): container finished" podID="107428a7-7b98-43a1-9bd1-a6c08c914569" containerID="f54a58795b5e4ce68dd4e6d2207fa05252d8467aba2521a969f655b5e7f5b43a" exitCode=0 Feb 25 06:59:44 crc kubenswrapper[4978]: I0225 06:59:44.989236 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" event={"ID":"107428a7-7b98-43a1-9bd1-a6c08c914569","Type":"ContainerDied","Data":"f54a58795b5e4ce68dd4e6d2207fa05252d8467aba2521a969f655b5e7f5b43a"} Feb 25 06:59:45 crc kubenswrapper[4978]: I0225 06:59:45.997911 4978 generic.go:334] "Generic (PLEG): container finished" podID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerID="813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b" exitCode=0 Feb 25 06:59:45 crc kubenswrapper[4978]: I0225 06:59:45.998014 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v29t" event={"ID":"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db","Type":"ContainerDied","Data":"813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b"} Feb 25 06:59:46 crc kubenswrapper[4978]: I0225 06:59:46.331507 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:46 crc kubenswrapper[4978]: I0225 06:59:46.456656 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-bundle\") pod \"107428a7-7b98-43a1-9bd1-a6c08c914569\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " Feb 25 06:59:46 crc kubenswrapper[4978]: I0225 06:59:46.456803 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlmzt\" (UniqueName: \"kubernetes.io/projected/107428a7-7b98-43a1-9bd1-a6c08c914569-kube-api-access-xlmzt\") pod \"107428a7-7b98-43a1-9bd1-a6c08c914569\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " Feb 25 06:59:46 crc kubenswrapper[4978]: I0225 06:59:46.456990 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-util\") pod \"107428a7-7b98-43a1-9bd1-a6c08c914569\" (UID: \"107428a7-7b98-43a1-9bd1-a6c08c914569\") " Feb 25 06:59:46 crc kubenswrapper[4978]: I0225 06:59:46.457794 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-bundle" (OuterVolumeSpecName: "bundle") pod "107428a7-7b98-43a1-9bd1-a6c08c914569" (UID: "107428a7-7b98-43a1-9bd1-a6c08c914569"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:59:46 crc kubenswrapper[4978]: I0225 06:59:46.468697 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/107428a7-7b98-43a1-9bd1-a6c08c914569-kube-api-access-xlmzt" (OuterVolumeSpecName: "kube-api-access-xlmzt") pod "107428a7-7b98-43a1-9bd1-a6c08c914569" (UID: "107428a7-7b98-43a1-9bd1-a6c08c914569"). InnerVolumeSpecName "kube-api-access-xlmzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 06:59:46 crc kubenswrapper[4978]: I0225 06:59:46.491625 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-util" (OuterVolumeSpecName: "util") pod "107428a7-7b98-43a1-9bd1-a6c08c914569" (UID: "107428a7-7b98-43a1-9bd1-a6c08c914569"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 06:59:46 crc kubenswrapper[4978]: I0225 06:59:46.558619 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlmzt\" (UniqueName: \"kubernetes.io/projected/107428a7-7b98-43a1-9bd1-a6c08c914569-kube-api-access-xlmzt\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:46 crc kubenswrapper[4978]: I0225 06:59:46.558672 4978 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-util\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:46 crc kubenswrapper[4978]: I0225 06:59:46.558688 4978 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/107428a7-7b98-43a1-9bd1-a6c08c914569-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 06:59:47 crc kubenswrapper[4978]: I0225 06:59:47.007726 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v29t" event={"ID":"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db","Type":"ContainerStarted","Data":"1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02"} Feb 25 06:59:47 crc kubenswrapper[4978]: I0225 06:59:47.011226 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" event={"ID":"107428a7-7b98-43a1-9bd1-a6c08c914569","Type":"ContainerDied","Data":"5031f8dfd274dd755644e9317f97da394e3d745d592e119636f964381630f490"} Feb 25 06:59:47 crc kubenswrapper[4978]: I0225 06:59:47.011333 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl" Feb 25 06:59:47 crc kubenswrapper[4978]: I0225 06:59:47.012134 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5031f8dfd274dd755644e9317f97da394e3d745d592e119636f964381630f490" Feb 25 06:59:47 crc kubenswrapper[4978]: I0225 06:59:47.042187 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5v29t" podStartSLOduration=2.612617346 podStartE2EDuration="5.042136071s" podCreationTimestamp="2026-02-25 06:59:42 +0000 UTC" firstStartedPulling="2026-02-25 06:59:43.979799714 +0000 UTC m=+877.419056173" lastFinishedPulling="2026-02-25 06:59:46.409318439 +0000 UTC m=+879.848574898" observedRunningTime="2026-02-25 06:59:47.033258249 +0000 UTC m=+880.472514768" watchObservedRunningTime="2026-02-25 06:59:47.042136071 +0000 UTC m=+880.481392560" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.152758 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-6cv5l"] Feb 25 06:59:51 crc kubenswrapper[4978]: E0225 06:59:51.153314 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="107428a7-7b98-43a1-9bd1-a6c08c914569" containerName="util" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.153330 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="107428a7-7b98-43a1-9bd1-a6c08c914569" containerName="util" Feb 25 06:59:51 crc kubenswrapper[4978]: E0225 06:59:51.153342 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="107428a7-7b98-43a1-9bd1-a6c08c914569" containerName="extract" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.153349 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="107428a7-7b98-43a1-9bd1-a6c08c914569" containerName="extract" Feb 25 06:59:51 crc kubenswrapper[4978]: E0225 06:59:51.153394 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="107428a7-7b98-43a1-9bd1-a6c08c914569" containerName="pull" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.153403 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="107428a7-7b98-43a1-9bd1-a6c08c914569" containerName="pull" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.153518 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="107428a7-7b98-43a1-9bd1-a6c08c914569" containerName="extract" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.153942 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-6cv5l" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.155729 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.156853 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.156941 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-9lcsx" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.176479 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-6cv5l"] Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.231952 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h627z\" (UniqueName: \"kubernetes.io/projected/107b6a29-7b92-4c24-a38e-5ae9ec345730-kube-api-access-h627z\") pod \"nmstate-operator-694c9596b7-6cv5l\" (UID: \"107b6a29-7b92-4c24-a38e-5ae9ec345730\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-6cv5l" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.332752 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h627z\" (UniqueName: \"kubernetes.io/projected/107b6a29-7b92-4c24-a38e-5ae9ec345730-kube-api-access-h627z\") pod \"nmstate-operator-694c9596b7-6cv5l\" (UID: \"107b6a29-7b92-4c24-a38e-5ae9ec345730\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-6cv5l" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.353018 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h627z\" (UniqueName: \"kubernetes.io/projected/107b6a29-7b92-4c24-a38e-5ae9ec345730-kube-api-access-h627z\") pod \"nmstate-operator-694c9596b7-6cv5l\" (UID: \"107b6a29-7b92-4c24-a38e-5ae9ec345730\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-6cv5l" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.474079 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-6cv5l" Feb 25 06:59:51 crc kubenswrapper[4978]: I0225 06:59:51.946970 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-6cv5l"] Feb 25 06:59:52 crc kubenswrapper[4978]: I0225 06:59:52.048810 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-6cv5l" event={"ID":"107b6a29-7b92-4c24-a38e-5ae9ec345730","Type":"ContainerStarted","Data":"e023bac53b542dca7b1dfd108244efdb585e59e845133da4c97a53c8836539c3"} Feb 25 06:59:52 crc kubenswrapper[4978]: I0225 06:59:52.932967 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:52 crc kubenswrapper[4978]: I0225 06:59:52.933484 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 06:59:53 crc kubenswrapper[4978]: I0225 06:59:53.999624 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5v29t" podUID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerName="registry-server" probeResult="failure" output=< Feb 25 06:59:53 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 06:59:53 crc kubenswrapper[4978]: > Feb 25 06:59:55 crc kubenswrapper[4978]: I0225 06:59:55.076215 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-6cv5l" event={"ID":"107b6a29-7b92-4c24-a38e-5ae9ec345730","Type":"ContainerStarted","Data":"ccf950fb2a8a981fbe99a5b8427e68579538f55abba44284c79a561b07ff26c8"} Feb 25 06:59:55 crc kubenswrapper[4978]: I0225 06:59:55.105286 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-6cv5l" podStartSLOduration=1.714175561 podStartE2EDuration="4.105259376s" podCreationTimestamp="2026-02-25 06:59:51 +0000 UTC" firstStartedPulling="2026-02-25 06:59:51.957441207 +0000 UTC m=+885.396697706" lastFinishedPulling="2026-02-25 06:59:54.348525032 +0000 UTC m=+887.787781521" observedRunningTime="2026-02-25 06:59:55.100693261 +0000 UTC m=+888.539949760" watchObservedRunningTime="2026-02-25 06:59:55.105259376 +0000 UTC m=+888.544515865" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.182492 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533380-mbpf9"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.184988 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533380-mbpf9" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.193230 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.195418 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.195698 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.211100 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.212186 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.214293 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.214712 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.263598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kprz6\" (UniqueName: \"kubernetes.io/projected/87f4f3a6-8318-45cd-8942-b3a58210c930-kube-api-access-kprz6\") pod \"auto-csr-approver-29533380-mbpf9\" (UID: \"87f4f3a6-8318-45cd-8942-b3a58210c930\") " pod="openshift-infra/auto-csr-approver-29533380-mbpf9" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.280752 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533380-mbpf9"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.286255 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.364309 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-secret-volume\") pod \"collect-profiles-29533380-7vwq8\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.364376 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnz66\" (UniqueName: \"kubernetes.io/projected/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-kube-api-access-cnz66\") pod \"collect-profiles-29533380-7vwq8\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.364666 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-config-volume\") pod \"collect-profiles-29533380-7vwq8\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.364919 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kprz6\" (UniqueName: \"kubernetes.io/projected/87f4f3a6-8318-45cd-8942-b3a58210c930-kube-api-access-kprz6\") pod \"auto-csr-approver-29533380-mbpf9\" (UID: \"87f4f3a6-8318-45cd-8942-b3a58210c930\") " pod="openshift-infra/auto-csr-approver-29533380-mbpf9" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.393651 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kprz6\" (UniqueName: \"kubernetes.io/projected/87f4f3a6-8318-45cd-8942-b3a58210c930-kube-api-access-kprz6\") pod \"auto-csr-approver-29533380-mbpf9\" (UID: \"87f4f3a6-8318-45cd-8942-b3a58210c930\") " pod="openshift-infra/auto-csr-approver-29533380-mbpf9" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.466271 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-secret-volume\") pod \"collect-profiles-29533380-7vwq8\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.466330 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnz66\" (UniqueName: \"kubernetes.io/projected/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-kube-api-access-cnz66\") pod \"collect-profiles-29533380-7vwq8\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.466555 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-config-volume\") pod \"collect-profiles-29533380-7vwq8\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.467512 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-config-volume\") pod \"collect-profiles-29533380-7vwq8\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.470175 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-secret-volume\") pod \"collect-profiles-29533380-7vwq8\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.481980 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnz66\" (UniqueName: \"kubernetes.io/projected/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-kube-api-access-cnz66\") pod \"collect-profiles-29533380-7vwq8\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.506982 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533380-mbpf9" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.528974 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.651808 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.668749 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.680734 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.681400 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-dkvsg" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.681672 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.682887 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.691180 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.693319 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-vjh9q"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.694286 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.698774 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.766630 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.768001 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.770746 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.771016 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-mlk4p" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.771021 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.778872 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/81d07653-9826-45dd-a0fd-68a2bd37b8e2-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-hlrw5\" (UID: \"81d07653-9826-45dd-a0fd-68a2bd37b8e2\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.778926 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-ovs-socket\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.778982 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-nmstate-lock\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.779003 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq6w9\" (UniqueName: \"kubernetes.io/projected/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-kube-api-access-lq6w9\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.779025 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcq7t\" (UniqueName: \"kubernetes.io/projected/81d07653-9826-45dd-a0fd-68a2bd37b8e2-kube-api-access-fcq7t\") pod \"nmstate-webhook-866bcb46dc-hlrw5\" (UID: \"81d07653-9826-45dd-a0fd-68a2bd37b8e2\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.779051 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p2jq\" (UniqueName: \"kubernetes.io/projected/31fedbf4-35ef-4440-9039-341b3a0ece67-kube-api-access-6p2jq\") pod \"nmstate-metrics-58c85c668d-m8ksd\" (UID: \"31fedbf4-35ef-4440-9039-341b3a0ece67\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.779071 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-dbus-socket\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.781103 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881212 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-nmstate-lock\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881321 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq6w9\" (UniqueName: \"kubernetes.io/projected/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-kube-api-access-lq6w9\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881358 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-nmstate-lock\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881394 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcq7t\" (UniqueName: \"kubernetes.io/projected/81d07653-9826-45dd-a0fd-68a2bd37b8e2-kube-api-access-fcq7t\") pod \"nmstate-webhook-866bcb46dc-hlrw5\" (UID: \"81d07653-9826-45dd-a0fd-68a2bd37b8e2\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881434 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14d5cc0e-ed1f-4894-afde-ae5164197176-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-b7qb4\" (UID: \"14d5cc0e-ed1f-4894-afde-ae5164197176\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881489 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p2jq\" (UniqueName: \"kubernetes.io/projected/31fedbf4-35ef-4440-9039-341b3a0ece67-kube-api-access-6p2jq\") pod \"nmstate-metrics-58c85c668d-m8ksd\" (UID: \"31fedbf4-35ef-4440-9039-341b3a0ece67\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881525 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-dbus-socket\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881583 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/81d07653-9826-45dd-a0fd-68a2bd37b8e2-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-hlrw5\" (UID: \"81d07653-9826-45dd-a0fd-68a2bd37b8e2\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881607 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-ovs-socket\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881675 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85c5b\" (UniqueName: \"kubernetes.io/projected/14d5cc0e-ed1f-4894-afde-ae5164197176-kube-api-access-85c5b\") pod \"nmstate-console-plugin-5c78fc5d65-b7qb4\" (UID: \"14d5cc0e-ed1f-4894-afde-ae5164197176\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881728 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/14d5cc0e-ed1f-4894-afde-ae5164197176-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-b7qb4\" (UID: \"14d5cc0e-ed1f-4894-afde-ae5164197176\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.881912 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-ovs-socket\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.882190 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-dbus-socket\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.888788 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/81d07653-9826-45dd-a0fd-68a2bd37b8e2-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-hlrw5\" (UID: \"81d07653-9826-45dd-a0fd-68a2bd37b8e2\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.904141 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcq7t\" (UniqueName: \"kubernetes.io/projected/81d07653-9826-45dd-a0fd-68a2bd37b8e2-kube-api-access-fcq7t\") pod \"nmstate-webhook-866bcb46dc-hlrw5\" (UID: \"81d07653-9826-45dd-a0fd-68a2bd37b8e2\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.909554 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq6w9\" (UniqueName: \"kubernetes.io/projected/5a297fa5-61f7-4feb-b6b0-532d3d8509f9-kube-api-access-lq6w9\") pod \"nmstate-handler-vjh9q\" (UID: \"5a297fa5-61f7-4feb-b6b0-532d3d8509f9\") " pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.912710 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p2jq\" (UniqueName: \"kubernetes.io/projected/31fedbf4-35ef-4440-9039-341b3a0ece67-kube-api-access-6p2jq\") pod \"nmstate-metrics-58c85c668d-m8ksd\" (UID: \"31fedbf4-35ef-4440-9039-341b3a0ece67\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.965488 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-574869f-nlz7w"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.966627 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.969300 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-574869f-nlz7w"] Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.983018 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85c5b\" (UniqueName: \"kubernetes.io/projected/14d5cc0e-ed1f-4894-afde-ae5164197176-kube-api-access-85c5b\") pod \"nmstate-console-plugin-5c78fc5d65-b7qb4\" (UID: \"14d5cc0e-ed1f-4894-afde-ae5164197176\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.983073 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/14d5cc0e-ed1f-4894-afde-ae5164197176-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-b7qb4\" (UID: \"14d5cc0e-ed1f-4894-afde-ae5164197176\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.983127 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14d5cc0e-ed1f-4894-afde-ae5164197176-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-b7qb4\" (UID: \"14d5cc0e-ed1f-4894-afde-ae5164197176\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.984215 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/14d5cc0e-ed1f-4894-afde-ae5164197176-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-b7qb4\" (UID: \"14d5cc0e-ed1f-4894-afde-ae5164197176\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:00 crc kubenswrapper[4978]: I0225 07:00:00.987066 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14d5cc0e-ed1f-4894-afde-ae5164197176-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-b7qb4\" (UID: \"14d5cc0e-ed1f-4894-afde-ae5164197176\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.003639 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85c5b\" (UniqueName: \"kubernetes.io/projected/14d5cc0e-ed1f-4894-afde-ae5164197176-kube-api-access-85c5b\") pod \"nmstate-console-plugin-5c78fc5d65-b7qb4\" (UID: \"14d5cc0e-ed1f-4894-afde-ae5164197176\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.010176 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.032747 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.045946 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533380-mbpf9"] Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.049557 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:01 crc kubenswrapper[4978]: W0225 07:00:01.054878 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87f4f3a6_8318_45cd_8942_b3a58210c930.slice/crio-4076e4f4aac5f52931af7640d625bce33ad3b5152795f8e81a2b5cd35640c244 WatchSource:0}: Error finding container 4076e4f4aac5f52931af7640d625bce33ad3b5152795f8e81a2b5cd35640c244: Status 404 returned error can't find the container with id 4076e4f4aac5f52931af7640d625bce33ad3b5152795f8e81a2b5cd35640c244 Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.059328 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8"] Feb 25 07:00:01 crc kubenswrapper[4978]: W0225 07:00:01.063540 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b2bef0e_1e45_4eb3_9525_3328e351fa3d.slice/crio-552b849e550bb747af8fada9a9078941f17ca3cae03c273596e4b477e627a35d WatchSource:0}: Error finding container 552b849e550bb747af8fada9a9078941f17ca3cae03c273596e4b477e627a35d: Status 404 returned error can't find the container with id 552b849e550bb747af8fada9a9078941f17ca3cae03c273596e4b477e627a35d Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.083844 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-service-ca\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.083918 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/90574f74-0b1a-43f2-87a6-916677274604-console-oauth-config\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.083948 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/90574f74-0b1a-43f2-87a6-916677274604-console-serving-cert\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.084024 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vjzn\" (UniqueName: \"kubernetes.io/projected/90574f74-0b1a-43f2-87a6-916677274604-kube-api-access-6vjzn\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.084063 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-trusted-ca-bundle\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.084082 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-console-config\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.084097 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-oauth-serving-cert\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.085289 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" Feb 25 07:00:01 crc kubenswrapper[4978]: W0225 07:00:01.085819 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a297fa5_61f7_4feb_b6b0_532d3d8509f9.slice/crio-eac06f671d396eae966e75a83408bfd768eeb4b96494d1977b3631eecc93161b WatchSource:0}: Error finding container eac06f671d396eae966e75a83408bfd768eeb4b96494d1977b3631eecc93161b: Status 404 returned error can't find the container with id eac06f671d396eae966e75a83408bfd768eeb4b96494d1977b3631eecc93161b Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.148010 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" event={"ID":"1b2bef0e-1e45-4eb3-9525-3328e351fa3d","Type":"ContainerStarted","Data":"552b849e550bb747af8fada9a9078941f17ca3cae03c273596e4b477e627a35d"} Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.153145 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vjh9q" event={"ID":"5a297fa5-61f7-4feb-b6b0-532d3d8509f9","Type":"ContainerStarted","Data":"eac06f671d396eae966e75a83408bfd768eeb4b96494d1977b3631eecc93161b"} Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.154140 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533380-mbpf9" event={"ID":"87f4f3a6-8318-45cd-8942-b3a58210c930","Type":"ContainerStarted","Data":"4076e4f4aac5f52931af7640d625bce33ad3b5152795f8e81a2b5cd35640c244"} Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.185449 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vjzn\" (UniqueName: \"kubernetes.io/projected/90574f74-0b1a-43f2-87a6-916677274604-kube-api-access-6vjzn\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.185505 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-trusted-ca-bundle\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.185528 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-console-config\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.185553 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-oauth-serving-cert\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.185610 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-service-ca\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.185664 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/90574f74-0b1a-43f2-87a6-916677274604-console-oauth-config\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.185688 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/90574f74-0b1a-43f2-87a6-916677274604-console-serving-cert\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.189192 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-trusted-ca-bundle\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.190288 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-service-ca\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.190744 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-console-config\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.190865 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/90574f74-0b1a-43f2-87a6-916677274604-oauth-serving-cert\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.212501 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/90574f74-0b1a-43f2-87a6-916677274604-console-serving-cert\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.212760 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vjzn\" (UniqueName: \"kubernetes.io/projected/90574f74-0b1a-43f2-87a6-916677274604-kube-api-access-6vjzn\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.213787 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/90574f74-0b1a-43f2-87a6-916677274604-console-oauth-config\") pod \"console-574869f-nlz7w\" (UID: \"90574f74-0b1a-43f2-87a6-916677274604\") " pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.238741 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd"] Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.272849 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5"] Feb 25 07:00:01 crc kubenswrapper[4978]: W0225 07:00:01.278318 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81d07653_9826_45dd_a0fd_68a2bd37b8e2.slice/crio-d63e5a1cdcaa7abd38c791be09de458ca2b4a50edc92c05f89dd16df58745dac WatchSource:0}: Error finding container d63e5a1cdcaa7abd38c791be09de458ca2b4a50edc92c05f89dd16df58745dac: Status 404 returned error can't find the container with id d63e5a1cdcaa7abd38c791be09de458ca2b4a50edc92c05f89dd16df58745dac Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.280957 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.320702 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4"] Feb 25 07:00:01 crc kubenswrapper[4978]: W0225 07:00:01.325810 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14d5cc0e_ed1f_4894_afde_ae5164197176.slice/crio-89244f0e1282515af0cc446ddd4568875660b84074ab90150b08f3ddfb7fa25c WatchSource:0}: Error finding container 89244f0e1282515af0cc446ddd4568875660b84074ab90150b08f3ddfb7fa25c: Status 404 returned error can't find the container with id 89244f0e1282515af0cc446ddd4568875660b84074ab90150b08f3ddfb7fa25c Feb 25 07:00:01 crc kubenswrapper[4978]: I0225 07:00:01.474070 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-574869f-nlz7w"] Feb 25 07:00:01 crc kubenswrapper[4978]: W0225 07:00:01.483597 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90574f74_0b1a_43f2_87a6_916677274604.slice/crio-ce19e576bd26732432294decb6600960cb397e5c4805275eb830ae31f4c58363 WatchSource:0}: Error finding container ce19e576bd26732432294decb6600960cb397e5c4805275eb830ae31f4c58363: Status 404 returned error can't find the container with id ce19e576bd26732432294decb6600960cb397e5c4805275eb830ae31f4c58363 Feb 25 07:00:02 crc kubenswrapper[4978]: I0225 07:00:02.165074 4978 generic.go:334] "Generic (PLEG): container finished" podID="1b2bef0e-1e45-4eb3-9525-3328e351fa3d" containerID="5027f2c80274584aa70cb25ea3445ad7171efdbc698dfc62136aa6c03ee2eec7" exitCode=0 Feb 25 07:00:02 crc kubenswrapper[4978]: I0225 07:00:02.165165 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" event={"ID":"1b2bef0e-1e45-4eb3-9525-3328e351fa3d","Type":"ContainerDied","Data":"5027f2c80274584aa70cb25ea3445ad7171efdbc698dfc62136aa6c03ee2eec7"} Feb 25 07:00:02 crc kubenswrapper[4978]: I0225 07:00:02.167894 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-574869f-nlz7w" event={"ID":"90574f74-0b1a-43f2-87a6-916677274604","Type":"ContainerStarted","Data":"bbcd5b955dbed64601b027a1d523b547481186066732699af7160e06408ed621"} Feb 25 07:00:02 crc kubenswrapper[4978]: I0225 07:00:02.167940 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-574869f-nlz7w" event={"ID":"90574f74-0b1a-43f2-87a6-916677274604","Type":"ContainerStarted","Data":"ce19e576bd26732432294decb6600960cb397e5c4805275eb830ae31f4c58363"} Feb 25 07:00:02 crc kubenswrapper[4978]: I0225 07:00:02.170815 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd" event={"ID":"31fedbf4-35ef-4440-9039-341b3a0ece67","Type":"ContainerStarted","Data":"faf6aca1b7ce9ac6dbbbac3b0ed988e9ad74bb1ee0b957bd842904ba7881de9e"} Feb 25 07:00:02 crc kubenswrapper[4978]: I0225 07:00:02.172994 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" event={"ID":"81d07653-9826-45dd-a0fd-68a2bd37b8e2","Type":"ContainerStarted","Data":"d63e5a1cdcaa7abd38c791be09de458ca2b4a50edc92c05f89dd16df58745dac"} Feb 25 07:00:02 crc kubenswrapper[4978]: I0225 07:00:02.175075 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" event={"ID":"14d5cc0e-ed1f-4894-afde-ae5164197176","Type":"ContainerStarted","Data":"89244f0e1282515af0cc446ddd4568875660b84074ab90150b08f3ddfb7fa25c"} Feb 25 07:00:02 crc kubenswrapper[4978]: I0225 07:00:02.226837 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-574869f-nlz7w" podStartSLOduration=2.226805572 podStartE2EDuration="2.226805572s" podCreationTimestamp="2026-02-25 07:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:00:02.220331257 +0000 UTC m=+895.659587796" watchObservedRunningTime="2026-02-25 07:00:02.226805572 +0000 UTC m=+895.666062071" Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.002249 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.075992 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.255807 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5v29t"] Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.529825 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.626885 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-secret-volume\") pod \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.626949 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-config-volume\") pod \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.626984 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnz66\" (UniqueName: \"kubernetes.io/projected/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-kube-api-access-cnz66\") pod \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\" (UID: \"1b2bef0e-1e45-4eb3-9525-3328e351fa3d\") " Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.628850 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-config-volume" (OuterVolumeSpecName: "config-volume") pod "1b2bef0e-1e45-4eb3-9525-3328e351fa3d" (UID: "1b2bef0e-1e45-4eb3-9525-3328e351fa3d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.640641 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1b2bef0e-1e45-4eb3-9525-3328e351fa3d" (UID: "1b2bef0e-1e45-4eb3-9525-3328e351fa3d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.640672 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-kube-api-access-cnz66" (OuterVolumeSpecName: "kube-api-access-cnz66") pod "1b2bef0e-1e45-4eb3-9525-3328e351fa3d" (UID: "1b2bef0e-1e45-4eb3-9525-3328e351fa3d"). InnerVolumeSpecName "kube-api-access-cnz66". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.729096 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.729661 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:03 crc kubenswrapper[4978]: I0225 07:00:03.729691 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnz66\" (UniqueName: \"kubernetes.io/projected/1b2bef0e-1e45-4eb3-9525-3328e351fa3d-kube-api-access-cnz66\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.196173 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.196165 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8" event={"ID":"1b2bef0e-1e45-4eb3-9525-3328e351fa3d","Type":"ContainerDied","Data":"552b849e550bb747af8fada9a9078941f17ca3cae03c273596e4b477e627a35d"} Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.196304 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="552b849e550bb747af8fada9a9078941f17ca3cae03c273596e4b477e627a35d" Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.196400 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5v29t" podUID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerName="registry-server" containerID="cri-o://1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02" gracePeriod=2 Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.674920 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.743283 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-catalog-content\") pod \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.743515 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5zcj\" (UniqueName: \"kubernetes.io/projected/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-kube-api-access-j5zcj\") pod \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.743634 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-utilities\") pod \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\" (UID: \"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db\") " Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.744785 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-utilities" (OuterVolumeSpecName: "utilities") pod "d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" (UID: "d3a1c35d-75dc-4f68-b5a1-bc40cb4555db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.751767 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-kube-api-access-j5zcj" (OuterVolumeSpecName: "kube-api-access-j5zcj") pod "d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" (UID: "d3a1c35d-75dc-4f68-b5a1-bc40cb4555db"). InnerVolumeSpecName "kube-api-access-j5zcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.844797 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.844829 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5zcj\" (UniqueName: \"kubernetes.io/projected/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-kube-api-access-j5zcj\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.874379 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" (UID: "d3a1c35d-75dc-4f68-b5a1-bc40cb4555db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:00:04 crc kubenswrapper[4978]: I0225 07:00:04.945508 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.209979 4978 generic.go:334] "Generic (PLEG): container finished" podID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerID="1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02" exitCode=0 Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.210042 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v29t" event={"ID":"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db","Type":"ContainerDied","Data":"1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02"} Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.210083 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5v29t" event={"ID":"d3a1c35d-75dc-4f68-b5a1-bc40cb4555db","Type":"ContainerDied","Data":"79d8af0785eb7fcdcd63f1486ced61ab609f438943f00652463fc3ac64f99837"} Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.210112 4978 scope.go:117] "RemoveContainer" containerID="1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.210170 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5v29t" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.243052 4978 scope.go:117] "RemoveContainer" containerID="813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.275680 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5v29t"] Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.286280 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5v29t"] Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.293955 4978 scope.go:117] "RemoveContainer" containerID="f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.316609 4978 scope.go:117] "RemoveContainer" containerID="1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02" Feb 25 07:00:05 crc kubenswrapper[4978]: E0225 07:00:05.317109 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02\": container with ID starting with 1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02 not found: ID does not exist" containerID="1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.317157 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02"} err="failed to get container status \"1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02\": rpc error: code = NotFound desc = could not find container \"1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02\": container with ID starting with 1ad263cadaa46ac424c7d96ce288e9c84a3ea02b6e5866d0e0c92fa12bf11c02 not found: ID does not exist" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.317185 4978 scope.go:117] "RemoveContainer" containerID="813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b" Feb 25 07:00:05 crc kubenswrapper[4978]: E0225 07:00:05.317574 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b\": container with ID starting with 813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b not found: ID does not exist" containerID="813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.317608 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b"} err="failed to get container status \"813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b\": rpc error: code = NotFound desc = could not find container \"813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b\": container with ID starting with 813df45e1a250398c3b49f6a109d8a4d8ebed8466846cf2cab9ea8afe7d75b7b not found: ID does not exist" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.317626 4978 scope.go:117] "RemoveContainer" containerID="f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836" Feb 25 07:00:05 crc kubenswrapper[4978]: E0225 07:00:05.317942 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836\": container with ID starting with f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836 not found: ID does not exist" containerID="f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.318010 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836"} err="failed to get container status \"f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836\": rpc error: code = NotFound desc = could not find container \"f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836\": container with ID starting with f9bfdcbeaa431e3f3fd0be0c1cc6eb3372cc68cfadcb34f2b0eca12d799e2836 not found: ID does not exist" Feb 25 07:00:05 crc kubenswrapper[4978]: I0225 07:00:05.335804 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" path="/var/lib/kubelet/pods/d3a1c35d-75dc-4f68-b5a1-bc40cb4555db/volumes" Feb 25 07:00:07 crc kubenswrapper[4978]: I0225 07:00:07.227278 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vjh9q" event={"ID":"5a297fa5-61f7-4feb-b6b0-532d3d8509f9","Type":"ContainerStarted","Data":"93fb5000b948d1d0abae53d593dd39c9b9aa36c2ced09787ea553b7f96ead3d2"} Feb 25 07:00:07 crc kubenswrapper[4978]: I0225 07:00:07.228319 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:07 crc kubenswrapper[4978]: I0225 07:00:07.230485 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533380-mbpf9" event={"ID":"87f4f3a6-8318-45cd-8942-b3a58210c930","Type":"ContainerStarted","Data":"1ceed6f5add84033ac8410f88f984ec20df27e6503464fac02cb4408959b2fd5"} Feb 25 07:00:07 crc kubenswrapper[4978]: I0225 07:00:07.251279 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-vjh9q" podStartSLOduration=1.83001933 podStartE2EDuration="7.251255682s" podCreationTimestamp="2026-02-25 07:00:00 +0000 UTC" firstStartedPulling="2026-02-25 07:00:01.094879583 +0000 UTC m=+894.534136062" lastFinishedPulling="2026-02-25 07:00:06.516115965 +0000 UTC m=+899.955372414" observedRunningTime="2026-02-25 07:00:07.249121515 +0000 UTC m=+900.688378024" watchObservedRunningTime="2026-02-25 07:00:07.251255682 +0000 UTC m=+900.690512151" Feb 25 07:00:07 crc kubenswrapper[4978]: I0225 07:00:07.269392 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533380-mbpf9" podStartSLOduration=1.5834046960000001 podStartE2EDuration="7.269339017s" podCreationTimestamp="2026-02-25 07:00:00 +0000 UTC" firstStartedPulling="2026-02-25 07:00:01.064247501 +0000 UTC m=+894.503503970" lastFinishedPulling="2026-02-25 07:00:06.750181792 +0000 UTC m=+900.189438291" observedRunningTime="2026-02-25 07:00:07.267090934 +0000 UTC m=+900.706347413" watchObservedRunningTime="2026-02-25 07:00:07.269339017 +0000 UTC m=+900.708595496" Feb 25 07:00:08 crc kubenswrapper[4978]: I0225 07:00:08.241784 4978 generic.go:334] "Generic (PLEG): container finished" podID="87f4f3a6-8318-45cd-8942-b3a58210c930" containerID="1ceed6f5add84033ac8410f88f984ec20df27e6503464fac02cb4408959b2fd5" exitCode=0 Feb 25 07:00:08 crc kubenswrapper[4978]: I0225 07:00:08.241845 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533380-mbpf9" event={"ID":"87f4f3a6-8318-45cd-8942-b3a58210c930","Type":"ContainerDied","Data":"1ceed6f5add84033ac8410f88f984ec20df27e6503464fac02cb4408959b2fd5"} Feb 25 07:00:09 crc kubenswrapper[4978]: I0225 07:00:09.639944 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533380-mbpf9" Feb 25 07:00:09 crc kubenswrapper[4978]: I0225 07:00:09.723889 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kprz6\" (UniqueName: \"kubernetes.io/projected/87f4f3a6-8318-45cd-8942-b3a58210c930-kube-api-access-kprz6\") pod \"87f4f3a6-8318-45cd-8942-b3a58210c930\" (UID: \"87f4f3a6-8318-45cd-8942-b3a58210c930\") " Feb 25 07:00:09 crc kubenswrapper[4978]: I0225 07:00:09.732014 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87f4f3a6-8318-45cd-8942-b3a58210c930-kube-api-access-kprz6" (OuterVolumeSpecName: "kube-api-access-kprz6") pod "87f4f3a6-8318-45cd-8942-b3a58210c930" (UID: "87f4f3a6-8318-45cd-8942-b3a58210c930"). InnerVolumeSpecName "kube-api-access-kprz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:00:09 crc kubenswrapper[4978]: I0225 07:00:09.825713 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kprz6\" (UniqueName: \"kubernetes.io/projected/87f4f3a6-8318-45cd-8942-b3a58210c930-kube-api-access-kprz6\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:10 crc kubenswrapper[4978]: I0225 07:00:10.259969 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533380-mbpf9" event={"ID":"87f4f3a6-8318-45cd-8942-b3a58210c930","Type":"ContainerDied","Data":"4076e4f4aac5f52931af7640d625bce33ad3b5152795f8e81a2b5cd35640c244"} Feb 25 07:00:10 crc kubenswrapper[4978]: I0225 07:00:10.260398 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4076e4f4aac5f52931af7640d625bce33ad3b5152795f8e81a2b5cd35640c244" Feb 25 07:00:10 crc kubenswrapper[4978]: I0225 07:00:10.260046 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533380-mbpf9" Feb 25 07:00:10 crc kubenswrapper[4978]: I0225 07:00:10.705889 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533374-fdvtk"] Feb 25 07:00:10 crc kubenswrapper[4978]: I0225 07:00:10.709907 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533374-fdvtk"] Feb 25 07:00:11 crc kubenswrapper[4978]: I0225 07:00:11.080321 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-vjh9q" Feb 25 07:00:11 crc kubenswrapper[4978]: I0225 07:00:11.269331 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" event={"ID":"14d5cc0e-ed1f-4894-afde-ae5164197176","Type":"ContainerStarted","Data":"9494871be357dd7cf6b5a89e73bb72aafd54dac76643496399c840866f0711f0"} Feb 25 07:00:11 crc kubenswrapper[4978]: I0225 07:00:11.340902 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e559d93f-7e03-4ee7-850c-7a305a76e3cd" path="/var/lib/kubelet/pods/e559d93f-7e03-4ee7-850c-7a305a76e3cd/volumes" Feb 25 07:00:11 crc kubenswrapper[4978]: I0225 07:00:11.502930 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:11 crc kubenswrapper[4978]: I0225 07:00:11.504203 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:11 crc kubenswrapper[4978]: I0225 07:00:11.515835 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:11 crc kubenswrapper[4978]: I0225 07:00:11.544154 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-b7qb4" podStartSLOduration=1.984802743 podStartE2EDuration="11.544124767s" podCreationTimestamp="2026-02-25 07:00:00 +0000 UTC" firstStartedPulling="2026-02-25 07:00:01.327443343 +0000 UTC m=+894.766699792" lastFinishedPulling="2026-02-25 07:00:10.886765317 +0000 UTC m=+904.326021816" observedRunningTime="2026-02-25 07:00:11.294540127 +0000 UTC m=+904.733796596" watchObservedRunningTime="2026-02-25 07:00:11.544124767 +0000 UTC m=+904.983381266" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.285397 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-574869f-nlz7w" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.356492 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-r5cx2"] Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.539743 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f9zf4"] Feb 25 07:00:12 crc kubenswrapper[4978]: E0225 07:00:12.540662 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerName="registry-server" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.540756 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerName="registry-server" Feb 25 07:00:12 crc kubenswrapper[4978]: E0225 07:00:12.540832 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87f4f3a6-8318-45cd-8942-b3a58210c930" containerName="oc" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.540932 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="87f4f3a6-8318-45cd-8942-b3a58210c930" containerName="oc" Feb 25 07:00:12 crc kubenswrapper[4978]: E0225 07:00:12.541042 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2bef0e-1e45-4eb3-9525-3328e351fa3d" containerName="collect-profiles" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.541141 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2bef0e-1e45-4eb3-9525-3328e351fa3d" containerName="collect-profiles" Feb 25 07:00:12 crc kubenswrapper[4978]: E0225 07:00:12.541243 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerName="extract-utilities" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.541333 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerName="extract-utilities" Feb 25 07:00:12 crc kubenswrapper[4978]: E0225 07:00:12.541456 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerName="extract-content" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.541568 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerName="extract-content" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.541788 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a1c35d-75dc-4f68-b5a1-bc40cb4555db" containerName="registry-server" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.541879 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="87f4f3a6-8318-45cd-8942-b3a58210c930" containerName="oc" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.541954 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2bef0e-1e45-4eb3-9525-3328e351fa3d" containerName="collect-profiles" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.543003 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.551307 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9zf4"] Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.623138 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fx85\" (UniqueName: \"kubernetes.io/projected/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-kube-api-access-4fx85\") pod \"redhat-marketplace-f9zf4\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.623218 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-catalog-content\") pod \"redhat-marketplace-f9zf4\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.623272 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-utilities\") pod \"redhat-marketplace-f9zf4\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.724534 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-utilities\") pod \"redhat-marketplace-f9zf4\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.724730 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fx85\" (UniqueName: \"kubernetes.io/projected/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-kube-api-access-4fx85\") pod \"redhat-marketplace-f9zf4\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.724811 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-catalog-content\") pod \"redhat-marketplace-f9zf4\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.725292 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-catalog-content\") pod \"redhat-marketplace-f9zf4\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.725299 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-utilities\") pod \"redhat-marketplace-f9zf4\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.752560 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fx85\" (UniqueName: \"kubernetes.io/projected/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-kube-api-access-4fx85\") pod \"redhat-marketplace-f9zf4\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:12 crc kubenswrapper[4978]: I0225 07:00:12.858211 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:13 crc kubenswrapper[4978]: I0225 07:00:13.159833 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9zf4"] Feb 25 07:00:13 crc kubenswrapper[4978]: W0225 07:00:13.166582 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode34d92ef_42a0_4519_b679_f0ad3bb0a60c.slice/crio-822e34f3bbc48a2c439c88f6f5d2e681fba41cbea593c9a3a94cc698612ad794 WatchSource:0}: Error finding container 822e34f3bbc48a2c439c88f6f5d2e681fba41cbea593c9a3a94cc698612ad794: Status 404 returned error can't find the container with id 822e34f3bbc48a2c439c88f6f5d2e681fba41cbea593c9a3a94cc698612ad794 Feb 25 07:00:13 crc kubenswrapper[4978]: I0225 07:00:13.285486 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9zf4" event={"ID":"e34d92ef-42a0-4519-b679-f0ad3bb0a60c","Type":"ContainerStarted","Data":"822e34f3bbc48a2c439c88f6f5d2e681fba41cbea593c9a3a94cc698612ad794"} Feb 25 07:00:14 crc kubenswrapper[4978]: I0225 07:00:14.297661 4978 generic.go:334] "Generic (PLEG): container finished" podID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerID="ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b" exitCode=0 Feb 25 07:00:14 crc kubenswrapper[4978]: I0225 07:00:14.297882 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9zf4" event={"ID":"e34d92ef-42a0-4519-b679-f0ad3bb0a60c","Type":"ContainerDied","Data":"ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b"} Feb 25 07:00:14 crc kubenswrapper[4978]: I0225 07:00:14.303197 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" event={"ID":"81d07653-9826-45dd-a0fd-68a2bd37b8e2","Type":"ContainerStarted","Data":"ae99cf4b75e01d6d85265e965196de92f62dd652ecf395614c4d60aa60f6dff5"} Feb 25 07:00:14 crc kubenswrapper[4978]: I0225 07:00:14.370910 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" podStartSLOduration=1.956682662 podStartE2EDuration="14.370884658s" podCreationTimestamp="2026-02-25 07:00:00 +0000 UTC" firstStartedPulling="2026-02-25 07:00:01.280605057 +0000 UTC m=+894.719861516" lastFinishedPulling="2026-02-25 07:00:13.694807013 +0000 UTC m=+907.134063512" observedRunningTime="2026-02-25 07:00:14.364313979 +0000 UTC m=+907.803570468" watchObservedRunningTime="2026-02-25 07:00:14.370884658 +0000 UTC m=+907.810141127" Feb 25 07:00:14 crc kubenswrapper[4978]: I0225 07:00:14.640453 4978 scope.go:117] "RemoveContainer" containerID="929f4573b7b94c9d2dd4dbe58627b2bd0a3d624d3974ea2369e01a289d5ba5b2" Feb 25 07:00:15 crc kubenswrapper[4978]: I0225 07:00:15.315904 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd" event={"ID":"31fedbf4-35ef-4440-9039-341b3a0ece67","Type":"ContainerStarted","Data":"40ec3d0701118dec61d2b74b2eaaa487761c4143fa5c17250a6fdf5f5b6e2146"} Feb 25 07:00:15 crc kubenswrapper[4978]: I0225 07:00:15.320076 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9zf4" event={"ID":"e34d92ef-42a0-4519-b679-f0ad3bb0a60c","Type":"ContainerStarted","Data":"d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17"} Feb 25 07:00:15 crc kubenswrapper[4978]: I0225 07:00:15.320272 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" Feb 25 07:00:16 crc kubenswrapper[4978]: I0225 07:00:16.326278 4978 generic.go:334] "Generic (PLEG): container finished" podID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerID="d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17" exitCode=0 Feb 25 07:00:16 crc kubenswrapper[4978]: I0225 07:00:16.326337 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9zf4" event={"ID":"e34d92ef-42a0-4519-b679-f0ad3bb0a60c","Type":"ContainerDied","Data":"d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17"} Feb 25 07:00:17 crc kubenswrapper[4978]: I0225 07:00:17.352958 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd" event={"ID":"31fedbf4-35ef-4440-9039-341b3a0ece67","Type":"ContainerStarted","Data":"f30e423dd399041088b178a74f5bd6045c7b2e8cd416d048fc371da8b6fd4036"} Feb 25 07:00:17 crc kubenswrapper[4978]: I0225 07:00:17.389875 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-m8ksd" podStartSLOduration=1.811493743 podStartE2EDuration="17.389822417s" podCreationTimestamp="2026-02-25 07:00:00 +0000 UTC" firstStartedPulling="2026-02-25 07:00:01.257020438 +0000 UTC m=+894.696276897" lastFinishedPulling="2026-02-25 07:00:16.835349112 +0000 UTC m=+910.274605571" observedRunningTime="2026-02-25 07:00:17.384438447 +0000 UTC m=+910.823694946" watchObservedRunningTime="2026-02-25 07:00:17.389822417 +0000 UTC m=+910.829078916" Feb 25 07:00:18 crc kubenswrapper[4978]: I0225 07:00:18.361863 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9zf4" event={"ID":"e34d92ef-42a0-4519-b679-f0ad3bb0a60c","Type":"ContainerStarted","Data":"441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2"} Feb 25 07:00:18 crc kubenswrapper[4978]: I0225 07:00:18.393216 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f9zf4" podStartSLOduration=3.218087021 podStartE2EDuration="6.393198847s" podCreationTimestamp="2026-02-25 07:00:12 +0000 UTC" firstStartedPulling="2026-02-25 07:00:14.300416641 +0000 UTC m=+907.739673130" lastFinishedPulling="2026-02-25 07:00:17.475528467 +0000 UTC m=+910.914784956" observedRunningTime="2026-02-25 07:00:18.38793639 +0000 UTC m=+911.827192859" watchObservedRunningTime="2026-02-25 07:00:18.393198847 +0000 UTC m=+911.832455316" Feb 25 07:00:22 crc kubenswrapper[4978]: I0225 07:00:22.859031 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:22 crc kubenswrapper[4978]: I0225 07:00:22.859824 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:22 crc kubenswrapper[4978]: I0225 07:00:22.930406 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:23 crc kubenswrapper[4978]: I0225 07:00:23.470766 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:23 crc kubenswrapper[4978]: I0225 07:00:23.539873 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9zf4"] Feb 25 07:00:25 crc kubenswrapper[4978]: I0225 07:00:25.415638 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f9zf4" podUID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerName="registry-server" containerID="cri-o://441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2" gracePeriod=2 Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.376113 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.426440 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fx85\" (UniqueName: \"kubernetes.io/projected/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-kube-api-access-4fx85\") pod \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.426506 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-catalog-content\") pod \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.426591 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-utilities\") pod \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\" (UID: \"e34d92ef-42a0-4519-b679-f0ad3bb0a60c\") " Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.427633 4978 generic.go:334] "Generic (PLEG): container finished" podID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerID="441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2" exitCode=0 Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.427691 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9zf4" event={"ID":"e34d92ef-42a0-4519-b679-f0ad3bb0a60c","Type":"ContainerDied","Data":"441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2"} Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.427737 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9zf4" event={"ID":"e34d92ef-42a0-4519-b679-f0ad3bb0a60c","Type":"ContainerDied","Data":"822e34f3bbc48a2c439c88f6f5d2e681fba41cbea593c9a3a94cc698612ad794"} Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.427746 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-utilities" (OuterVolumeSpecName: "utilities") pod "e34d92ef-42a0-4519-b679-f0ad3bb0a60c" (UID: "e34d92ef-42a0-4519-b679-f0ad3bb0a60c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.427769 4978 scope.go:117] "RemoveContainer" containerID="441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.427868 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9zf4" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.436339 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-kube-api-access-4fx85" (OuterVolumeSpecName: "kube-api-access-4fx85") pod "e34d92ef-42a0-4519-b679-f0ad3bb0a60c" (UID: "e34d92ef-42a0-4519-b679-f0ad3bb0a60c"). InnerVolumeSpecName "kube-api-access-4fx85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.456159 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e34d92ef-42a0-4519-b679-f0ad3bb0a60c" (UID: "e34d92ef-42a0-4519-b679-f0ad3bb0a60c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.456883 4978 scope.go:117] "RemoveContainer" containerID="d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.480747 4978 scope.go:117] "RemoveContainer" containerID="ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.503306 4978 scope.go:117] "RemoveContainer" containerID="441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2" Feb 25 07:00:26 crc kubenswrapper[4978]: E0225 07:00:26.503822 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2\": container with ID starting with 441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2 not found: ID does not exist" containerID="441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.503892 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2"} err="failed to get container status \"441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2\": rpc error: code = NotFound desc = could not find container \"441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2\": container with ID starting with 441b3fd965dca4207b7b85c276db5e70fd4ffd600453f8be571cc87aa258a1c2 not found: ID does not exist" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.503933 4978 scope.go:117] "RemoveContainer" containerID="d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17" Feb 25 07:00:26 crc kubenswrapper[4978]: E0225 07:00:26.504436 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17\": container with ID starting with d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17 not found: ID does not exist" containerID="d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.504475 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17"} err="failed to get container status \"d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17\": rpc error: code = NotFound desc = could not find container \"d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17\": container with ID starting with d5836be00e0df3459cae1c3c623d2603f43a3c19d3b463cf0578966912643c17 not found: ID does not exist" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.504504 4978 scope.go:117] "RemoveContainer" containerID="ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b" Feb 25 07:00:26 crc kubenswrapper[4978]: E0225 07:00:26.504967 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b\": container with ID starting with ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b not found: ID does not exist" containerID="ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.505015 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b"} err="failed to get container status \"ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b\": rpc error: code = NotFound desc = could not find container \"ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b\": container with ID starting with ef0e4563348a54f6504c0ff818347db8e9ea0f4ee15801706d3e22f289a5980b not found: ID does not exist" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.528536 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fx85\" (UniqueName: \"kubernetes.io/projected/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-kube-api-access-4fx85\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.528563 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.528575 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e34d92ef-42a0-4519-b679-f0ad3bb0a60c-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.776892 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9zf4"] Feb 25 07:00:26 crc kubenswrapper[4978]: I0225 07:00:26.787875 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9zf4"] Feb 25 07:00:27 crc kubenswrapper[4978]: I0225 07:00:27.342705 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" path="/var/lib/kubelet/pods/e34d92ef-42a0-4519-b679-f0ad3bb0a60c/volumes" Feb 25 07:00:31 crc kubenswrapper[4978]: I0225 07:00:31.040951 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-hlrw5" Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.425962 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-r5cx2" podUID="b51a2671-2dea-418e-8f99-512e4ba4e34c" containerName="console" containerID="cri-o://b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f" gracePeriod=15 Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.852184 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-r5cx2_b51a2671-2dea-418e-8f99-512e4ba4e34c/console/0.log" Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.852416 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.925242 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-serving-cert\") pod \"b51a2671-2dea-418e-8f99-512e4ba4e34c\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.925852 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-trusted-ca-bundle\") pod \"b51a2671-2dea-418e-8f99-512e4ba4e34c\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.925879 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-oauth-config\") pod \"b51a2671-2dea-418e-8f99-512e4ba4e34c\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.925941 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-service-ca\") pod \"b51a2671-2dea-418e-8f99-512e4ba4e34c\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.925969 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4m8t\" (UniqueName: \"kubernetes.io/projected/b51a2671-2dea-418e-8f99-512e4ba4e34c-kube-api-access-c4m8t\") pod \"b51a2671-2dea-418e-8f99-512e4ba4e34c\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.926009 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-oauth-serving-cert\") pod \"b51a2671-2dea-418e-8f99-512e4ba4e34c\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.926387 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-config\") pod \"b51a2671-2dea-418e-8f99-512e4ba4e34c\" (UID: \"b51a2671-2dea-418e-8f99-512e4ba4e34c\") " Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.926958 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-service-ca" (OuterVolumeSpecName: "service-ca") pod "b51a2671-2dea-418e-8f99-512e4ba4e34c" (UID: "b51a2671-2dea-418e-8f99-512e4ba4e34c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.927173 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-config" (OuterVolumeSpecName: "console-config") pod "b51a2671-2dea-418e-8f99-512e4ba4e34c" (UID: "b51a2671-2dea-418e-8f99-512e4ba4e34c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.927441 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b51a2671-2dea-418e-8f99-512e4ba4e34c" (UID: "b51a2671-2dea-418e-8f99-512e4ba4e34c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.933008 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b51a2671-2dea-418e-8f99-512e4ba4e34c" (UID: "b51a2671-2dea-418e-8f99-512e4ba4e34c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.934000 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b51a2671-2dea-418e-8f99-512e4ba4e34c" (UID: "b51a2671-2dea-418e-8f99-512e4ba4e34c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.939522 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b51a2671-2dea-418e-8f99-512e4ba4e34c-kube-api-access-c4m8t" (OuterVolumeSpecName: "kube-api-access-c4m8t") pod "b51a2671-2dea-418e-8f99-512e4ba4e34c" (UID: "b51a2671-2dea-418e-8f99-512e4ba4e34c"). InnerVolumeSpecName "kube-api-access-c4m8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:00:37 crc kubenswrapper[4978]: I0225 07:00:37.943177 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b51a2671-2dea-418e-8f99-512e4ba4e34c" (UID: "b51a2671-2dea-418e-8f99-512e4ba4e34c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.028255 4978 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.028295 4978 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.028306 4978 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.028315 4978 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-service-ca\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.028326 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4m8t\" (UniqueName: \"kubernetes.io/projected/b51a2671-2dea-418e-8f99-512e4ba4e34c-kube-api-access-c4m8t\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.028338 4978 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.028346 4978 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b51a2671-2dea-418e-8f99-512e4ba4e34c-console-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.514566 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-r5cx2_b51a2671-2dea-418e-8f99-512e4ba4e34c/console/0.log" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.514646 4978 generic.go:334] "Generic (PLEG): container finished" podID="b51a2671-2dea-418e-8f99-512e4ba4e34c" containerID="b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f" exitCode=2 Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.514690 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r5cx2" event={"ID":"b51a2671-2dea-418e-8f99-512e4ba4e34c","Type":"ContainerDied","Data":"b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f"} Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.514727 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-r5cx2" event={"ID":"b51a2671-2dea-418e-8f99-512e4ba4e34c","Type":"ContainerDied","Data":"d32c7a64124df0faedcc97b91c960c3bcc579e6f30b205831e9757413888ae7b"} Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.514755 4978 scope.go:117] "RemoveContainer" containerID="b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.514929 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-r5cx2" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.544957 4978 scope.go:117] "RemoveContainer" containerID="b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f" Feb 25 07:00:38 crc kubenswrapper[4978]: E0225 07:00:38.545798 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f\": container with ID starting with b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f not found: ID does not exist" containerID="b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.545875 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f"} err="failed to get container status \"b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f\": rpc error: code = NotFound desc = could not find container \"b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f\": container with ID starting with b87feadcede5483bd302846b45f5d28212eaa6b99ce1581a882effc745873d3f not found: ID does not exist" Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.570878 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-r5cx2"] Feb 25 07:00:38 crc kubenswrapper[4978]: I0225 07:00:38.576133 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-r5cx2"] Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.337983 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b51a2671-2dea-418e-8f99-512e4ba4e34c" path="/var/lib/kubelet/pods/b51a2671-2dea-418e-8f99-512e4ba4e34c/volumes" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.758424 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8cmwz"] Feb 25 07:00:39 crc kubenswrapper[4978]: E0225 07:00:39.758989 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerName="extract-utilities" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.759005 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerName="extract-utilities" Feb 25 07:00:39 crc kubenswrapper[4978]: E0225 07:00:39.759032 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51a2671-2dea-418e-8f99-512e4ba4e34c" containerName="console" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.759039 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51a2671-2dea-418e-8f99-512e4ba4e34c" containerName="console" Feb 25 07:00:39 crc kubenswrapper[4978]: E0225 07:00:39.759052 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerName="extract-content" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.759061 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerName="extract-content" Feb 25 07:00:39 crc kubenswrapper[4978]: E0225 07:00:39.759072 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerName="registry-server" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.759080 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerName="registry-server" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.759197 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51a2671-2dea-418e-8f99-512e4ba4e34c" containerName="console" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.759208 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e34d92ef-42a0-4519-b679-f0ad3bb0a60c" containerName="registry-server" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.760130 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.766287 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8cmwz"] Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.855232 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjt9v\" (UniqueName: \"kubernetes.io/projected/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-kube-api-access-pjt9v\") pod \"certified-operators-8cmwz\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.855281 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-utilities\") pod \"certified-operators-8cmwz\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.855303 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-catalog-content\") pod \"certified-operators-8cmwz\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.955941 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjt9v\" (UniqueName: \"kubernetes.io/projected/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-kube-api-access-pjt9v\") pod \"certified-operators-8cmwz\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.955989 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-utilities\") pod \"certified-operators-8cmwz\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.956012 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-catalog-content\") pod \"certified-operators-8cmwz\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.956457 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-catalog-content\") pod \"certified-operators-8cmwz\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.956665 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-utilities\") pod \"certified-operators-8cmwz\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:39 crc kubenswrapper[4978]: I0225 07:00:39.982489 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjt9v\" (UniqueName: \"kubernetes.io/projected/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-kube-api-access-pjt9v\") pod \"certified-operators-8cmwz\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:40 crc kubenswrapper[4978]: I0225 07:00:40.088429 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:40 crc kubenswrapper[4978]: I0225 07:00:40.409956 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8cmwz"] Feb 25 07:00:40 crc kubenswrapper[4978]: I0225 07:00:40.552646 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cmwz" event={"ID":"f5ec50dd-fb4f-46db-84cf-9da25e3760bb","Type":"ContainerStarted","Data":"7369152be492e22841762ef868ae5a675a10158b36a641d55f1ff961a152c338"} Feb 25 07:00:41 crc kubenswrapper[4978]: I0225 07:00:41.566571 4978 generic.go:334] "Generic (PLEG): container finished" podID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerID="a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe" exitCode=0 Feb 25 07:00:41 crc kubenswrapper[4978]: I0225 07:00:41.566986 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cmwz" event={"ID":"f5ec50dd-fb4f-46db-84cf-9da25e3760bb","Type":"ContainerDied","Data":"a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe"} Feb 25 07:00:42 crc kubenswrapper[4978]: I0225 07:00:42.580091 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cmwz" event={"ID":"f5ec50dd-fb4f-46db-84cf-9da25e3760bb","Type":"ContainerStarted","Data":"bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21"} Feb 25 07:00:43 crc kubenswrapper[4978]: I0225 07:00:43.589449 4978 generic.go:334] "Generic (PLEG): container finished" podID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerID="bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21" exitCode=0 Feb 25 07:00:43 crc kubenswrapper[4978]: I0225 07:00:43.589557 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cmwz" event={"ID":"f5ec50dd-fb4f-46db-84cf-9da25e3760bb","Type":"ContainerDied","Data":"bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21"} Feb 25 07:00:44 crc kubenswrapper[4978]: I0225 07:00:44.606515 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cmwz" event={"ID":"f5ec50dd-fb4f-46db-84cf-9da25e3760bb","Type":"ContainerStarted","Data":"4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34"} Feb 25 07:00:44 crc kubenswrapper[4978]: I0225 07:00:44.648429 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8cmwz" podStartSLOduration=3.014311732 podStartE2EDuration="5.648400459s" podCreationTimestamp="2026-02-25 07:00:39 +0000 UTC" firstStartedPulling="2026-02-25 07:00:41.569616831 +0000 UTC m=+935.008873330" lastFinishedPulling="2026-02-25 07:00:44.203705598 +0000 UTC m=+937.642962057" observedRunningTime="2026-02-25 07:00:44.645049712 +0000 UTC m=+938.084306251" watchObservedRunningTime="2026-02-25 07:00:44.648400459 +0000 UTC m=+938.087656968" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.520304 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-45974"] Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.522629 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.539056 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-45974"] Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.540116 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.540197 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.571830 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4pw9\" (UniqueName: \"kubernetes.io/projected/1b792399-8138-43b9-a6a8-75eeec8b567e-kube-api-access-b4pw9\") pod \"community-operators-45974\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.571928 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-utilities\") pod \"community-operators-45974\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.572005 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-catalog-content\") pod \"community-operators-45974\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.673671 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-catalog-content\") pod \"community-operators-45974\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.673762 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4pw9\" (UniqueName: \"kubernetes.io/projected/1b792399-8138-43b9-a6a8-75eeec8b567e-kube-api-access-b4pw9\") pod \"community-operators-45974\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.673864 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-utilities\") pod \"community-operators-45974\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.674682 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-utilities\") pod \"community-operators-45974\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.674700 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-catalog-content\") pod \"community-operators-45974\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.719536 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4pw9\" (UniqueName: \"kubernetes.io/projected/1b792399-8138-43b9-a6a8-75eeec8b567e-kube-api-access-b4pw9\") pod \"community-operators-45974\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:46 crc kubenswrapper[4978]: I0225 07:00:46.881150 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:47 crc kubenswrapper[4978]: I0225 07:00:47.198630 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-45974"] Feb 25 07:00:47 crc kubenswrapper[4978]: I0225 07:00:47.628860 4978 generic.go:334] "Generic (PLEG): container finished" podID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerID="8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82" exitCode=0 Feb 25 07:00:47 crc kubenswrapper[4978]: I0225 07:00:47.628945 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45974" event={"ID":"1b792399-8138-43b9-a6a8-75eeec8b567e","Type":"ContainerDied","Data":"8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82"} Feb 25 07:00:47 crc kubenswrapper[4978]: I0225 07:00:47.629214 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45974" event={"ID":"1b792399-8138-43b9-a6a8-75eeec8b567e","Type":"ContainerStarted","Data":"b813e5a747ce2adfba910ebf18b3c69749da3e7d21512f695d418c3b10061fe6"} Feb 25 07:00:48 crc kubenswrapper[4978]: I0225 07:00:48.639197 4978 generic.go:334] "Generic (PLEG): container finished" podID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerID="b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c" exitCode=0 Feb 25 07:00:48 crc kubenswrapper[4978]: I0225 07:00:48.639245 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45974" event={"ID":"1b792399-8138-43b9-a6a8-75eeec8b567e","Type":"ContainerDied","Data":"b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c"} Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.178964 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt"] Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.202139 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.204719 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.213121 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt"] Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.319254 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.319343 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.319426 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvbvf\" (UniqueName: \"kubernetes.io/projected/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-kube-api-access-jvbvf\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.420436 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.420540 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.420585 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvbvf\" (UniqueName: \"kubernetes.io/projected/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-kube-api-access-jvbvf\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.421100 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.421201 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.453677 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvbvf\" (UniqueName: \"kubernetes.io/projected/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-kube-api-access-jvbvf\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.519266 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.656072 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45974" event={"ID":"1b792399-8138-43b9-a6a8-75eeec8b567e","Type":"ContainerStarted","Data":"a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c"} Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.686805 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-45974" podStartSLOduration=2.27684945 podStartE2EDuration="3.686786062s" podCreationTimestamp="2026-02-25 07:00:46 +0000 UTC" firstStartedPulling="2026-02-25 07:00:47.631024477 +0000 UTC m=+941.070280976" lastFinishedPulling="2026-02-25 07:00:49.040961119 +0000 UTC m=+942.480217588" observedRunningTime="2026-02-25 07:00:49.67914819 +0000 UTC m=+943.118404659" watchObservedRunningTime="2026-02-25 07:00:49.686786062 +0000 UTC m=+943.126042531" Feb 25 07:00:49 crc kubenswrapper[4978]: I0225 07:00:49.778677 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt"] Feb 25 07:00:50 crc kubenswrapper[4978]: I0225 07:00:50.088802 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:50 crc kubenswrapper[4978]: I0225 07:00:50.089721 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:50 crc kubenswrapper[4978]: I0225 07:00:50.137140 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:50 crc kubenswrapper[4978]: I0225 07:00:50.667465 4978 generic.go:334] "Generic (PLEG): container finished" podID="0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" containerID="29b9254f41ae2c5fbbdfc2b8c10f83ea4ffa51cedfc13c16602987b2850dbefb" exitCode=0 Feb 25 07:00:50 crc kubenswrapper[4978]: I0225 07:00:50.667603 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" event={"ID":"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4","Type":"ContainerDied","Data":"29b9254f41ae2c5fbbdfc2b8c10f83ea4ffa51cedfc13c16602987b2850dbefb"} Feb 25 07:00:50 crc kubenswrapper[4978]: I0225 07:00:50.667903 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" event={"ID":"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4","Type":"ContainerStarted","Data":"215832df65c3bea86e0d69a6770d25c9e9822e5475262ef01acbb954adb50847"} Feb 25 07:00:50 crc kubenswrapper[4978]: I0225 07:00:50.760647 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:52 crc kubenswrapper[4978]: I0225 07:00:52.688006 4978 generic.go:334] "Generic (PLEG): container finished" podID="0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" containerID="f6b2e4cc968f6eda4920503a57fee0150d30b72fc65f414c357388093bd47849" exitCode=0 Feb 25 07:00:52 crc kubenswrapper[4978]: I0225 07:00:52.688101 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" event={"ID":"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4","Type":"ContainerDied","Data":"f6b2e4cc968f6eda4920503a57fee0150d30b72fc65f414c357388093bd47849"} Feb 25 07:00:53 crc kubenswrapper[4978]: I0225 07:00:53.698983 4978 generic.go:334] "Generic (PLEG): container finished" podID="0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" containerID="35f7ca106b7ab47fc8949495f056614cf76ae8aa85bbf8e978f5297c90158514" exitCode=0 Feb 25 07:00:53 crc kubenswrapper[4978]: I0225 07:00:53.699108 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" event={"ID":"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4","Type":"ContainerDied","Data":"35f7ca106b7ab47fc8949495f056614cf76ae8aa85bbf8e978f5297c90158514"} Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.027480 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.102894 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-util\") pod \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.102973 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvbvf\" (UniqueName: \"kubernetes.io/projected/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-kube-api-access-jvbvf\") pod \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.103085 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-bundle\") pod \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\" (UID: \"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4\") " Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.104477 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-bundle" (OuterVolumeSpecName: "bundle") pod "0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" (UID: "0c1cb1bf-977e-400b-a0c6-2058d62e1ff4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.114788 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-kube-api-access-jvbvf" (OuterVolumeSpecName: "kube-api-access-jvbvf") pod "0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" (UID: "0c1cb1bf-977e-400b-a0c6-2058d62e1ff4"). InnerVolumeSpecName "kube-api-access-jvbvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.204720 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvbvf\" (UniqueName: \"kubernetes.io/projected/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-kube-api-access-jvbvf\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.204786 4978 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.429588 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-util" (OuterVolumeSpecName: "util") pod "0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" (UID: "0c1cb1bf-977e-400b-a0c6-2058d62e1ff4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.509708 4978 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0c1cb1bf-977e-400b-a0c6-2058d62e1ff4-util\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.714456 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8cmwz"] Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.715043 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8cmwz" podUID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerName="registry-server" containerID="cri-o://4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34" gracePeriod=2 Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.722218 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" event={"ID":"0c1cb1bf-977e-400b-a0c6-2058d62e1ff4","Type":"ContainerDied","Data":"215832df65c3bea86e0d69a6770d25c9e9822e5475262ef01acbb954adb50847"} Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.722281 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="215832df65c3bea86e0d69a6770d25c9e9822e5475262ef01acbb954adb50847" Feb 25 07:00:55 crc kubenswrapper[4978]: I0225 07:00:55.722306 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.198858 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.321078 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjt9v\" (UniqueName: \"kubernetes.io/projected/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-kube-api-access-pjt9v\") pod \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.321177 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-utilities\") pod \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.321249 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-catalog-content\") pod \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\" (UID: \"f5ec50dd-fb4f-46db-84cf-9da25e3760bb\") " Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.322641 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-utilities" (OuterVolumeSpecName: "utilities") pod "f5ec50dd-fb4f-46db-84cf-9da25e3760bb" (UID: "f5ec50dd-fb4f-46db-84cf-9da25e3760bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.328932 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-kube-api-access-pjt9v" (OuterVolumeSpecName: "kube-api-access-pjt9v") pod "f5ec50dd-fb4f-46db-84cf-9da25e3760bb" (UID: "f5ec50dd-fb4f-46db-84cf-9da25e3760bb"). InnerVolumeSpecName "kube-api-access-pjt9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.369016 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5ec50dd-fb4f-46db-84cf-9da25e3760bb" (UID: "f5ec50dd-fb4f-46db-84cf-9da25e3760bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.422501 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjt9v\" (UniqueName: \"kubernetes.io/projected/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-kube-api-access-pjt9v\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.422898 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.422979 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ec50dd-fb4f-46db-84cf-9da25e3760bb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.735815 4978 generic.go:334] "Generic (PLEG): container finished" podID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerID="4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34" exitCode=0 Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.735905 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cmwz" event={"ID":"f5ec50dd-fb4f-46db-84cf-9da25e3760bb","Type":"ContainerDied","Data":"4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34"} Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.735975 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8cmwz" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.736733 4978 scope.go:117] "RemoveContainer" containerID="4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.737504 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8cmwz" event={"ID":"f5ec50dd-fb4f-46db-84cf-9da25e3760bb","Type":"ContainerDied","Data":"7369152be492e22841762ef868ae5a675a10158b36a641d55f1ff961a152c338"} Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.762824 4978 scope.go:117] "RemoveContainer" containerID="bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.795463 4978 scope.go:117] "RemoveContainer" containerID="a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.810457 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8cmwz"] Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.832900 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8cmwz"] Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.835154 4978 scope.go:117] "RemoveContainer" containerID="4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34" Feb 25 07:00:56 crc kubenswrapper[4978]: E0225 07:00:56.836248 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34\": container with ID starting with 4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34 not found: ID does not exist" containerID="4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.836357 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34"} err="failed to get container status \"4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34\": rpc error: code = NotFound desc = could not find container \"4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34\": container with ID starting with 4e1b7c435f603b86e25a37db2c948394306b6f094941ea297723b1bc03b86c34 not found: ID does not exist" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.836485 4978 scope.go:117] "RemoveContainer" containerID="bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21" Feb 25 07:00:56 crc kubenswrapper[4978]: E0225 07:00:56.837068 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21\": container with ID starting with bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21 not found: ID does not exist" containerID="bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.837133 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21"} err="failed to get container status \"bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21\": rpc error: code = NotFound desc = could not find container \"bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21\": container with ID starting with bbdefe11c18c72f905c89967449a173b1de3ebbe8180e48057e10822ba883f21 not found: ID does not exist" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.837176 4978 scope.go:117] "RemoveContainer" containerID="a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe" Feb 25 07:00:56 crc kubenswrapper[4978]: E0225 07:00:56.838115 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe\": container with ID starting with a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe not found: ID does not exist" containerID="a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.838198 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe"} err="failed to get container status \"a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe\": rpc error: code = NotFound desc = could not find container \"a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe\": container with ID starting with a2ecff3f99e8e1bba7421f86a0e968538985330b79361cfd3ebd090b12139efe not found: ID does not exist" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.882177 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.882258 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:56 crc kubenswrapper[4978]: I0225 07:00:56.936132 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-45974" Feb 25 07:00:57 crc kubenswrapper[4978]: I0225 07:00:57.338439 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" path="/var/lib/kubelet/pods/f5ec50dd-fb4f-46db-84cf-9da25e3760bb/volumes" Feb 25 07:00:57 crc kubenswrapper[4978]: I0225 07:00:57.822130 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-45974" Feb 25 07:01:00 crc kubenswrapper[4978]: I0225 07:01:00.515340 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-45974"] Feb 25 07:01:00 crc kubenswrapper[4978]: I0225 07:01:00.767267 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-45974" podUID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerName="registry-server" containerID="cri-o://a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c" gracePeriod=2 Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.236107 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-45974" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.299362 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4pw9\" (UniqueName: \"kubernetes.io/projected/1b792399-8138-43b9-a6a8-75eeec8b567e-kube-api-access-b4pw9\") pod \"1b792399-8138-43b9-a6a8-75eeec8b567e\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.299843 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-utilities\") pod \"1b792399-8138-43b9-a6a8-75eeec8b567e\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.299961 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-catalog-content\") pod \"1b792399-8138-43b9-a6a8-75eeec8b567e\" (UID: \"1b792399-8138-43b9-a6a8-75eeec8b567e\") " Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.301067 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-utilities" (OuterVolumeSpecName: "utilities") pod "1b792399-8138-43b9-a6a8-75eeec8b567e" (UID: "1b792399-8138-43b9-a6a8-75eeec8b567e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.305479 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b792399-8138-43b9-a6a8-75eeec8b567e-kube-api-access-b4pw9" (OuterVolumeSpecName: "kube-api-access-b4pw9") pod "1b792399-8138-43b9-a6a8-75eeec8b567e" (UID: "1b792399-8138-43b9-a6a8-75eeec8b567e"). InnerVolumeSpecName "kube-api-access-b4pw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.379000 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b792399-8138-43b9-a6a8-75eeec8b567e" (UID: "1b792399-8138-43b9-a6a8-75eeec8b567e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.400999 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.401043 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b792399-8138-43b9-a6a8-75eeec8b567e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.401060 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4pw9\" (UniqueName: \"kubernetes.io/projected/1b792399-8138-43b9-a6a8-75eeec8b567e-kube-api-access-b4pw9\") on node \"crc\" DevicePath \"\"" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.775453 4978 generic.go:334] "Generic (PLEG): container finished" podID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerID="a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c" exitCode=0 Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.775494 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45974" event={"ID":"1b792399-8138-43b9-a6a8-75eeec8b567e","Type":"ContainerDied","Data":"a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c"} Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.775518 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-45974" event={"ID":"1b792399-8138-43b9-a6a8-75eeec8b567e","Type":"ContainerDied","Data":"b813e5a747ce2adfba910ebf18b3c69749da3e7d21512f695d418c3b10061fe6"} Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.775536 4978 scope.go:117] "RemoveContainer" containerID="a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.775558 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-45974" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.803575 4978 scope.go:117] "RemoveContainer" containerID="b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.819773 4978 scope.go:117] "RemoveContainer" containerID="8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.825515 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-45974"] Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.832291 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-45974"] Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.843123 4978 scope.go:117] "RemoveContainer" containerID="a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c" Feb 25 07:01:01 crc kubenswrapper[4978]: E0225 07:01:01.843665 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c\": container with ID starting with a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c not found: ID does not exist" containerID="a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.843706 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c"} err="failed to get container status \"a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c\": rpc error: code = NotFound desc = could not find container \"a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c\": container with ID starting with a4329a7b8c9d12bf148ffdbe42f6ea1d834d1122ea42f6a8a70acfbf7169104c not found: ID does not exist" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.843736 4978 scope.go:117] "RemoveContainer" containerID="b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c" Feb 25 07:01:01 crc kubenswrapper[4978]: E0225 07:01:01.844052 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c\": container with ID starting with b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c not found: ID does not exist" containerID="b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.844088 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c"} err="failed to get container status \"b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c\": rpc error: code = NotFound desc = could not find container \"b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c\": container with ID starting with b63162c1ca05385df3bc111b139be1804c83e40fe6c6cf0f52af52c6ff6c583c not found: ID does not exist" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.844116 4978 scope.go:117] "RemoveContainer" containerID="8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82" Feb 25 07:01:01 crc kubenswrapper[4978]: E0225 07:01:01.847699 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82\": container with ID starting with 8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82 not found: ID does not exist" containerID="8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82" Feb 25 07:01:01 crc kubenswrapper[4978]: I0225 07:01:01.847732 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82"} err="failed to get container status \"8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82\": rpc error: code = NotFound desc = could not find container \"8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82\": container with ID starting with 8b3d38c603e72c8e6d8f5d9366740e34d51272dcdff1b07f31272652c6978f82 not found: ID does not exist" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.335325 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b792399-8138-43b9-a6a8-75eeec8b567e" path="/var/lib/kubelet/pods/1b792399-8138-43b9-a6a8-75eeec8b567e/volumes" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.734835 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5769477c66-cbbds"] Feb 25 07:01:03 crc kubenswrapper[4978]: E0225 07:01:03.735376 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerName="extract-content" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735389 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerName="extract-content" Feb 25 07:01:03 crc kubenswrapper[4978]: E0225 07:01:03.735402 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" containerName="pull" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735407 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" containerName="pull" Feb 25 07:01:03 crc kubenswrapper[4978]: E0225 07:01:03.735417 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerName="extract-utilities" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735423 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerName="extract-utilities" Feb 25 07:01:03 crc kubenswrapper[4978]: E0225 07:01:03.735434 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" containerName="util" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735439 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" containerName="util" Feb 25 07:01:03 crc kubenswrapper[4978]: E0225 07:01:03.735449 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" containerName="extract" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735456 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" containerName="extract" Feb 25 07:01:03 crc kubenswrapper[4978]: E0225 07:01:03.735468 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerName="registry-server" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735476 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerName="registry-server" Feb 25 07:01:03 crc kubenswrapper[4978]: E0225 07:01:03.735487 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerName="extract-utilities" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735493 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerName="extract-utilities" Feb 25 07:01:03 crc kubenswrapper[4978]: E0225 07:01:03.735500 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerName="extract-content" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735508 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerName="extract-content" Feb 25 07:01:03 crc kubenswrapper[4978]: E0225 07:01:03.735517 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerName="registry-server" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735523 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerName="registry-server" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735609 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b792399-8138-43b9-a6a8-75eeec8b567e" containerName="registry-server" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735621 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1cb1bf-977e-400b-a0c6-2058d62e1ff4" containerName="extract" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.735631 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5ec50dd-fb4f-46db-84cf-9da25e3760bb" containerName="registry-server" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.736040 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.738001 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.738433 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.738496 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.740919 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-tsdr5" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.741388 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.753983 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5769477c66-cbbds"] Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.831639 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2khnv\" (UniqueName: \"kubernetes.io/projected/296657b4-46fe-4780-b077-65d11eeab81d-kube-api-access-2khnv\") pod \"metallb-operator-controller-manager-5769477c66-cbbds\" (UID: \"296657b4-46fe-4780-b077-65d11eeab81d\") " pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.831702 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/296657b4-46fe-4780-b077-65d11eeab81d-webhook-cert\") pod \"metallb-operator-controller-manager-5769477c66-cbbds\" (UID: \"296657b4-46fe-4780-b077-65d11eeab81d\") " pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.831777 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/296657b4-46fe-4780-b077-65d11eeab81d-apiservice-cert\") pod \"metallb-operator-controller-manager-5769477c66-cbbds\" (UID: \"296657b4-46fe-4780-b077-65d11eeab81d\") " pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.933094 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2khnv\" (UniqueName: \"kubernetes.io/projected/296657b4-46fe-4780-b077-65d11eeab81d-kube-api-access-2khnv\") pod \"metallb-operator-controller-manager-5769477c66-cbbds\" (UID: \"296657b4-46fe-4780-b077-65d11eeab81d\") " pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.933188 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/296657b4-46fe-4780-b077-65d11eeab81d-webhook-cert\") pod \"metallb-operator-controller-manager-5769477c66-cbbds\" (UID: \"296657b4-46fe-4780-b077-65d11eeab81d\") " pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.933308 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/296657b4-46fe-4780-b077-65d11eeab81d-apiservice-cert\") pod \"metallb-operator-controller-manager-5769477c66-cbbds\" (UID: \"296657b4-46fe-4780-b077-65d11eeab81d\") " pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.940810 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/296657b4-46fe-4780-b077-65d11eeab81d-apiservice-cert\") pod \"metallb-operator-controller-manager-5769477c66-cbbds\" (UID: \"296657b4-46fe-4780-b077-65d11eeab81d\") " pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.941259 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/296657b4-46fe-4780-b077-65d11eeab81d-webhook-cert\") pod \"metallb-operator-controller-manager-5769477c66-cbbds\" (UID: \"296657b4-46fe-4780-b077-65d11eeab81d\") " pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:03 crc kubenswrapper[4978]: I0225 07:01:03.949797 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2khnv\" (UniqueName: \"kubernetes.io/projected/296657b4-46fe-4780-b077-65d11eeab81d-kube-api-access-2khnv\") pod \"metallb-operator-controller-manager-5769477c66-cbbds\" (UID: \"296657b4-46fe-4780-b077-65d11eeab81d\") " pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.048710 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.275809 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4"] Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.277216 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.281166 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.281474 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.282544 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-f79r8" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.296092 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4"] Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.340926 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4887490-f9e3-4a5f-be93-ba244beaf4e7-webhook-cert\") pod \"metallb-operator-webhook-server-5b647b685b-2jdt4\" (UID: \"c4887490-f9e3-4a5f-be93-ba244beaf4e7\") " pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.340969 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv8t4\" (UniqueName: \"kubernetes.io/projected/c4887490-f9e3-4a5f-be93-ba244beaf4e7-kube-api-access-fv8t4\") pod \"metallb-operator-webhook-server-5b647b685b-2jdt4\" (UID: \"c4887490-f9e3-4a5f-be93-ba244beaf4e7\") " pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.340996 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4887490-f9e3-4a5f-be93-ba244beaf4e7-apiservice-cert\") pod \"metallb-operator-webhook-server-5b647b685b-2jdt4\" (UID: \"c4887490-f9e3-4a5f-be93-ba244beaf4e7\") " pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.441945 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv8t4\" (UniqueName: \"kubernetes.io/projected/c4887490-f9e3-4a5f-be93-ba244beaf4e7-kube-api-access-fv8t4\") pod \"metallb-operator-webhook-server-5b647b685b-2jdt4\" (UID: \"c4887490-f9e3-4a5f-be93-ba244beaf4e7\") " pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.442002 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4887490-f9e3-4a5f-be93-ba244beaf4e7-apiservice-cert\") pod \"metallb-operator-webhook-server-5b647b685b-2jdt4\" (UID: \"c4887490-f9e3-4a5f-be93-ba244beaf4e7\") " pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.442075 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4887490-f9e3-4a5f-be93-ba244beaf4e7-webhook-cert\") pod \"metallb-operator-webhook-server-5b647b685b-2jdt4\" (UID: \"c4887490-f9e3-4a5f-be93-ba244beaf4e7\") " pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.450094 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4887490-f9e3-4a5f-be93-ba244beaf4e7-apiservice-cert\") pod \"metallb-operator-webhook-server-5b647b685b-2jdt4\" (UID: \"c4887490-f9e3-4a5f-be93-ba244beaf4e7\") " pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.461190 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4887490-f9e3-4a5f-be93-ba244beaf4e7-webhook-cert\") pod \"metallb-operator-webhook-server-5b647b685b-2jdt4\" (UID: \"c4887490-f9e3-4a5f-be93-ba244beaf4e7\") " pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.461996 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv8t4\" (UniqueName: \"kubernetes.io/projected/c4887490-f9e3-4a5f-be93-ba244beaf4e7-kube-api-access-fv8t4\") pod \"metallb-operator-webhook-server-5b647b685b-2jdt4\" (UID: \"c4887490-f9e3-4a5f-be93-ba244beaf4e7\") " pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.573743 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5769477c66-cbbds"] Feb 25 07:01:04 crc kubenswrapper[4978]: W0225 07:01:04.585926 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod296657b4_46fe_4780_b077_65d11eeab81d.slice/crio-d2b1b9ac4f673fbd0af1100361d05d161f49375b83ec13bb6a108d617ea5de76 WatchSource:0}: Error finding container d2b1b9ac4f673fbd0af1100361d05d161f49375b83ec13bb6a108d617ea5de76: Status 404 returned error can't find the container with id d2b1b9ac4f673fbd0af1100361d05d161f49375b83ec13bb6a108d617ea5de76 Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.605161 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.796234 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" event={"ID":"296657b4-46fe-4780-b077-65d11eeab81d","Type":"ContainerStarted","Data":"d2b1b9ac4f673fbd0af1100361d05d161f49375b83ec13bb6a108d617ea5de76"} Feb 25 07:01:04 crc kubenswrapper[4978]: I0225 07:01:04.894727 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4"] Feb 25 07:01:05 crc kubenswrapper[4978]: I0225 07:01:05.817748 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" event={"ID":"c4887490-f9e3-4a5f-be93-ba244beaf4e7","Type":"ContainerStarted","Data":"d8dfa65d2ecd59188fd01ab269204be165e43c6d9388b1b5a2cd3901a974f158"} Feb 25 07:01:09 crc kubenswrapper[4978]: I0225 07:01:09.863218 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" event={"ID":"c4887490-f9e3-4a5f-be93-ba244beaf4e7","Type":"ContainerStarted","Data":"9f998c13edd93ad79fbe972cf3c9ed56a56a0dd4117120405d78d849dba1cbb8"} Feb 25 07:01:09 crc kubenswrapper[4978]: I0225 07:01:09.863737 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:09 crc kubenswrapper[4978]: I0225 07:01:09.865501 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" event={"ID":"296657b4-46fe-4780-b077-65d11eeab81d","Type":"ContainerStarted","Data":"1102c4f0b510a4d0a096a995c15cf027c464dba4a305f23edb32e1b02149b4b7"} Feb 25 07:01:09 crc kubenswrapper[4978]: I0225 07:01:09.865681 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:09 crc kubenswrapper[4978]: I0225 07:01:09.901761 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" podStartSLOduration=1.539068119 podStartE2EDuration="5.901738589s" podCreationTimestamp="2026-02-25 07:01:04 +0000 UTC" firstStartedPulling="2026-02-25 07:01:04.905073471 +0000 UTC m=+958.344329930" lastFinishedPulling="2026-02-25 07:01:09.267743941 +0000 UTC m=+962.707000400" observedRunningTime="2026-02-25 07:01:09.896491773 +0000 UTC m=+963.335748252" watchObservedRunningTime="2026-02-25 07:01:09.901738589 +0000 UTC m=+963.340995058" Feb 25 07:01:09 crc kubenswrapper[4978]: I0225 07:01:09.946715 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" podStartSLOduration=2.277373336 podStartE2EDuration="6.946690796s" podCreationTimestamp="2026-02-25 07:01:03 +0000 UTC" firstStartedPulling="2026-02-25 07:01:04.590384536 +0000 UTC m=+958.029640985" lastFinishedPulling="2026-02-25 07:01:09.259701986 +0000 UTC m=+962.698958445" observedRunningTime="2026-02-25 07:01:09.940974804 +0000 UTC m=+963.380231283" watchObservedRunningTime="2026-02-25 07:01:09.946690796 +0000 UTC m=+963.385947275" Feb 25 07:01:16 crc kubenswrapper[4978]: I0225 07:01:16.540815 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:01:16 crc kubenswrapper[4978]: I0225 07:01:16.541476 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:01:24 crc kubenswrapper[4978]: I0225 07:01:24.611791 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5b647b685b-2jdt4" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.052356 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5769477c66-cbbds" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.826555 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl"] Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.827616 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.832538 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.834461 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-k8t8n" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.845664 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-xb4dk"] Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.848483 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.850513 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl"] Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.851829 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.852128 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.853690 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-reloader\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.853789 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a29c614-8113-4dd5-8209-3bf914925474-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-xwfgl\" (UID: \"3a29c614-8113-4dd5-8209-3bf914925474\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.853861 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqz4k\" (UniqueName: \"kubernetes.io/projected/3a29c614-8113-4dd5-8209-3bf914925474-kube-api-access-sqz4k\") pod \"frr-k8s-webhook-server-78b44bf5bb-xwfgl\" (UID: \"3a29c614-8113-4dd5-8209-3bf914925474\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.853884 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-frr-conf\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.853920 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-metrics-certs\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.853968 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-frr-startup\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.854026 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-metrics\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.854064 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfrtx\" (UniqueName: \"kubernetes.io/projected/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-kube-api-access-mfrtx\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.854085 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-frr-sockets\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.920497 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-bkcdb"] Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.921330 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bkcdb" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.923383 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.923388 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.923509 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dwxvz" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.923529 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.933147 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-6fkw5"] Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.933936 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.936448 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.944507 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-6fkw5"] Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.954939 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-metrics-certs\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.955012 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-frr-startup\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.955052 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-metrics\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.955072 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfrtx\" (UniqueName: \"kubernetes.io/projected/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-kube-api-access-mfrtx\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.955087 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-frr-sockets\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.955114 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-reloader\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.955140 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a29c614-8113-4dd5-8209-3bf914925474-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-xwfgl\" (UID: \"3a29c614-8113-4dd5-8209-3bf914925474\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.955177 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqz4k\" (UniqueName: \"kubernetes.io/projected/3a29c614-8113-4dd5-8209-3bf914925474-kube-api-access-sqz4k\") pod \"frr-k8s-webhook-server-78b44bf5bb-xwfgl\" (UID: \"3a29c614-8113-4dd5-8209-3bf914925474\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.955198 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-frr-conf\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.955699 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-frr-conf\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: E0225 07:01:44.955803 4978 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Feb 25 07:01:44 crc kubenswrapper[4978]: E0225 07:01:44.955852 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-metrics-certs podName:65f4ea4a-d2f4-4180-adbd-e464b90c84a0 nodeName:}" failed. No retries permitted until 2026-02-25 07:01:45.455835176 +0000 UTC m=+998.895091635 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-metrics-certs") pod "frr-k8s-xb4dk" (UID: "65f4ea4a-d2f4-4180-adbd-e464b90c84a0") : secret "frr-k8s-certs-secret" not found Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.956830 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-frr-startup\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.957026 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-metrics\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.957344 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-frr-sockets\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.957570 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-reloader\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.966216 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a29c614-8113-4dd5-8209-3bf914925474-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-xwfgl\" (UID: \"3a29c614-8113-4dd5-8209-3bf914925474\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.976009 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfrtx\" (UniqueName: \"kubernetes.io/projected/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-kube-api-access-mfrtx\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:44 crc kubenswrapper[4978]: I0225 07:01:44.979299 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqz4k\" (UniqueName: \"kubernetes.io/projected/3a29c614-8113-4dd5-8209-3bf914925474-kube-api-access-sqz4k\") pod \"frr-k8s-webhook-server-78b44bf5bb-xwfgl\" (UID: \"3a29c614-8113-4dd5-8209-3bf914925474\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.056441 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-metallb-excludel2\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.056573 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-metrics-certs\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.056628 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22ded0ab-9ff4-42c2-80fd-7423e610c25b-metrics-certs\") pod \"controller-69bbfbf88f-6fkw5\" (UID: \"22ded0ab-9ff4-42c2-80fd-7423e610c25b\") " pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.056661 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-memberlist\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.056679 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22ded0ab-9ff4-42c2-80fd-7423e610c25b-cert\") pod \"controller-69bbfbf88f-6fkw5\" (UID: \"22ded0ab-9ff4-42c2-80fd-7423e610c25b\") " pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.056702 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlw98\" (UniqueName: \"kubernetes.io/projected/22ded0ab-9ff4-42c2-80fd-7423e610c25b-kube-api-access-wlw98\") pod \"controller-69bbfbf88f-6fkw5\" (UID: \"22ded0ab-9ff4-42c2-80fd-7423e610c25b\") " pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.056728 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s77jw\" (UniqueName: \"kubernetes.io/projected/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-kube-api-access-s77jw\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.145865 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.157650 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-metrics-certs\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.157710 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22ded0ab-9ff4-42c2-80fd-7423e610c25b-metrics-certs\") pod \"controller-69bbfbf88f-6fkw5\" (UID: \"22ded0ab-9ff4-42c2-80fd-7423e610c25b\") " pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.157743 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-memberlist\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.157760 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22ded0ab-9ff4-42c2-80fd-7423e610c25b-cert\") pod \"controller-69bbfbf88f-6fkw5\" (UID: \"22ded0ab-9ff4-42c2-80fd-7423e610c25b\") " pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.157777 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlw98\" (UniqueName: \"kubernetes.io/projected/22ded0ab-9ff4-42c2-80fd-7423e610c25b-kube-api-access-wlw98\") pod \"controller-69bbfbf88f-6fkw5\" (UID: \"22ded0ab-9ff4-42c2-80fd-7423e610c25b\") " pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.157793 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s77jw\" (UniqueName: \"kubernetes.io/projected/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-kube-api-access-s77jw\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.157835 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-metallb-excludel2\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.158519 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-metallb-excludel2\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: E0225 07:01:45.158606 4978 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 25 07:01:45 crc kubenswrapper[4978]: E0225 07:01:45.158650 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-metrics-certs podName:0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb nodeName:}" failed. No retries permitted until 2026-02-25 07:01:45.658638342 +0000 UTC m=+999.097894801 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-metrics-certs") pod "speaker-bkcdb" (UID: "0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb") : secret "speaker-certs-secret" not found Feb 25 07:01:45 crc kubenswrapper[4978]: E0225 07:01:45.158842 4978 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Feb 25 07:01:45 crc kubenswrapper[4978]: E0225 07:01:45.158873 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22ded0ab-9ff4-42c2-80fd-7423e610c25b-metrics-certs podName:22ded0ab-9ff4-42c2-80fd-7423e610c25b nodeName:}" failed. No retries permitted until 2026-02-25 07:01:45.658865779 +0000 UTC m=+999.098122238 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/22ded0ab-9ff4-42c2-80fd-7423e610c25b-metrics-certs") pod "controller-69bbfbf88f-6fkw5" (UID: "22ded0ab-9ff4-42c2-80fd-7423e610c25b") : secret "controller-certs-secret" not found Feb 25 07:01:45 crc kubenswrapper[4978]: E0225 07:01:45.158910 4978 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 25 07:01:45 crc kubenswrapper[4978]: E0225 07:01:45.158927 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-memberlist podName:0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb nodeName:}" failed. No retries permitted until 2026-02-25 07:01:45.658922031 +0000 UTC m=+999.098178490 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-memberlist") pod "speaker-bkcdb" (UID: "0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb") : secret "metallb-memberlist" not found Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.166058 4978 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.178209 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/22ded0ab-9ff4-42c2-80fd-7423e610c25b-cert\") pod \"controller-69bbfbf88f-6fkw5\" (UID: \"22ded0ab-9ff4-42c2-80fd-7423e610c25b\") " pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.182155 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s77jw\" (UniqueName: \"kubernetes.io/projected/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-kube-api-access-s77jw\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.189159 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlw98\" (UniqueName: \"kubernetes.io/projected/22ded0ab-9ff4-42c2-80fd-7423e610c25b-kube-api-access-wlw98\") pod \"controller-69bbfbf88f-6fkw5\" (UID: \"22ded0ab-9ff4-42c2-80fd-7423e610c25b\") " pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.371867 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl"] Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.463056 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-metrics-certs\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.469943 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/65f4ea4a-d2f4-4180-adbd-e464b90c84a0-metrics-certs\") pod \"frr-k8s-xb4dk\" (UID: \"65f4ea4a-d2f4-4180-adbd-e464b90c84a0\") " pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.470287 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.665670 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-memberlist\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.665849 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-metrics-certs\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: E0225 07:01:45.665894 4978 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.665932 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22ded0ab-9ff4-42c2-80fd-7423e610c25b-metrics-certs\") pod \"controller-69bbfbf88f-6fkw5\" (UID: \"22ded0ab-9ff4-42c2-80fd-7423e610c25b\") " pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:45 crc kubenswrapper[4978]: E0225 07:01:45.665982 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-memberlist podName:0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb nodeName:}" failed. No retries permitted until 2026-02-25 07:01:46.66595313 +0000 UTC m=+1000.105209619 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-memberlist") pod "speaker-bkcdb" (UID: "0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb") : secret "metallb-memberlist" not found Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.667465 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xb4dk" event={"ID":"65f4ea4a-d2f4-4180-adbd-e464b90c84a0","Type":"ContainerStarted","Data":"d230eb06f5641f55538b73bf53bc23fcf48cbf06815db0cfa638f20ddaf4fcf6"} Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.670449 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" event={"ID":"3a29c614-8113-4dd5-8209-3bf914925474","Type":"ContainerStarted","Data":"1cbf321919ad5053e0e344994dce23cc68adbb09f8df42aac60c5028726eb80e"} Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.672445 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-metrics-certs\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.673316 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/22ded0ab-9ff4-42c2-80fd-7423e610c25b-metrics-certs\") pod \"controller-69bbfbf88f-6fkw5\" (UID: \"22ded0ab-9ff4-42c2-80fd-7423e610c25b\") " pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:45 crc kubenswrapper[4978]: I0225 07:01:45.848442 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.141445 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-6fkw5"] Feb 25 07:01:46 crc kubenswrapper[4978]: W0225 07:01:46.143833 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22ded0ab_9ff4_42c2_80fd_7423e610c25b.slice/crio-1360ad27214843842423a098e96cf2805dccc1689175103b37e3d45e49eb1942 WatchSource:0}: Error finding container 1360ad27214843842423a098e96cf2805dccc1689175103b37e3d45e49eb1942: Status 404 returned error can't find the container with id 1360ad27214843842423a098e96cf2805dccc1689175103b37e3d45e49eb1942 Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.540389 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.540750 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.540799 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.541456 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ea6d2ab6806f2c754878e481d2101f535bcfdd355af23df81d835564bbd06358"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.541517 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://ea6d2ab6806f2c754878e481d2101f535bcfdd355af23df81d835564bbd06358" gracePeriod=600 Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.682073 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-memberlist\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.688925 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb-memberlist\") pod \"speaker-bkcdb\" (UID: \"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb\") " pod="metallb-system/speaker-bkcdb" Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.697469 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="ea6d2ab6806f2c754878e481d2101f535bcfdd355af23df81d835564bbd06358" exitCode=0 Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.697515 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"ea6d2ab6806f2c754878e481d2101f535bcfdd355af23df81d835564bbd06358"} Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.697568 4978 scope.go:117] "RemoveContainer" containerID="d6649b854ec6cf99def7e6418068fe617fbe3b2759595971796ed8d3fffb9745" Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.703919 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-6fkw5" event={"ID":"22ded0ab-9ff4-42c2-80fd-7423e610c25b","Type":"ContainerStarted","Data":"c08a3c1c93b0546649b783659eee7b1c83e7c554311bae61f846a452e13466cb"} Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.703977 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-6fkw5" event={"ID":"22ded0ab-9ff4-42c2-80fd-7423e610c25b","Type":"ContainerStarted","Data":"5d7ef49eba4130bd1f0112079895980d56210270e072ff30acb00257e92517a1"} Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.703997 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-6fkw5" event={"ID":"22ded0ab-9ff4-42c2-80fd-7423e610c25b","Type":"ContainerStarted","Data":"1360ad27214843842423a098e96cf2805dccc1689175103b37e3d45e49eb1942"} Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.705061 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.723719 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-6fkw5" podStartSLOduration=2.723704689 podStartE2EDuration="2.723704689s" podCreationTimestamp="2026-02-25 07:01:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:01:46.720127526 +0000 UTC m=+1000.159383985" watchObservedRunningTime="2026-02-25 07:01:46.723704689 +0000 UTC m=+1000.162961148" Feb 25 07:01:46 crc kubenswrapper[4978]: I0225 07:01:46.734627 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bkcdb" Feb 25 07:01:47 crc kubenswrapper[4978]: I0225 07:01:47.717629 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bkcdb" event={"ID":"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb","Type":"ContainerStarted","Data":"96e91662958681f4377ac89bf1cfbbabd2356935b47fc322a3d226614e625f15"} Feb 25 07:01:47 crc kubenswrapper[4978]: I0225 07:01:47.718094 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bkcdb" event={"ID":"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb","Type":"ContainerStarted","Data":"ba1a9cbdb3997e0ff2ea399d4bbf01c33f03a9e7f47eedaa3c87e15beb476858"} Feb 25 07:01:47 crc kubenswrapper[4978]: I0225 07:01:47.718106 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bkcdb" event={"ID":"0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb","Type":"ContainerStarted","Data":"15a11e41db008e070a91b1ac4b4f29523c49664d79ba24c804d1af2a0306b2e3"} Feb 25 07:01:47 crc kubenswrapper[4978]: I0225 07:01:47.718290 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-bkcdb" Feb 25 07:01:47 crc kubenswrapper[4978]: I0225 07:01:47.722740 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"03d7d67e86736ca0a34d745efcdc9b40b54189fb150b16a7aa6067c654318766"} Feb 25 07:01:47 crc kubenswrapper[4978]: I0225 07:01:47.734657 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-bkcdb" podStartSLOduration=3.734639938 podStartE2EDuration="3.734639938s" podCreationTimestamp="2026-02-25 07:01:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:01:47.733973467 +0000 UTC m=+1001.173229926" watchObservedRunningTime="2026-02-25 07:01:47.734639938 +0000 UTC m=+1001.173896397" Feb 25 07:01:52 crc kubenswrapper[4978]: I0225 07:01:52.770430 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" event={"ID":"3a29c614-8113-4dd5-8209-3bf914925474","Type":"ContainerStarted","Data":"5e6fb384a9a7c2bcf95983fe1cdf09794e7939bce6a12d74e2f759c92f1f0c66"} Feb 25 07:01:52 crc kubenswrapper[4978]: I0225 07:01:52.771461 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" Feb 25 07:01:52 crc kubenswrapper[4978]: I0225 07:01:52.773006 4978 generic.go:334] "Generic (PLEG): container finished" podID="65f4ea4a-d2f4-4180-adbd-e464b90c84a0" containerID="22ccc70b49db26e3dd4d277919cc36cb5c57d757976264635a00e9fc792b1c54" exitCode=0 Feb 25 07:01:52 crc kubenswrapper[4978]: I0225 07:01:52.773095 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xb4dk" event={"ID":"65f4ea4a-d2f4-4180-adbd-e464b90c84a0","Type":"ContainerDied","Data":"22ccc70b49db26e3dd4d277919cc36cb5c57d757976264635a00e9fc792b1c54"} Feb 25 07:01:52 crc kubenswrapper[4978]: I0225 07:01:52.802763 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" podStartSLOduration=1.773508211 podStartE2EDuration="8.802737817s" podCreationTimestamp="2026-02-25 07:01:44 +0000 UTC" firstStartedPulling="2026-02-25 07:01:45.38400666 +0000 UTC m=+998.823263119" lastFinishedPulling="2026-02-25 07:01:52.413236236 +0000 UTC m=+1005.852492725" observedRunningTime="2026-02-25 07:01:52.797170101 +0000 UTC m=+1006.236426570" watchObservedRunningTime="2026-02-25 07:01:52.802737817 +0000 UTC m=+1006.241994286" Feb 25 07:01:53 crc kubenswrapper[4978]: I0225 07:01:53.783422 4978 generic.go:334] "Generic (PLEG): container finished" podID="65f4ea4a-d2f4-4180-adbd-e464b90c84a0" containerID="60ae7e8fdd562776cf9e9fe8d04d6c83dcb5f3cc8f4efb091423fdcaa916a9b8" exitCode=0 Feb 25 07:01:53 crc kubenswrapper[4978]: I0225 07:01:53.783522 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xb4dk" event={"ID":"65f4ea4a-d2f4-4180-adbd-e464b90c84a0","Type":"ContainerDied","Data":"60ae7e8fdd562776cf9e9fe8d04d6c83dcb5f3cc8f4efb091423fdcaa916a9b8"} Feb 25 07:01:54 crc kubenswrapper[4978]: I0225 07:01:54.796516 4978 generic.go:334] "Generic (PLEG): container finished" podID="65f4ea4a-d2f4-4180-adbd-e464b90c84a0" containerID="4e80545f1d8762abd4864884ba152c9d178b9fed4dd4fca017223db4748b1787" exitCode=0 Feb 25 07:01:54 crc kubenswrapper[4978]: I0225 07:01:54.796616 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xb4dk" event={"ID":"65f4ea4a-d2f4-4180-adbd-e464b90c84a0","Type":"ContainerDied","Data":"4e80545f1d8762abd4864884ba152c9d178b9fed4dd4fca017223db4748b1787"} Feb 25 07:01:55 crc kubenswrapper[4978]: I0225 07:01:55.811689 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xb4dk" event={"ID":"65f4ea4a-d2f4-4180-adbd-e464b90c84a0","Type":"ContainerStarted","Data":"2f6528e866c27f2f132589690b32f0cc3e7f6443e059c703a4d9ee262ad6268b"} Feb 25 07:01:55 crc kubenswrapper[4978]: I0225 07:01:55.812149 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xb4dk" event={"ID":"65f4ea4a-d2f4-4180-adbd-e464b90c84a0","Type":"ContainerStarted","Data":"b4581c15209b041816d63ce3f10ccd70c31654e8002cfc82a2d577a6dd9c8b59"} Feb 25 07:01:55 crc kubenswrapper[4978]: I0225 07:01:55.812196 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xb4dk" event={"ID":"65f4ea4a-d2f4-4180-adbd-e464b90c84a0","Type":"ContainerStarted","Data":"a4d3ffdb5df7c34bf3cab0567f5f6311b0c0a62ed71bd6878277d8baf46ca830"} Feb 25 07:01:55 crc kubenswrapper[4978]: I0225 07:01:55.812219 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xb4dk" event={"ID":"65f4ea4a-d2f4-4180-adbd-e464b90c84a0","Type":"ContainerStarted","Data":"f8fb76d28848b269a8c335013152312e8100846740a5896fe7bcc58dea36b9bc"} Feb 25 07:01:56 crc kubenswrapper[4978]: I0225 07:01:56.824398 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xb4dk" event={"ID":"65f4ea4a-d2f4-4180-adbd-e464b90c84a0","Type":"ContainerStarted","Data":"10959d08b1007b13582e7fa72cc912a26413da1b1accb328d9582efeb64bf499"} Feb 25 07:01:56 crc kubenswrapper[4978]: I0225 07:01:56.825008 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-xb4dk" event={"ID":"65f4ea4a-d2f4-4180-adbd-e464b90c84a0","Type":"ContainerStarted","Data":"c757608104bb1ce5c08c58123ef8601e9989294998d2e07d65a310b6279c407a"} Feb 25 07:01:56 crc kubenswrapper[4978]: I0225 07:01:56.825045 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:01:56 crc kubenswrapper[4978]: I0225 07:01:56.865813 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-xb4dk" podStartSLOduration=6.072754339 podStartE2EDuration="12.865789763s" podCreationTimestamp="2026-02-25 07:01:44 +0000 UTC" firstStartedPulling="2026-02-25 07:01:45.618216509 +0000 UTC m=+999.057472978" lastFinishedPulling="2026-02-25 07:01:52.411251913 +0000 UTC m=+1005.850508402" observedRunningTime="2026-02-25 07:01:56.858680769 +0000 UTC m=+1010.297937268" watchObservedRunningTime="2026-02-25 07:01:56.865789763 +0000 UTC m=+1010.305046252" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.139694 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533382-pvr9s"] Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.142873 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533382-pvr9s" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.145136 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.145694 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.146180 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.158281 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533382-pvr9s"] Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.295590 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkk56\" (UniqueName: \"kubernetes.io/projected/1cc2f335-e592-465e-93e9-ce8ba085dca6-kube-api-access-lkk56\") pod \"auto-csr-approver-29533382-pvr9s\" (UID: \"1cc2f335-e592-465e-93e9-ce8ba085dca6\") " pod="openshift-infra/auto-csr-approver-29533382-pvr9s" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.397981 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkk56\" (UniqueName: \"kubernetes.io/projected/1cc2f335-e592-465e-93e9-ce8ba085dca6-kube-api-access-lkk56\") pod \"auto-csr-approver-29533382-pvr9s\" (UID: \"1cc2f335-e592-465e-93e9-ce8ba085dca6\") " pod="openshift-infra/auto-csr-approver-29533382-pvr9s" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.438556 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkk56\" (UniqueName: \"kubernetes.io/projected/1cc2f335-e592-465e-93e9-ce8ba085dca6-kube-api-access-lkk56\") pod \"auto-csr-approver-29533382-pvr9s\" (UID: \"1cc2f335-e592-465e-93e9-ce8ba085dca6\") " pod="openshift-infra/auto-csr-approver-29533382-pvr9s" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.470740 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.476231 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533382-pvr9s" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.538667 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.777226 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533382-pvr9s"] Feb 25 07:02:00 crc kubenswrapper[4978]: I0225 07:02:00.869923 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533382-pvr9s" event={"ID":"1cc2f335-e592-465e-93e9-ce8ba085dca6","Type":"ContainerStarted","Data":"7e9cda354125825d735a6801efd579c49a8d82747b2249654ebfafd4f93033ab"} Feb 25 07:02:02 crc kubenswrapper[4978]: I0225 07:02:02.889249 4978 generic.go:334] "Generic (PLEG): container finished" podID="1cc2f335-e592-465e-93e9-ce8ba085dca6" containerID="6a5aed0cd78d1335f2c61b8685492f41cff7d1fc97c36517a6514e5c9e63dfe8" exitCode=0 Feb 25 07:02:02 crc kubenswrapper[4978]: I0225 07:02:02.889566 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533382-pvr9s" event={"ID":"1cc2f335-e592-465e-93e9-ce8ba085dca6","Type":"ContainerDied","Data":"6a5aed0cd78d1335f2c61b8685492f41cff7d1fc97c36517a6514e5c9e63dfe8"} Feb 25 07:02:04 crc kubenswrapper[4978]: I0225 07:02:04.264329 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533382-pvr9s" Feb 25 07:02:04 crc kubenswrapper[4978]: I0225 07:02:04.402776 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkk56\" (UniqueName: \"kubernetes.io/projected/1cc2f335-e592-465e-93e9-ce8ba085dca6-kube-api-access-lkk56\") pod \"1cc2f335-e592-465e-93e9-ce8ba085dca6\" (UID: \"1cc2f335-e592-465e-93e9-ce8ba085dca6\") " Feb 25 07:02:04 crc kubenswrapper[4978]: I0225 07:02:04.414671 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cc2f335-e592-465e-93e9-ce8ba085dca6-kube-api-access-lkk56" (OuterVolumeSpecName: "kube-api-access-lkk56") pod "1cc2f335-e592-465e-93e9-ce8ba085dca6" (UID: "1cc2f335-e592-465e-93e9-ce8ba085dca6"). InnerVolumeSpecName "kube-api-access-lkk56". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:02:04 crc kubenswrapper[4978]: I0225 07:02:04.506584 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkk56\" (UniqueName: \"kubernetes.io/projected/1cc2f335-e592-465e-93e9-ce8ba085dca6-kube-api-access-lkk56\") on node \"crc\" DevicePath \"\"" Feb 25 07:02:04 crc kubenswrapper[4978]: I0225 07:02:04.907473 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533382-pvr9s" event={"ID":"1cc2f335-e592-465e-93e9-ce8ba085dca6","Type":"ContainerDied","Data":"7e9cda354125825d735a6801efd579c49a8d82747b2249654ebfafd4f93033ab"} Feb 25 07:02:04 crc kubenswrapper[4978]: I0225 07:02:04.907519 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e9cda354125825d735a6801efd579c49a8d82747b2249654ebfafd4f93033ab" Feb 25 07:02:04 crc kubenswrapper[4978]: I0225 07:02:04.907529 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533382-pvr9s" Feb 25 07:02:05 crc kubenswrapper[4978]: I0225 07:02:05.151239 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-xwfgl" Feb 25 07:02:05 crc kubenswrapper[4978]: I0225 07:02:05.355698 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533376-xrqml"] Feb 25 07:02:05 crc kubenswrapper[4978]: I0225 07:02:05.362816 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533376-xrqml"] Feb 25 07:02:05 crc kubenswrapper[4978]: I0225 07:02:05.475185 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-xb4dk" Feb 25 07:02:05 crc kubenswrapper[4978]: I0225 07:02:05.854307 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-6fkw5" Feb 25 07:02:06 crc kubenswrapper[4978]: I0225 07:02:06.740423 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-bkcdb" Feb 25 07:02:07 crc kubenswrapper[4978]: I0225 07:02:07.336253 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dc77d60-3127-4d1d-acfa-87d1a8822d8a" path="/var/lib/kubelet/pods/7dc77d60-3127-4d1d-acfa-87d1a8822d8a/volumes" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.383591 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v"] Feb 25 07:02:08 crc kubenswrapper[4978]: E0225 07:02:08.383969 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cc2f335-e592-465e-93e9-ce8ba085dca6" containerName="oc" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.383990 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cc2f335-e592-465e-93e9-ce8ba085dca6" containerName="oc" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.384186 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cc2f335-e592-465e-93e9-ce8ba085dca6" containerName="oc" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.385594 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.391563 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v"] Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.419564 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.460799 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.460900 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbxdh\" (UniqueName: \"kubernetes.io/projected/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-kube-api-access-vbxdh\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.460932 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.561787 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbxdh\" (UniqueName: \"kubernetes.io/projected/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-kube-api-access-vbxdh\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.562044 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.562148 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.562700 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.562700 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.592880 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbxdh\" (UniqueName: \"kubernetes.io/projected/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-kube-api-access-vbxdh\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:08 crc kubenswrapper[4978]: I0225 07:02:08.745923 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:09 crc kubenswrapper[4978]: I0225 07:02:09.001537 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v"] Feb 25 07:02:09 crc kubenswrapper[4978]: W0225 07:02:09.011515 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a65fcb7_9cf8_40b8_96d5_4b520ad4357e.slice/crio-d897194559141a0baca30e9df747072f57f95b819bc76559ac3a14286ae8cff8 WatchSource:0}: Error finding container d897194559141a0baca30e9df747072f57f95b819bc76559ac3a14286ae8cff8: Status 404 returned error can't find the container with id d897194559141a0baca30e9df747072f57f95b819bc76559ac3a14286ae8cff8 Feb 25 07:02:09 crc kubenswrapper[4978]: I0225 07:02:09.951826 4978 generic.go:334] "Generic (PLEG): container finished" podID="6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" containerID="ba9f66416fc4a7a6b61d9b43f13941015fa7c3d3c0f4618b6f493e58d0bffa8b" exitCode=0 Feb 25 07:02:09 crc kubenswrapper[4978]: I0225 07:02:09.952178 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" event={"ID":"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e","Type":"ContainerDied","Data":"ba9f66416fc4a7a6b61d9b43f13941015fa7c3d3c0f4618b6f493e58d0bffa8b"} Feb 25 07:02:09 crc kubenswrapper[4978]: I0225 07:02:09.952215 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" event={"ID":"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e","Type":"ContainerStarted","Data":"d897194559141a0baca30e9df747072f57f95b819bc76559ac3a14286ae8cff8"} Feb 25 07:02:13 crc kubenswrapper[4978]: I0225 07:02:13.987318 4978 generic.go:334] "Generic (PLEG): container finished" podID="6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" containerID="3ddcd73df042628f315ff569b236ad3e32f0085353fe90aa466203185b5a58d3" exitCode=0 Feb 25 07:02:13 crc kubenswrapper[4978]: I0225 07:02:13.987445 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" event={"ID":"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e","Type":"ContainerDied","Data":"3ddcd73df042628f315ff569b236ad3e32f0085353fe90aa466203185b5a58d3"} Feb 25 07:02:14 crc kubenswrapper[4978]: I0225 07:02:14.798307 4978 scope.go:117] "RemoveContainer" containerID="767d35074d443f0c5027d813c91e7792fdc019d8ca89a519280927f80bb82cf8" Feb 25 07:02:14 crc kubenswrapper[4978]: I0225 07:02:14.995395 4978 generic.go:334] "Generic (PLEG): container finished" podID="6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" containerID="54b3e3a574e501c8c6f14df26788687d569c0ae2d6e97e18e0d9750b7f5d1658" exitCode=0 Feb 25 07:02:14 crc kubenswrapper[4978]: I0225 07:02:14.995442 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" event={"ID":"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e","Type":"ContainerDied","Data":"54b3e3a574e501c8c6f14df26788687d569c0ae2d6e97e18e0d9750b7f5d1658"} Feb 25 07:02:16 crc kubenswrapper[4978]: I0225 07:02:16.488296 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:16 crc kubenswrapper[4978]: I0225 07:02:16.489351 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-bundle\") pod \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " Feb 25 07:02:16 crc kubenswrapper[4978]: I0225 07:02:16.489490 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbxdh\" (UniqueName: \"kubernetes.io/projected/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-kube-api-access-vbxdh\") pod \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " Feb 25 07:02:16 crc kubenswrapper[4978]: I0225 07:02:16.489548 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-util\") pod \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\" (UID: \"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e\") " Feb 25 07:02:16 crc kubenswrapper[4978]: I0225 07:02:16.490974 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-bundle" (OuterVolumeSpecName: "bundle") pod "6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" (UID: "6a65fcb7-9cf8-40b8-96d5-4b520ad4357e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:02:16 crc kubenswrapper[4978]: I0225 07:02:16.498107 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-kube-api-access-vbxdh" (OuterVolumeSpecName: "kube-api-access-vbxdh") pod "6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" (UID: "6a65fcb7-9cf8-40b8-96d5-4b520ad4357e"). InnerVolumeSpecName "kube-api-access-vbxdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:02:16 crc kubenswrapper[4978]: I0225 07:02:16.519627 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-util" (OuterVolumeSpecName: "util") pod "6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" (UID: "6a65fcb7-9cf8-40b8-96d5-4b520ad4357e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:02:16 crc kubenswrapper[4978]: I0225 07:02:16.591606 4978 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:02:16 crc kubenswrapper[4978]: I0225 07:02:16.591700 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbxdh\" (UniqueName: \"kubernetes.io/projected/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-kube-api-access-vbxdh\") on node \"crc\" DevicePath \"\"" Feb 25 07:02:16 crc kubenswrapper[4978]: I0225 07:02:16.591733 4978 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6a65fcb7-9cf8-40b8-96d5-4b520ad4357e-util\") on node \"crc\" DevicePath \"\"" Feb 25 07:02:17 crc kubenswrapper[4978]: I0225 07:02:17.018463 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" event={"ID":"6a65fcb7-9cf8-40b8-96d5-4b520ad4357e","Type":"ContainerDied","Data":"d897194559141a0baca30e9df747072f57f95b819bc76559ac3a14286ae8cff8"} Feb 25 07:02:17 crc kubenswrapper[4978]: I0225 07:02:17.018510 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v" Feb 25 07:02:17 crc kubenswrapper[4978]: I0225 07:02:17.018502 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d897194559141a0baca30e9df747072f57f95b819bc76559ac3a14286ae8cff8" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.160669 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn"] Feb 25 07:02:22 crc kubenswrapper[4978]: E0225 07:02:22.161614 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" containerName="util" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.161635 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" containerName="util" Feb 25 07:02:22 crc kubenswrapper[4978]: E0225 07:02:22.161655 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" containerName="pull" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.161668 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" containerName="pull" Feb 25 07:02:22 crc kubenswrapper[4978]: E0225 07:02:22.161699 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" containerName="extract" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.161715 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" containerName="extract" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.161899 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a65fcb7-9cf8-40b8-96d5-4b520ad4357e" containerName="extract" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.162619 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.168042 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.168260 4978 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-bkxlr" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.172692 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.177432 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgk7s\" (UniqueName: \"kubernetes.io/projected/8b7a3242-944b-4f7b-b6db-d9cca7112665-kube-api-access-bgk7s\") pod \"cert-manager-operator-controller-manager-66c8bdd694-5rkhn\" (UID: \"8b7a3242-944b-4f7b-b6db-d9cca7112665\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.177593 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8b7a3242-944b-4f7b-b6db-d9cca7112665-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-5rkhn\" (UID: \"8b7a3242-944b-4f7b-b6db-d9cca7112665\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.231398 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn"] Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.279641 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgk7s\" (UniqueName: \"kubernetes.io/projected/8b7a3242-944b-4f7b-b6db-d9cca7112665-kube-api-access-bgk7s\") pod \"cert-manager-operator-controller-manager-66c8bdd694-5rkhn\" (UID: \"8b7a3242-944b-4f7b-b6db-d9cca7112665\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.279735 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8b7a3242-944b-4f7b-b6db-d9cca7112665-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-5rkhn\" (UID: \"8b7a3242-944b-4f7b-b6db-d9cca7112665\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.280394 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8b7a3242-944b-4f7b-b6db-d9cca7112665-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-5rkhn\" (UID: \"8b7a3242-944b-4f7b-b6db-d9cca7112665\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.316413 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgk7s\" (UniqueName: \"kubernetes.io/projected/8b7a3242-944b-4f7b-b6db-d9cca7112665-kube-api-access-bgk7s\") pod \"cert-manager-operator-controller-manager-66c8bdd694-5rkhn\" (UID: \"8b7a3242-944b-4f7b-b6db-d9cca7112665\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.484234 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" Feb 25 07:02:22 crc kubenswrapper[4978]: I0225 07:02:22.771638 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn"] Feb 25 07:02:23 crc kubenswrapper[4978]: I0225 07:02:23.068654 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" event={"ID":"8b7a3242-944b-4f7b-b6db-d9cca7112665","Type":"ContainerStarted","Data":"b3e1bec43f5ce773dc238c45bc91e8ed8903d842e0fcc7decdf85032b501f058"} Feb 25 07:02:26 crc kubenswrapper[4978]: I0225 07:02:26.092870 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" event={"ID":"8b7a3242-944b-4f7b-b6db-d9cca7112665","Type":"ContainerStarted","Data":"6b76c1c9f65e6209d6f017f7752fec9aed646f36531be1e35b2c0b4bebcf8aa9"} Feb 25 07:02:26 crc kubenswrapper[4978]: I0225 07:02:26.121012 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-5rkhn" podStartSLOduration=1.410664183 podStartE2EDuration="4.120994839s" podCreationTimestamp="2026-02-25 07:02:22 +0000 UTC" firstStartedPulling="2026-02-25 07:02:22.783429271 +0000 UTC m=+1036.222685720" lastFinishedPulling="2026-02-25 07:02:25.493759907 +0000 UTC m=+1038.933016376" observedRunningTime="2026-02-25 07:02:26.116995133 +0000 UTC m=+1039.556251582" watchObservedRunningTime="2026-02-25 07:02:26.120994839 +0000 UTC m=+1039.560251298" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.476945 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-9sx64"] Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.478502 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.481252 4978 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-x97lb" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.481544 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.483850 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-9sx64"] Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.493060 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.622225 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a262ae0-ec64-4848-9283-92b155feb047-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-9sx64\" (UID: \"8a262ae0-ec64-4848-9283-92b155feb047\") " pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.622328 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzhv4\" (UniqueName: \"kubernetes.io/projected/8a262ae0-ec64-4848-9283-92b155feb047-kube-api-access-lzhv4\") pod \"cert-manager-webhook-6888856db4-9sx64\" (UID: \"8a262ae0-ec64-4848-9283-92b155feb047\") " pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.724047 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a262ae0-ec64-4848-9283-92b155feb047-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-9sx64\" (UID: \"8a262ae0-ec64-4848-9283-92b155feb047\") " pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.724420 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzhv4\" (UniqueName: \"kubernetes.io/projected/8a262ae0-ec64-4848-9283-92b155feb047-kube-api-access-lzhv4\") pod \"cert-manager-webhook-6888856db4-9sx64\" (UID: \"8a262ae0-ec64-4848-9283-92b155feb047\") " pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.746198 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzhv4\" (UniqueName: \"kubernetes.io/projected/8a262ae0-ec64-4848-9283-92b155feb047-kube-api-access-lzhv4\") pod \"cert-manager-webhook-6888856db4-9sx64\" (UID: \"8a262ae0-ec64-4848-9283-92b155feb047\") " pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.750415 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a262ae0-ec64-4848-9283-92b155feb047-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-9sx64\" (UID: \"8a262ae0-ec64-4848-9283-92b155feb047\") " pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" Feb 25 07:02:29 crc kubenswrapper[4978]: I0225 07:02:29.794721 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" Feb 25 07:02:30 crc kubenswrapper[4978]: I0225 07:02:30.278821 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-9sx64"] Feb 25 07:02:31 crc kubenswrapper[4978]: I0225 07:02:31.121299 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" event={"ID":"8a262ae0-ec64-4848-9283-92b155feb047","Type":"ContainerStarted","Data":"7c92fed5eabcf3886170e79ac95e4f4af14cc555dfe2fcbb9bd7f1178b441b8f"} Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.042236 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-xplnd"] Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.043131 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.045519 4978 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-rtj9t" Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.055539 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-xplnd"] Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.157285 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f499cab6-b321-4fac-808f-c2ff15ca12ff-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-xplnd\" (UID: \"f499cab6-b321-4fac-808f-c2ff15ca12ff\") " pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.157350 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rfvq\" (UniqueName: \"kubernetes.io/projected/f499cab6-b321-4fac-808f-c2ff15ca12ff-kube-api-access-4rfvq\") pod \"cert-manager-cainjector-5545bd876-xplnd\" (UID: \"f499cab6-b321-4fac-808f-c2ff15ca12ff\") " pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.258654 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rfvq\" (UniqueName: \"kubernetes.io/projected/f499cab6-b321-4fac-808f-c2ff15ca12ff-kube-api-access-4rfvq\") pod \"cert-manager-cainjector-5545bd876-xplnd\" (UID: \"f499cab6-b321-4fac-808f-c2ff15ca12ff\") " pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.258783 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f499cab6-b321-4fac-808f-c2ff15ca12ff-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-xplnd\" (UID: \"f499cab6-b321-4fac-808f-c2ff15ca12ff\") " pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.278824 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f499cab6-b321-4fac-808f-c2ff15ca12ff-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-xplnd\" (UID: \"f499cab6-b321-4fac-808f-c2ff15ca12ff\") " pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.279003 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rfvq\" (UniqueName: \"kubernetes.io/projected/f499cab6-b321-4fac-808f-c2ff15ca12ff-kube-api-access-4rfvq\") pod \"cert-manager-cainjector-5545bd876-xplnd\" (UID: \"f499cab6-b321-4fac-808f-c2ff15ca12ff\") " pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.367169 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" Feb 25 07:02:32 crc kubenswrapper[4978]: I0225 07:02:32.685315 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-xplnd"] Feb 25 07:02:32 crc kubenswrapper[4978]: W0225 07:02:32.700390 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf499cab6_b321_4fac_808f_c2ff15ca12ff.slice/crio-972a9b0d814e0d3ca07938868680050dc5d8e079220bdc31e1391760a418e8f2 WatchSource:0}: Error finding container 972a9b0d814e0d3ca07938868680050dc5d8e079220bdc31e1391760a418e8f2: Status 404 returned error can't find the container with id 972a9b0d814e0d3ca07938868680050dc5d8e079220bdc31e1391760a418e8f2 Feb 25 07:02:33 crc kubenswrapper[4978]: I0225 07:02:33.134939 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" event={"ID":"f499cab6-b321-4fac-808f-c2ff15ca12ff","Type":"ContainerStarted","Data":"972a9b0d814e0d3ca07938868680050dc5d8e079220bdc31e1391760a418e8f2"} Feb 25 07:02:35 crc kubenswrapper[4978]: I0225 07:02:35.150723 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" event={"ID":"8a262ae0-ec64-4848-9283-92b155feb047","Type":"ContainerStarted","Data":"5af270cdef853dbabeb482fbd0812cb1bfcb6b7f704d3002e830dd23a9aae6b9"} Feb 25 07:02:35 crc kubenswrapper[4978]: I0225 07:02:35.151095 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" Feb 25 07:02:35 crc kubenswrapper[4978]: I0225 07:02:35.154794 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" event={"ID":"f499cab6-b321-4fac-808f-c2ff15ca12ff","Type":"ContainerStarted","Data":"1755138b2bd8b871ed392ff31967b7c00e5a7628bc5a8d261af63ab827e47d7c"} Feb 25 07:02:35 crc kubenswrapper[4978]: I0225 07:02:35.174213 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" podStartSLOduration=1.778437396 podStartE2EDuration="6.174182816s" podCreationTimestamp="2026-02-25 07:02:29 +0000 UTC" firstStartedPulling="2026-02-25 07:02:30.286996141 +0000 UTC m=+1043.726252600" lastFinishedPulling="2026-02-25 07:02:34.682741571 +0000 UTC m=+1048.121998020" observedRunningTime="2026-02-25 07:02:35.171819851 +0000 UTC m=+1048.611076330" watchObservedRunningTime="2026-02-25 07:02:35.174182816 +0000 UTC m=+1048.613439315" Feb 25 07:02:35 crc kubenswrapper[4978]: I0225 07:02:35.191783 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-xplnd" podStartSLOduration=1.2188542039999999 podStartE2EDuration="3.191753682s" podCreationTimestamp="2026-02-25 07:02:32 +0000 UTC" firstStartedPulling="2026-02-25 07:02:32.703359007 +0000 UTC m=+1046.142615466" lastFinishedPulling="2026-02-25 07:02:34.676258455 +0000 UTC m=+1048.115514944" observedRunningTime="2026-02-25 07:02:35.188924343 +0000 UTC m=+1048.628180822" watchObservedRunningTime="2026-02-25 07:02:35.191753682 +0000 UTC m=+1048.631010181" Feb 25 07:02:39 crc kubenswrapper[4978]: I0225 07:02:39.798502 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-9sx64" Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.067718 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-z9ghg"] Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.069677 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-z9ghg" Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.075322 4978 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-r77ns" Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.075506 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-z9ghg"] Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.196126 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vppqv\" (UniqueName: \"kubernetes.io/projected/89615955-fd40-4bf0-9bc1-b29188ff03c7-kube-api-access-vppqv\") pod \"cert-manager-545d4d4674-z9ghg\" (UID: \"89615955-fd40-4bf0-9bc1-b29188ff03c7\") " pod="cert-manager/cert-manager-545d4d4674-z9ghg" Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.196741 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/89615955-fd40-4bf0-9bc1-b29188ff03c7-bound-sa-token\") pod \"cert-manager-545d4d4674-z9ghg\" (UID: \"89615955-fd40-4bf0-9bc1-b29188ff03c7\") " pod="cert-manager/cert-manager-545d4d4674-z9ghg" Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.299212 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/89615955-fd40-4bf0-9bc1-b29188ff03c7-bound-sa-token\") pod \"cert-manager-545d4d4674-z9ghg\" (UID: \"89615955-fd40-4bf0-9bc1-b29188ff03c7\") " pod="cert-manager/cert-manager-545d4d4674-z9ghg" Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.299315 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vppqv\" (UniqueName: \"kubernetes.io/projected/89615955-fd40-4bf0-9bc1-b29188ff03c7-kube-api-access-vppqv\") pod \"cert-manager-545d4d4674-z9ghg\" (UID: \"89615955-fd40-4bf0-9bc1-b29188ff03c7\") " pod="cert-manager/cert-manager-545d4d4674-z9ghg" Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.335134 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/89615955-fd40-4bf0-9bc1-b29188ff03c7-bound-sa-token\") pod \"cert-manager-545d4d4674-z9ghg\" (UID: \"89615955-fd40-4bf0-9bc1-b29188ff03c7\") " pod="cert-manager/cert-manager-545d4d4674-z9ghg" Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.335524 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vppqv\" (UniqueName: \"kubernetes.io/projected/89615955-fd40-4bf0-9bc1-b29188ff03c7-kube-api-access-vppqv\") pod \"cert-manager-545d4d4674-z9ghg\" (UID: \"89615955-fd40-4bf0-9bc1-b29188ff03c7\") " pod="cert-manager/cert-manager-545d4d4674-z9ghg" Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.394561 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-z9ghg" Feb 25 07:02:48 crc kubenswrapper[4978]: I0225 07:02:48.955107 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-z9ghg"] Feb 25 07:02:48 crc kubenswrapper[4978]: W0225 07:02:48.964273 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89615955_fd40_4bf0_9bc1_b29188ff03c7.slice/crio-ce9063e6726cf04a3fab968e4c016195313c48dbfcec6bf16f4f5bf4870a30fb WatchSource:0}: Error finding container ce9063e6726cf04a3fab968e4c016195313c48dbfcec6bf16f4f5bf4870a30fb: Status 404 returned error can't find the container with id ce9063e6726cf04a3fab968e4c016195313c48dbfcec6bf16f4f5bf4870a30fb Feb 25 07:02:49 crc kubenswrapper[4978]: I0225 07:02:49.265358 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-z9ghg" event={"ID":"89615955-fd40-4bf0-9bc1-b29188ff03c7","Type":"ContainerStarted","Data":"49260e2c14b669387bb8b62877db32ab51c442e36919d43c19b568f839259cec"} Feb 25 07:02:49 crc kubenswrapper[4978]: I0225 07:02:49.265450 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-z9ghg" event={"ID":"89615955-fd40-4bf0-9bc1-b29188ff03c7","Type":"ContainerStarted","Data":"ce9063e6726cf04a3fab968e4c016195313c48dbfcec6bf16f4f5bf4870a30fb"} Feb 25 07:02:49 crc kubenswrapper[4978]: I0225 07:02:49.281558 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-z9ghg" podStartSLOduration=1.281539121 podStartE2EDuration="1.281539121s" podCreationTimestamp="2026-02-25 07:02:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:02:49.280648574 +0000 UTC m=+1062.719905043" watchObservedRunningTime="2026-02-25 07:02:49.281539121 +0000 UTC m=+1062.720795580" Feb 25 07:02:53 crc kubenswrapper[4978]: I0225 07:02:53.584415 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-svzw5"] Feb 25 07:02:53 crc kubenswrapper[4978]: I0225 07:02:53.586226 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-svzw5" Feb 25 07:02:53 crc kubenswrapper[4978]: I0225 07:02:53.588042 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-lldng" Feb 25 07:02:53 crc kubenswrapper[4978]: I0225 07:02:53.594250 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 25 07:02:53 crc kubenswrapper[4978]: I0225 07:02:53.594425 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 25 07:02:53 crc kubenswrapper[4978]: I0225 07:02:53.595779 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-svzw5"] Feb 25 07:02:53 crc kubenswrapper[4978]: I0225 07:02:53.722162 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxjr2\" (UniqueName: \"kubernetes.io/projected/f8954d98-7d6d-4978-8f14-6c3ecaae134c-kube-api-access-cxjr2\") pod \"openstack-operator-index-svzw5\" (UID: \"f8954d98-7d6d-4978-8f14-6c3ecaae134c\") " pod="openstack-operators/openstack-operator-index-svzw5" Feb 25 07:02:53 crc kubenswrapper[4978]: I0225 07:02:53.823221 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxjr2\" (UniqueName: \"kubernetes.io/projected/f8954d98-7d6d-4978-8f14-6c3ecaae134c-kube-api-access-cxjr2\") pod \"openstack-operator-index-svzw5\" (UID: \"f8954d98-7d6d-4978-8f14-6c3ecaae134c\") " pod="openstack-operators/openstack-operator-index-svzw5" Feb 25 07:02:53 crc kubenswrapper[4978]: I0225 07:02:53.845040 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxjr2\" (UniqueName: \"kubernetes.io/projected/f8954d98-7d6d-4978-8f14-6c3ecaae134c-kube-api-access-cxjr2\") pod \"openstack-operator-index-svzw5\" (UID: \"f8954d98-7d6d-4978-8f14-6c3ecaae134c\") " pod="openstack-operators/openstack-operator-index-svzw5" Feb 25 07:02:53 crc kubenswrapper[4978]: I0225 07:02:53.923249 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-svzw5" Feb 25 07:02:54 crc kubenswrapper[4978]: I0225 07:02:54.184660 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-svzw5"] Feb 25 07:02:54 crc kubenswrapper[4978]: W0225 07:02:54.193648 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8954d98_7d6d_4978_8f14_6c3ecaae134c.slice/crio-ee422792466719b1ec18d0466a385437e7cce8df51a4d07d3f77478d41e29807 WatchSource:0}: Error finding container ee422792466719b1ec18d0466a385437e7cce8df51a4d07d3f77478d41e29807: Status 404 returned error can't find the container with id ee422792466719b1ec18d0466a385437e7cce8df51a4d07d3f77478d41e29807 Feb 25 07:02:54 crc kubenswrapper[4978]: I0225 07:02:54.325570 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-svzw5" event={"ID":"f8954d98-7d6d-4978-8f14-6c3ecaae134c","Type":"ContainerStarted","Data":"ee422792466719b1ec18d0466a385437e7cce8df51a4d07d3f77478d41e29807"} Feb 25 07:02:55 crc kubenswrapper[4978]: I0225 07:02:55.951649 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-svzw5"] Feb 25 07:02:56 crc kubenswrapper[4978]: I0225 07:02:56.344637 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-svzw5" event={"ID":"f8954d98-7d6d-4978-8f14-6c3ecaae134c","Type":"ContainerStarted","Data":"7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf"} Feb 25 07:02:56 crc kubenswrapper[4978]: I0225 07:02:56.366587 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fhmb8"] Feb 25 07:02:56 crc kubenswrapper[4978]: I0225 07:02:56.368148 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fhmb8" Feb 25 07:02:56 crc kubenswrapper[4978]: I0225 07:02:56.377921 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fhmb8"] Feb 25 07:02:56 crc kubenswrapper[4978]: I0225 07:02:56.377926 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-svzw5" podStartSLOduration=2.172748507 podStartE2EDuration="3.377907491s" podCreationTimestamp="2026-02-25 07:02:53 +0000 UTC" firstStartedPulling="2026-02-25 07:02:54.196986402 +0000 UTC m=+1067.636242901" lastFinishedPulling="2026-02-25 07:02:55.402145396 +0000 UTC m=+1068.841401885" observedRunningTime="2026-02-25 07:02:56.374541005 +0000 UTC m=+1069.813797524" watchObservedRunningTime="2026-02-25 07:02:56.377907491 +0000 UTC m=+1069.817163990" Feb 25 07:02:56 crc kubenswrapper[4978]: I0225 07:02:56.469546 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s8b8\" (UniqueName: \"kubernetes.io/projected/bbee41d9-6d3d-4c1a-ad09-2bc260c74b4e-kube-api-access-2s8b8\") pod \"openstack-operator-index-fhmb8\" (UID: \"bbee41d9-6d3d-4c1a-ad09-2bc260c74b4e\") " pod="openstack-operators/openstack-operator-index-fhmb8" Feb 25 07:02:56 crc kubenswrapper[4978]: I0225 07:02:56.571910 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s8b8\" (UniqueName: \"kubernetes.io/projected/bbee41d9-6d3d-4c1a-ad09-2bc260c74b4e-kube-api-access-2s8b8\") pod \"openstack-operator-index-fhmb8\" (UID: \"bbee41d9-6d3d-4c1a-ad09-2bc260c74b4e\") " pod="openstack-operators/openstack-operator-index-fhmb8" Feb 25 07:02:56 crc kubenswrapper[4978]: I0225 07:02:56.605909 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s8b8\" (UniqueName: \"kubernetes.io/projected/bbee41d9-6d3d-4c1a-ad09-2bc260c74b4e-kube-api-access-2s8b8\") pod \"openstack-operator-index-fhmb8\" (UID: \"bbee41d9-6d3d-4c1a-ad09-2bc260c74b4e\") " pod="openstack-operators/openstack-operator-index-fhmb8" Feb 25 07:02:56 crc kubenswrapper[4978]: I0225 07:02:56.694271 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fhmb8" Feb 25 07:02:56 crc kubenswrapper[4978]: I0225 07:02:56.996181 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fhmb8"] Feb 25 07:02:57 crc kubenswrapper[4978]: W0225 07:02:57.001491 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbee41d9_6d3d_4c1a_ad09_2bc260c74b4e.slice/crio-b9ffb40a0315ebd3382ceff638d34ec494ebd7e0cca15acd6374365709d7fb15 WatchSource:0}: Error finding container b9ffb40a0315ebd3382ceff638d34ec494ebd7e0cca15acd6374365709d7fb15: Status 404 returned error can't find the container with id b9ffb40a0315ebd3382ceff638d34ec494ebd7e0cca15acd6374365709d7fb15 Feb 25 07:02:57 crc kubenswrapper[4978]: I0225 07:02:57.356255 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fhmb8" event={"ID":"bbee41d9-6d3d-4c1a-ad09-2bc260c74b4e","Type":"ContainerStarted","Data":"b9ffb40a0315ebd3382ceff638d34ec494ebd7e0cca15acd6374365709d7fb15"} Feb 25 07:02:57 crc kubenswrapper[4978]: I0225 07:02:57.356428 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-svzw5" podUID="f8954d98-7d6d-4978-8f14-6c3ecaae134c" containerName="registry-server" containerID="cri-o://7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf" gracePeriod=2 Feb 25 07:02:57 crc kubenswrapper[4978]: I0225 07:02:57.867130 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-svzw5" Feb 25 07:02:57 crc kubenswrapper[4978]: I0225 07:02:57.996962 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxjr2\" (UniqueName: \"kubernetes.io/projected/f8954d98-7d6d-4978-8f14-6c3ecaae134c-kube-api-access-cxjr2\") pod \"f8954d98-7d6d-4978-8f14-6c3ecaae134c\" (UID: \"f8954d98-7d6d-4978-8f14-6c3ecaae134c\") " Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.004601 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8954d98-7d6d-4978-8f14-6c3ecaae134c-kube-api-access-cxjr2" (OuterVolumeSpecName: "kube-api-access-cxjr2") pod "f8954d98-7d6d-4978-8f14-6c3ecaae134c" (UID: "f8954d98-7d6d-4978-8f14-6c3ecaae134c"). InnerVolumeSpecName "kube-api-access-cxjr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.099029 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxjr2\" (UniqueName: \"kubernetes.io/projected/f8954d98-7d6d-4978-8f14-6c3ecaae134c-kube-api-access-cxjr2\") on node \"crc\" DevicePath \"\"" Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.365985 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fhmb8" event={"ID":"bbee41d9-6d3d-4c1a-ad09-2bc260c74b4e","Type":"ContainerStarted","Data":"9b4de61a0ab43cfadf0ee2031bc1324d24f7c898611c044e50f44236f31794af"} Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.371092 4978 generic.go:334] "Generic (PLEG): container finished" podID="f8954d98-7d6d-4978-8f14-6c3ecaae134c" containerID="7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf" exitCode=0 Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.371158 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-svzw5" event={"ID":"f8954d98-7d6d-4978-8f14-6c3ecaae134c","Type":"ContainerDied","Data":"7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf"} Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.371168 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-svzw5" Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.371196 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-svzw5" event={"ID":"f8954d98-7d6d-4978-8f14-6c3ecaae134c","Type":"ContainerDied","Data":"ee422792466719b1ec18d0466a385437e7cce8df51a4d07d3f77478d41e29807"} Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.371234 4978 scope.go:117] "RemoveContainer" containerID="7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf" Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.402997 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fhmb8" podStartSLOduration=1.97271368 podStartE2EDuration="2.40296992s" podCreationTimestamp="2026-02-25 07:02:56 +0000 UTC" firstStartedPulling="2026-02-25 07:02:57.007632922 +0000 UTC m=+1070.446889391" lastFinishedPulling="2026-02-25 07:02:57.437889162 +0000 UTC m=+1070.877145631" observedRunningTime="2026-02-25 07:02:58.391898059 +0000 UTC m=+1071.831154588" watchObservedRunningTime="2026-02-25 07:02:58.40296992 +0000 UTC m=+1071.842226409" Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.415817 4978 scope.go:117] "RemoveContainer" containerID="7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf" Feb 25 07:02:58 crc kubenswrapper[4978]: E0225 07:02:58.416451 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf\": container with ID starting with 7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf not found: ID does not exist" containerID="7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf" Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.416510 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf"} err="failed to get container status \"7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf\": rpc error: code = NotFound desc = could not find container \"7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf\": container with ID starting with 7bffa27a2bf142c0f8e3581bd6672b2306e9b163f46094238332ef0be31747cf not found: ID does not exist" Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.439081 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-svzw5"] Feb 25 07:02:58 crc kubenswrapper[4978]: I0225 07:02:58.446233 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-svzw5"] Feb 25 07:02:59 crc kubenswrapper[4978]: I0225 07:02:59.339212 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8954d98-7d6d-4978-8f14-6c3ecaae134c" path="/var/lib/kubelet/pods/f8954d98-7d6d-4978-8f14-6c3ecaae134c/volumes" Feb 25 07:03:06 crc kubenswrapper[4978]: I0225 07:03:06.695456 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-fhmb8" Feb 25 07:03:06 crc kubenswrapper[4978]: I0225 07:03:06.696249 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-fhmb8" Feb 25 07:03:06 crc kubenswrapper[4978]: I0225 07:03:06.743985 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-fhmb8" Feb 25 07:03:07 crc kubenswrapper[4978]: I0225 07:03:07.485266 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-fhmb8" Feb 25 07:03:13 crc kubenswrapper[4978]: I0225 07:03:13.808575 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq"] Feb 25 07:03:13 crc kubenswrapper[4978]: E0225 07:03:13.809637 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8954d98-7d6d-4978-8f14-6c3ecaae134c" containerName="registry-server" Feb 25 07:03:13 crc kubenswrapper[4978]: I0225 07:03:13.809664 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8954d98-7d6d-4978-8f14-6c3ecaae134c" containerName="registry-server" Feb 25 07:03:13 crc kubenswrapper[4978]: I0225 07:03:13.809895 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8954d98-7d6d-4978-8f14-6c3ecaae134c" containerName="registry-server" Feb 25 07:03:13 crc kubenswrapper[4978]: I0225 07:03:13.811550 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:13 crc kubenswrapper[4978]: I0225 07:03:13.814409 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qtdjt" Feb 25 07:03:13 crc kubenswrapper[4978]: I0225 07:03:13.829887 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq"] Feb 25 07:03:13 crc kubenswrapper[4978]: I0225 07:03:13.944817 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-util\") pod \"11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:13 crc kubenswrapper[4978]: I0225 07:03:13.945396 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgkg9\" (UniqueName: \"kubernetes.io/projected/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-kube-api-access-hgkg9\") pod \"11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:13 crc kubenswrapper[4978]: I0225 07:03:13.945596 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-bundle\") pod \"11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:14 crc kubenswrapper[4978]: I0225 07:03:14.046926 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgkg9\" (UniqueName: \"kubernetes.io/projected/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-kube-api-access-hgkg9\") pod \"11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:14 crc kubenswrapper[4978]: I0225 07:03:14.046995 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-bundle\") pod \"11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:14 crc kubenswrapper[4978]: I0225 07:03:14.047093 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-util\") pod \"11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:14 crc kubenswrapper[4978]: I0225 07:03:14.047665 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-util\") pod \"11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:14 crc kubenswrapper[4978]: I0225 07:03:14.047953 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-bundle\") pod \"11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:14 crc kubenswrapper[4978]: I0225 07:03:14.080765 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgkg9\" (UniqueName: \"kubernetes.io/projected/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-kube-api-access-hgkg9\") pod \"11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:14 crc kubenswrapper[4978]: I0225 07:03:14.136808 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:14 crc kubenswrapper[4978]: I0225 07:03:14.624728 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq"] Feb 25 07:03:15 crc kubenswrapper[4978]: I0225 07:03:15.517221 4978 generic.go:334] "Generic (PLEG): container finished" podID="9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" containerID="963b1d40bd875923981ac7f24d5bccc2a67093c0e5159f4becce39c145ef4976" exitCode=0 Feb 25 07:03:15 crc kubenswrapper[4978]: I0225 07:03:15.517361 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" event={"ID":"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a","Type":"ContainerDied","Data":"963b1d40bd875923981ac7f24d5bccc2a67093c0e5159f4becce39c145ef4976"} Feb 25 07:03:15 crc kubenswrapper[4978]: I0225 07:03:15.517787 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" event={"ID":"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a","Type":"ContainerStarted","Data":"d6a1e8dc7e6946e63ef322bbbd6fa778220ca4f1e099810a668d8f8a954f613a"} Feb 25 07:03:16 crc kubenswrapper[4978]: I0225 07:03:16.526868 4978 generic.go:334] "Generic (PLEG): container finished" podID="9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" containerID="9598b23b80b3262c8312ffa3959b3e726ba68c90383419994faa55b2fdbbd8d9" exitCode=0 Feb 25 07:03:16 crc kubenswrapper[4978]: I0225 07:03:16.526922 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" event={"ID":"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a","Type":"ContainerDied","Data":"9598b23b80b3262c8312ffa3959b3e726ba68c90383419994faa55b2fdbbd8d9"} Feb 25 07:03:17 crc kubenswrapper[4978]: I0225 07:03:17.541650 4978 generic.go:334] "Generic (PLEG): container finished" podID="9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" containerID="406332b61e830c1f55321c64752de87a1333268af2d780e522bd7881fd5ea1d7" exitCode=0 Feb 25 07:03:17 crc kubenswrapper[4978]: I0225 07:03:17.541730 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" event={"ID":"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a","Type":"ContainerDied","Data":"406332b61e830c1f55321c64752de87a1333268af2d780e522bd7881fd5ea1d7"} Feb 25 07:03:18 crc kubenswrapper[4978]: I0225 07:03:18.883183 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.033071 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgkg9\" (UniqueName: \"kubernetes.io/projected/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-kube-api-access-hgkg9\") pod \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.033350 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-util\") pod \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.033433 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-bundle\") pod \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\" (UID: \"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a\") " Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.034254 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-bundle" (OuterVolumeSpecName: "bundle") pod "9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" (UID: "9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.045859 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-kube-api-access-hgkg9" (OuterVolumeSpecName: "kube-api-access-hgkg9") pod "9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" (UID: "9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a"). InnerVolumeSpecName "kube-api-access-hgkg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.046906 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-util" (OuterVolumeSpecName: "util") pod "9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" (UID: "9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.135362 4978 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-util\") on node \"crc\" DevicePath \"\"" Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.135423 4978 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.135439 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgkg9\" (UniqueName: \"kubernetes.io/projected/9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a-kube-api-access-hgkg9\") on node \"crc\" DevicePath \"\"" Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.564748 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" event={"ID":"9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a","Type":"ContainerDied","Data":"d6a1e8dc7e6946e63ef322bbbd6fa778220ca4f1e099810a668d8f8a954f613a"} Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.564813 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6a1e8dc7e6946e63ef322bbbd6fa778220ca4f1e099810a668d8f8a954f613a" Feb 25 07:03:19 crc kubenswrapper[4978]: I0225 07:03:19.565350 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.275845 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-55c649df44-nggwh"] Feb 25 07:03:27 crc kubenswrapper[4978]: E0225 07:03:27.276952 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" containerName="pull" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.276978 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" containerName="pull" Feb 25 07:03:27 crc kubenswrapper[4978]: E0225 07:03:27.277017 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" containerName="extract" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.277031 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" containerName="extract" Feb 25 07:03:27 crc kubenswrapper[4978]: E0225 07:03:27.277053 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" containerName="util" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.277066 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" containerName="util" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.277260 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a" containerName="extract" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.277923 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-55c649df44-nggwh" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.281356 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-4vphn" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.320577 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-55c649df44-nggwh"] Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.464244 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zfg2\" (UniqueName: \"kubernetes.io/projected/18a604af-ab94-4884-9750-63f4dd830912-kube-api-access-7zfg2\") pod \"openstack-operator-controller-init-55c649df44-nggwh\" (UID: \"18a604af-ab94-4884-9750-63f4dd830912\") " pod="openstack-operators/openstack-operator-controller-init-55c649df44-nggwh" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.565793 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zfg2\" (UniqueName: \"kubernetes.io/projected/18a604af-ab94-4884-9750-63f4dd830912-kube-api-access-7zfg2\") pod \"openstack-operator-controller-init-55c649df44-nggwh\" (UID: \"18a604af-ab94-4884-9750-63f4dd830912\") " pod="openstack-operators/openstack-operator-controller-init-55c649df44-nggwh" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.603414 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zfg2\" (UniqueName: \"kubernetes.io/projected/18a604af-ab94-4884-9750-63f4dd830912-kube-api-access-7zfg2\") pod \"openstack-operator-controller-init-55c649df44-nggwh\" (UID: \"18a604af-ab94-4884-9750-63f4dd830912\") " pod="openstack-operators/openstack-operator-controller-init-55c649df44-nggwh" Feb 25 07:03:27 crc kubenswrapper[4978]: I0225 07:03:27.895462 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-55c649df44-nggwh" Feb 25 07:03:28 crc kubenswrapper[4978]: I0225 07:03:28.431530 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-55c649df44-nggwh"] Feb 25 07:03:28 crc kubenswrapper[4978]: I0225 07:03:28.654247 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-55c649df44-nggwh" event={"ID":"18a604af-ab94-4884-9750-63f4dd830912","Type":"ContainerStarted","Data":"db9651a4ebc2646a4a44244807e8e8cf376a225795335e8e6bee302d908a7970"} Feb 25 07:03:33 crc kubenswrapper[4978]: I0225 07:03:33.702063 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-55c649df44-nggwh" event={"ID":"18a604af-ab94-4884-9750-63f4dd830912","Type":"ContainerStarted","Data":"022eba2fa1d8b4b99baf25d8eec926ae0ebee099bb9fe099f324bb3cc0128969"} Feb 25 07:03:33 crc kubenswrapper[4978]: I0225 07:03:33.703569 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-55c649df44-nggwh" Feb 25 07:03:46 crc kubenswrapper[4978]: I0225 07:03:46.541288 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:03:46 crc kubenswrapper[4978]: I0225 07:03:46.542123 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:03:47 crc kubenswrapper[4978]: I0225 07:03:47.899640 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-55c649df44-nggwh" Feb 25 07:03:47 crc kubenswrapper[4978]: I0225 07:03:47.974639 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-55c649df44-nggwh" podStartSLOduration=16.333813311 podStartE2EDuration="20.974605347s" podCreationTimestamp="2026-02-25 07:03:27 +0000 UTC" firstStartedPulling="2026-02-25 07:03:28.443504361 +0000 UTC m=+1101.882760850" lastFinishedPulling="2026-02-25 07:03:33.084296417 +0000 UTC m=+1106.523552886" observedRunningTime="2026-02-25 07:03:33.747981993 +0000 UTC m=+1107.187238452" watchObservedRunningTime="2026-02-25 07:03:47.974605347 +0000 UTC m=+1121.413861836" Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.153051 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533384-9q4l8"] Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.154543 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533384-9q4l8" Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.159236 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.160496 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.160892 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.172454 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533384-9q4l8"] Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.291252 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvl6s\" (UniqueName: \"kubernetes.io/projected/28930518-6de3-48af-94ff-63e0a1608fc4-kube-api-access-hvl6s\") pod \"auto-csr-approver-29533384-9q4l8\" (UID: \"28930518-6de3-48af-94ff-63e0a1608fc4\") " pod="openshift-infra/auto-csr-approver-29533384-9q4l8" Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.392959 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvl6s\" (UniqueName: \"kubernetes.io/projected/28930518-6de3-48af-94ff-63e0a1608fc4-kube-api-access-hvl6s\") pod \"auto-csr-approver-29533384-9q4l8\" (UID: \"28930518-6de3-48af-94ff-63e0a1608fc4\") " pod="openshift-infra/auto-csr-approver-29533384-9q4l8" Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.430013 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvl6s\" (UniqueName: \"kubernetes.io/projected/28930518-6de3-48af-94ff-63e0a1608fc4-kube-api-access-hvl6s\") pod \"auto-csr-approver-29533384-9q4l8\" (UID: \"28930518-6de3-48af-94ff-63e0a1608fc4\") " pod="openshift-infra/auto-csr-approver-29533384-9q4l8" Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.475690 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533384-9q4l8" Feb 25 07:04:00 crc kubenswrapper[4978]: I0225 07:04:00.968488 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533384-9q4l8"] Feb 25 07:04:01 crc kubenswrapper[4978]: I0225 07:04:01.956533 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533384-9q4l8" event={"ID":"28930518-6de3-48af-94ff-63e0a1608fc4","Type":"ContainerStarted","Data":"65a3e7c0120a2d4034073727b9e26d92443668d7d014b6536fe5c73e4b781921"} Feb 25 07:04:02 crc kubenswrapper[4978]: I0225 07:04:02.963004 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533384-9q4l8" event={"ID":"28930518-6de3-48af-94ff-63e0a1608fc4","Type":"ContainerStarted","Data":"52e71a0bce55e4c42a5a657ce506d03913c188711c74ade67560c87430666754"} Feb 25 07:04:02 crc kubenswrapper[4978]: I0225 07:04:02.986792 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533384-9q4l8" podStartSLOduration=1.42384338 podStartE2EDuration="2.986776176s" podCreationTimestamp="2026-02-25 07:04:00 +0000 UTC" firstStartedPulling="2026-02-25 07:04:00.9779029 +0000 UTC m=+1134.417159349" lastFinishedPulling="2026-02-25 07:04:02.540835696 +0000 UTC m=+1135.980092145" observedRunningTime="2026-02-25 07:04:02.98532592 +0000 UTC m=+1136.424582379" watchObservedRunningTime="2026-02-25 07:04:02.986776176 +0000 UTC m=+1136.426032635" Feb 25 07:04:03 crc kubenswrapper[4978]: I0225 07:04:03.969144 4978 generic.go:334] "Generic (PLEG): container finished" podID="28930518-6de3-48af-94ff-63e0a1608fc4" containerID="52e71a0bce55e4c42a5a657ce506d03913c188711c74ade67560c87430666754" exitCode=0 Feb 25 07:04:03 crc kubenswrapper[4978]: I0225 07:04:03.969238 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533384-9q4l8" event={"ID":"28930518-6de3-48af-94ff-63e0a1608fc4","Type":"ContainerDied","Data":"52e71a0bce55e4c42a5a657ce506d03913c188711c74ade67560c87430666754"} Feb 25 07:04:05 crc kubenswrapper[4978]: I0225 07:04:05.413221 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533384-9q4l8" Feb 25 07:04:05 crc kubenswrapper[4978]: I0225 07:04:05.562936 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvl6s\" (UniqueName: \"kubernetes.io/projected/28930518-6de3-48af-94ff-63e0a1608fc4-kube-api-access-hvl6s\") pod \"28930518-6de3-48af-94ff-63e0a1608fc4\" (UID: \"28930518-6de3-48af-94ff-63e0a1608fc4\") " Feb 25 07:04:05 crc kubenswrapper[4978]: I0225 07:04:05.572330 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28930518-6de3-48af-94ff-63e0a1608fc4-kube-api-access-hvl6s" (OuterVolumeSpecName: "kube-api-access-hvl6s") pod "28930518-6de3-48af-94ff-63e0a1608fc4" (UID: "28930518-6de3-48af-94ff-63e0a1608fc4"). InnerVolumeSpecName "kube-api-access-hvl6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:04:05 crc kubenswrapper[4978]: I0225 07:04:05.664825 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvl6s\" (UniqueName: \"kubernetes.io/projected/28930518-6de3-48af-94ff-63e0a1608fc4-kube-api-access-hvl6s\") on node \"crc\" DevicePath \"\"" Feb 25 07:04:05 crc kubenswrapper[4978]: I0225 07:04:05.985469 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533384-9q4l8" event={"ID":"28930518-6de3-48af-94ff-63e0a1608fc4","Type":"ContainerDied","Data":"65a3e7c0120a2d4034073727b9e26d92443668d7d014b6536fe5c73e4b781921"} Feb 25 07:04:05 crc kubenswrapper[4978]: I0225 07:04:05.985514 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65a3e7c0120a2d4034073727b9e26d92443668d7d014b6536fe5c73e4b781921" Feb 25 07:04:05 crc kubenswrapper[4978]: I0225 07:04:05.985630 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533384-9q4l8" Feb 25 07:04:06 crc kubenswrapper[4978]: I0225 07:04:06.072749 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533378-wmlm2"] Feb 25 07:04:06 crc kubenswrapper[4978]: I0225 07:04:06.079500 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533378-wmlm2"] Feb 25 07:04:07 crc kubenswrapper[4978]: I0225 07:04:07.349544 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09a0d1ce-3f33-46a6-98a9-a0a2337e7a33" path="/var/lib/kubelet/pods/09a0d1ce-3f33-46a6-98a9-a0a2337e7a33/volumes" Feb 25 07:04:14 crc kubenswrapper[4978]: I0225 07:04:14.930882 4978 scope.go:117] "RemoveContainer" containerID="d5c59d5bffe4a5b678647ead046fd39dc8f8128c1eb6627de9e03bd857a35439" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.419128 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g"] Feb 25 07:04:15 crc kubenswrapper[4978]: E0225 07:04:15.419729 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28930518-6de3-48af-94ff-63e0a1608fc4" containerName="oc" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.419743 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="28930518-6de3-48af-94ff-63e0a1608fc4" containerName="oc" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.419853 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="28930518-6de3-48af-94ff-63e0a1608fc4" containerName="oc" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.420324 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.422612 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.423334 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.425540 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-hj7zd" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.425706 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-qnlxb" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.425970 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.427790 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.431978 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-lwslg" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.437877 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.438779 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.440952 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-6qtbb" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.441623 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.445078 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.460203 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.470223 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.487449 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.488228 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.490426 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-jtk96" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.506122 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.514130 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.515025 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.518202 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-56vdr" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.534813 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.535795 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.539843 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-fn7tr" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.543450 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.543861 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.547666 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.562228 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.563030 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.572052 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-sjgnm" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.579410 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.595490 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.597110 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.606114 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-f6xc9" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.609486 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.611908 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.617731 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh968\" (UniqueName: \"kubernetes.io/projected/b81c87fe-0c25-4ff3-b962-d1bab3e3a91a-kube-api-access-nh968\") pod \"heat-operator-controller-manager-69f49c598c-gmzk4\" (UID: \"b81c87fe-0c25-4ff3-b962-d1bab3e3a91a\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.617958 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-687sm\" (UniqueName: \"kubernetes.io/projected/42305c1a-c5e6-467b-9f92-75b822ede523-kube-api-access-687sm\") pod \"barbican-operator-controller-manager-868647ff47-xr2cd\" (UID: \"42305c1a-c5e6-467b-9f92-75b822ede523\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.618045 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtgjp\" (UniqueName: \"kubernetes.io/projected/2a48bde4-1ad7-41fd-937e-223509b1b712-kube-api-access-dtgjp\") pod \"glance-operator-controller-manager-784b5bb6c5-rsf6k\" (UID: \"2a48bde4-1ad7-41fd-937e-223509b1b712\") " pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.618161 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc6t5\" (UniqueName: \"kubernetes.io/projected/6dc1c1a6-79f3-48bd-8eab-574e65c9152f-kube-api-access-qc6t5\") pod \"designate-operator-controller-manager-6d8bf5c495-jwrjk\" (UID: \"6dc1c1a6-79f3-48bd-8eab-574e65c9152f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.624144 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n97z6\" (UniqueName: \"kubernetes.io/projected/43011269-2605-4cc7-9475-46996e0210e3-kube-api-access-n97z6\") pod \"cinder-operator-controller-manager-55d77d7b5c-qk97g\" (UID: \"43011269-2605-4cc7-9475-46996e0210e3\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.624207 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlr4l\" (UniqueName: \"kubernetes.io/projected/6299ec19-6e59-472a-a09c-a698bbc547e8-kube-api-access-vlr4l\") pod \"horizon-operator-controller-manager-5b9b8895d5-whqp4\" (UID: \"6299ec19-6e59-472a-a09c-a698bbc547e8\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.626169 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-d9898" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.626819 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.639258 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.651625 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7kk5w" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.656269 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.696493 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.707906 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.724912 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh968\" (UniqueName: \"kubernetes.io/projected/b81c87fe-0c25-4ff3-b962-d1bab3e3a91a-kube-api-access-nh968\") pod \"heat-operator-controller-manager-69f49c598c-gmzk4\" (UID: \"b81c87fe-0c25-4ff3-b962-d1bab3e3a91a\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.724989 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz79m\" (UniqueName: \"kubernetes.io/projected/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-kube-api-access-sz79m\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.725030 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-687sm\" (UniqueName: \"kubernetes.io/projected/42305c1a-c5e6-467b-9f92-75b822ede523-kube-api-access-687sm\") pod \"barbican-operator-controller-manager-868647ff47-xr2cd\" (UID: \"42305c1a-c5e6-467b-9f92-75b822ede523\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.725064 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtgjp\" (UniqueName: \"kubernetes.io/projected/2a48bde4-1ad7-41fd-937e-223509b1b712-kube-api-access-dtgjp\") pod \"glance-operator-controller-manager-784b5bb6c5-rsf6k\" (UID: \"2a48bde4-1ad7-41fd-937e-223509b1b712\") " pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.725093 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.725123 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbknl\" (UniqueName: \"kubernetes.io/projected/02395a30-3d75-4d54-b057-422d43427f1c-kube-api-access-cbknl\") pod \"mariadb-operator-controller-manager-6994f66f48-c2vsw\" (UID: \"02395a30-3d75-4d54-b057-422d43427f1c\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.725158 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d4wv\" (UniqueName: \"kubernetes.io/projected/62fb8662-3720-4e7c-a556-08d9b441d1fe-kube-api-access-2d4wv\") pod \"manila-operator-controller-manager-67d996989d-bqwwg\" (UID: \"62fb8662-3720-4e7c-a556-08d9b441d1fe\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.725200 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znmdn\" (UniqueName: \"kubernetes.io/projected/2baccbb6-56c4-49ff-aa9b-d5f08b1c6b48-kube-api-access-znmdn\") pod \"keystone-operator-controller-manager-b4d948c87-7xrm6\" (UID: \"2baccbb6-56c4-49ff-aa9b-d5f08b1c6b48\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.725236 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc6t5\" (UniqueName: \"kubernetes.io/projected/6dc1c1a6-79f3-48bd-8eab-574e65c9152f-kube-api-access-qc6t5\") pod \"designate-operator-controller-manager-6d8bf5c495-jwrjk\" (UID: \"6dc1c1a6-79f3-48bd-8eab-574e65c9152f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.725264 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n97z6\" (UniqueName: \"kubernetes.io/projected/43011269-2605-4cc7-9475-46996e0210e3-kube-api-access-n97z6\") pod \"cinder-operator-controller-manager-55d77d7b5c-qk97g\" (UID: \"43011269-2605-4cc7-9475-46996e0210e3\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.725288 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7wqz\" (UniqueName: \"kubernetes.io/projected/3f1862db-5ef7-4749-b7a2-1ee2d65dd3c6-kube-api-access-p7wqz\") pod \"ironic-operator-controller-manager-554564d7fc-xw8bv\" (UID: \"3f1862db-5ef7-4749-b7a2-1ee2d65dd3c6\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.725309 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlr4l\" (UniqueName: \"kubernetes.io/projected/6299ec19-6e59-472a-a09c-a698bbc547e8-kube-api-access-vlr4l\") pod \"horizon-operator-controller-manager-5b9b8895d5-whqp4\" (UID: \"6299ec19-6e59-472a-a09c-a698bbc547e8\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.740439 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.741364 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.748767 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-djhrv" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.749207 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlr4l\" (UniqueName: \"kubernetes.io/projected/6299ec19-6e59-472a-a09c-a698bbc547e8-kube-api-access-vlr4l\") pod \"horizon-operator-controller-manager-5b9b8895d5-whqp4\" (UID: \"6299ec19-6e59-472a-a09c-a698bbc547e8\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.766691 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.767858 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.774220 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc6t5\" (UniqueName: \"kubernetes.io/projected/6dc1c1a6-79f3-48bd-8eab-574e65c9152f-kube-api-access-qc6t5\") pod \"designate-operator-controller-manager-6d8bf5c495-jwrjk\" (UID: \"6dc1c1a6-79f3-48bd-8eab-574e65c9152f\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.774336 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-687sm\" (UniqueName: \"kubernetes.io/projected/42305c1a-c5e6-467b-9f92-75b822ede523-kube-api-access-687sm\") pod \"barbican-operator-controller-manager-868647ff47-xr2cd\" (UID: \"42305c1a-c5e6-467b-9f92-75b822ede523\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.774963 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-rsb7c" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.776551 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh968\" (UniqueName: \"kubernetes.io/projected/b81c87fe-0c25-4ff3-b962-d1bab3e3a91a-kube-api-access-nh968\") pod \"heat-operator-controller-manager-69f49c598c-gmzk4\" (UID: \"b81c87fe-0c25-4ff3-b962-d1bab3e3a91a\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.778009 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.779531 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n97z6\" (UniqueName: \"kubernetes.io/projected/43011269-2605-4cc7-9475-46996e0210e3-kube-api-access-n97z6\") pod \"cinder-operator-controller-manager-55d77d7b5c-qk97g\" (UID: \"43011269-2605-4cc7-9475-46996e0210e3\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.789544 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.790886 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.792160 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtgjp\" (UniqueName: \"kubernetes.io/projected/2a48bde4-1ad7-41fd-937e-223509b1b712-kube-api-access-dtgjp\") pod \"glance-operator-controller-manager-784b5bb6c5-rsf6k\" (UID: \"2a48bde4-1ad7-41fd-937e-223509b1b712\") " pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.792586 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.795425 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-jlx4v" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.811641 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.813526 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.819697 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.820487 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.824222 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-fmlzc" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.825651 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mdm2\" (UniqueName: \"kubernetes.io/projected/19b7fc05-08e1-4982-9248-4f380e7733e6-kube-api-access-7mdm2\") pod \"neutron-operator-controller-manager-6bd4687957-wwpmj\" (UID: \"19b7fc05-08e1-4982-9248-4f380e7733e6\") " pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.825696 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7wqz\" (UniqueName: \"kubernetes.io/projected/3f1862db-5ef7-4749-b7a2-1ee2d65dd3c6-kube-api-access-p7wqz\") pod \"ironic-operator-controller-manager-554564d7fc-xw8bv\" (UID: \"3f1862db-5ef7-4749-b7a2-1ee2d65dd3c6\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.825715 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzfdh\" (UniqueName: \"kubernetes.io/projected/d1a081d5-4069-4ee0-9f2f-66460514ebe6-kube-api-access-qzfdh\") pod \"nova-operator-controller-manager-567668f5cf-grcw6\" (UID: \"d1a081d5-4069-4ee0-9f2f-66460514ebe6\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.825834 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz79m\" (UniqueName: \"kubernetes.io/projected/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-kube-api-access-sz79m\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.825880 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcmfk\" (UniqueName: \"kubernetes.io/projected/72b3dc0a-de07-42e9-b0b6-0897aa5c17c0-kube-api-access-vcmfk\") pod \"octavia-operator-controller-manager-659dc6bbfc-pwgkf\" (UID: \"72b3dc0a-de07-42e9-b0b6-0897aa5c17c0\") " pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.825924 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sfmh\" (UniqueName: \"kubernetes.io/projected/3a6484c8-775f-4c06-817b-2f89501e1c71-kube-api-access-4sfmh\") pod \"ovn-operator-controller-manager-5955d8c787-vdnbw\" (UID: \"3a6484c8-775f-4c06-817b-2f89501e1c71\") " pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.825949 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.825983 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbknl\" (UniqueName: \"kubernetes.io/projected/02395a30-3d75-4d54-b057-422d43427f1c-kube-api-access-cbknl\") pod \"mariadb-operator-controller-manager-6994f66f48-c2vsw\" (UID: \"02395a30-3d75-4d54-b057-422d43427f1c\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.826030 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d4wv\" (UniqueName: \"kubernetes.io/projected/62fb8662-3720-4e7c-a556-08d9b441d1fe-kube-api-access-2d4wv\") pod \"manila-operator-controller-manager-67d996989d-bqwwg\" (UID: \"62fb8662-3720-4e7c-a556-08d9b441d1fe\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.826090 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znmdn\" (UniqueName: \"kubernetes.io/projected/2baccbb6-56c4-49ff-aa9b-d5f08b1c6b48-kube-api-access-znmdn\") pod \"keystone-operator-controller-manager-b4d948c87-7xrm6\" (UID: \"2baccbb6-56c4-49ff-aa9b-d5f08b1c6b48\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6" Feb 25 07:04:15 crc kubenswrapper[4978]: E0225 07:04:15.826422 4978 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:15 crc kubenswrapper[4978]: E0225 07:04:15.826464 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert podName:6aa1040d-b0e6-412b-ad79-67e0d2e90b63 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:16.326446509 +0000 UTC m=+1149.765702968 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert") pod "infra-operator-controller-manager-79d975b745-8k5bx" (UID: "6aa1040d-b0e6-412b-ad79-67e0d2e90b63") : secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.830238 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.833432 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.836599 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-5r4pk" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.838484 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.839319 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.843681 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-cdjgs" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.843896 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.847814 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-9752r"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.848654 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.850489 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.851327 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7wqz\" (UniqueName: \"kubernetes.io/projected/3f1862db-5ef7-4749-b7a2-1ee2d65dd3c6-kube-api-access-p7wqz\") pod \"ironic-operator-controller-manager-554564d7fc-xw8bv\" (UID: \"3f1862db-5ef7-4749-b7a2-1ee2d65dd3c6\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.851764 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-q5drw" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.848442 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d4wv\" (UniqueName: \"kubernetes.io/projected/62fb8662-3720-4e7c-a556-08d9b441d1fe-kube-api-access-2d4wv\") pod \"manila-operator-controller-manager-67d996989d-bqwwg\" (UID: \"62fb8662-3720-4e7c-a556-08d9b441d1fe\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.859111 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znmdn\" (UniqueName: \"kubernetes.io/projected/2baccbb6-56c4-49ff-aa9b-d5f08b1c6b48-kube-api-access-znmdn\") pod \"keystone-operator-controller-manager-b4d948c87-7xrm6\" (UID: \"2baccbb6-56c4-49ff-aa9b-d5f08b1c6b48\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.859621 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbknl\" (UniqueName: \"kubernetes.io/projected/02395a30-3d75-4d54-b057-422d43427f1c-kube-api-access-cbknl\") pod \"mariadb-operator-controller-manager-6994f66f48-c2vsw\" (UID: \"02395a30-3d75-4d54-b057-422d43427f1c\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.870453 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.877685 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.883077 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz79m\" (UniqueName: \"kubernetes.io/projected/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-kube-api-access-sz79m\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.883249 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.897002 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.901294 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-9752r"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.927789 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q"] Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.928014 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbqqc\" (UniqueName: \"kubernetes.io/projected/fc4de411-fd98-4a38-b7a9-1137da7e8064-kube-api-access-dbqqc\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.928179 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.928232 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mdm2\" (UniqueName: \"kubernetes.io/projected/19b7fc05-08e1-4982-9248-4f380e7733e6-kube-api-access-7mdm2\") pod \"neutron-operator-controller-manager-6bd4687957-wwpmj\" (UID: \"19b7fc05-08e1-4982-9248-4f380e7733e6\") " pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.928302 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzfdh\" (UniqueName: \"kubernetes.io/projected/d1a081d5-4069-4ee0-9f2f-66460514ebe6-kube-api-access-qzfdh\") pod \"nova-operator-controller-manager-567668f5cf-grcw6\" (UID: \"d1a081d5-4069-4ee0-9f2f-66460514ebe6\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.928434 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcmfk\" (UniqueName: \"kubernetes.io/projected/72b3dc0a-de07-42e9-b0b6-0897aa5c17c0-kube-api-access-vcmfk\") pod \"octavia-operator-controller-manager-659dc6bbfc-pwgkf\" (UID: \"72b3dc0a-de07-42e9-b0b6-0897aa5c17c0\") " pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.928598 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sfmh\" (UniqueName: \"kubernetes.io/projected/3a6484c8-775f-4c06-817b-2f89501e1c71-kube-api-access-4sfmh\") pod \"ovn-operator-controller-manager-5955d8c787-vdnbw\" (UID: \"3a6484c8-775f-4c06-817b-2f89501e1c71\") " pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.928825 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.931763 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-6mmtj" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.955201 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mdm2\" (UniqueName: \"kubernetes.io/projected/19b7fc05-08e1-4982-9248-4f380e7733e6-kube-api-access-7mdm2\") pod \"neutron-operator-controller-manager-6bd4687957-wwpmj\" (UID: \"19b7fc05-08e1-4982-9248-4f380e7733e6\") " pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.956803 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcmfk\" (UniqueName: \"kubernetes.io/projected/72b3dc0a-de07-42e9-b0b6-0897aa5c17c0-kube-api-access-vcmfk\") pod \"octavia-operator-controller-manager-659dc6bbfc-pwgkf\" (UID: \"72b3dc0a-de07-42e9-b0b6-0897aa5c17c0\") " pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.960265 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzfdh\" (UniqueName: \"kubernetes.io/projected/d1a081d5-4069-4ee0-9f2f-66460514ebe6-kube-api-access-qzfdh\") pod \"nova-operator-controller-manager-567668f5cf-grcw6\" (UID: \"d1a081d5-4069-4ee0-9f2f-66460514ebe6\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.961951 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sfmh\" (UniqueName: \"kubernetes.io/projected/3a6484c8-775f-4c06-817b-2f89501e1c71-kube-api-access-4sfmh\") pod \"ovn-operator-controller-manager-5955d8c787-vdnbw\" (UID: \"3a6484c8-775f-4c06-817b-2f89501e1c71\") " pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw" Feb 25 07:04:15 crc kubenswrapper[4978]: I0225 07:04:15.972635 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.004593 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.004714 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.014545 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.029203 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.031040 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.032434 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.032498 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6vhv\" (UniqueName: \"kubernetes.io/projected/555963fa-06fb-4dd4-854c-9dc86a6dec12-kube-api-access-f6vhv\") pod \"swift-operator-controller-manager-68f46476f-qjw8s\" (UID: \"555963fa-06fb-4dd4-854c-9dc86a6dec12\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.032517 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq2zd\" (UniqueName: \"kubernetes.io/projected/29e606cc-8ae3-4065-8cf1-182f40ab4b98-kube-api-access-qq2zd\") pod \"placement-operator-controller-manager-8497b45c89-9752r\" (UID: \"29e606cc-8ae3-4065-8cf1-182f40ab4b98\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.032535 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntbzk\" (UniqueName: \"kubernetes.io/projected/b17d0775-1ea1-43b3-9813-a58f12645b64-kube-api-access-ntbzk\") pod \"telemetry-operator-controller-manager-589c568786-hpd2q\" (UID: \"b17d0775-1ea1-43b3-9813-a58f12645b64\") " pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.032582 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbqqc\" (UniqueName: \"kubernetes.io/projected/fc4de411-fd98-4a38-b7a9-1137da7e8064-kube-api-access-dbqqc\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.032827 4978 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.032867 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert podName:fc4de411-fd98-4a38-b7a9-1137da7e8064 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:16.532852618 +0000 UTC m=+1149.972109077 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert") pod "openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" (UID: "fc4de411-fd98-4a38-b7a9-1137da7e8064") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.036819 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bck94" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.040698 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.044036 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.050658 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.065690 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.070619 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.074947 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbqqc\" (UniqueName: \"kubernetes.io/projected/fc4de411-fd98-4a38-b7a9-1137da7e8064-kube-api-access-dbqqc\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.098789 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.100100 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.105171 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-6nt57" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.125909 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.132654 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.139541 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk962\" (UniqueName: \"kubernetes.io/projected/5de9178c-622e-4c57-93a8-dcda89354d5e-kube-api-access-jk962\") pod \"test-operator-controller-manager-5dc6794d5b-8xm7d\" (UID: \"5de9178c-622e-4c57-93a8-dcda89354d5e\") " pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.139612 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6vhv\" (UniqueName: \"kubernetes.io/projected/555963fa-06fb-4dd4-854c-9dc86a6dec12-kube-api-access-f6vhv\") pod \"swift-operator-controller-manager-68f46476f-qjw8s\" (UID: \"555963fa-06fb-4dd4-854c-9dc86a6dec12\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.139645 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq2zd\" (UniqueName: \"kubernetes.io/projected/29e606cc-8ae3-4065-8cf1-182f40ab4b98-kube-api-access-qq2zd\") pod \"placement-operator-controller-manager-8497b45c89-9752r\" (UID: \"29e606cc-8ae3-4065-8cf1-182f40ab4b98\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.139791 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntbzk\" (UniqueName: \"kubernetes.io/projected/b17d0775-1ea1-43b3-9813-a58f12645b64-kube-api-access-ntbzk\") pod \"telemetry-operator-controller-manager-589c568786-hpd2q\" (UID: \"b17d0775-1ea1-43b3-9813-a58f12645b64\") " pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.161834 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.187897 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.202176 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.215185 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntbzk\" (UniqueName: \"kubernetes.io/projected/b17d0775-1ea1-43b3-9813-a58f12645b64-kube-api-access-ntbzk\") pod \"telemetry-operator-controller-manager-589c568786-hpd2q\" (UID: \"b17d0775-1ea1-43b3-9813-a58f12645b64\") " pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.221632 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6vhv\" (UniqueName: \"kubernetes.io/projected/555963fa-06fb-4dd4-854c-9dc86a6dec12-kube-api-access-f6vhv\") pod \"swift-operator-controller-manager-68f46476f-qjw8s\" (UID: \"555963fa-06fb-4dd4-854c-9dc86a6dec12\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.222913 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq2zd\" (UniqueName: \"kubernetes.io/projected/29e606cc-8ae3-4065-8cf1-182f40ab4b98-kube-api-access-qq2zd\") pod \"placement-operator-controller-manager-8497b45c89-9752r\" (UID: \"29e606cc-8ae3-4065-8cf1-182f40ab4b98\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.224877 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.255302 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.265908 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrkxv\" (UniqueName: \"kubernetes.io/projected/03f30036-f34a-4751-a133-4ed9cd390ac9-kube-api-access-nrkxv\") pod \"watcher-operator-controller-manager-bccc79885-njfb8\" (UID: \"03f30036-f34a-4751-a133-4ed9cd390ac9\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.266001 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk962\" (UniqueName: \"kubernetes.io/projected/5de9178c-622e-4c57-93a8-dcda89354d5e-kube-api-access-jk962\") pod \"test-operator-controller-manager-5dc6794d5b-8xm7d\" (UID: \"5de9178c-622e-4c57-93a8-dcda89354d5e\") " pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.270991 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.273788 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.275547 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.293192 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.293801 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.294002 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-kbhvj" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.294122 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.305112 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk962\" (UniqueName: \"kubernetes.io/projected/5de9178c-622e-4c57-93a8-dcda89354d5e-kube-api-access-jk962\") pod \"test-operator-controller-manager-5dc6794d5b-8xm7d\" (UID: \"5de9178c-622e-4c57-93a8-dcda89354d5e\") " pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.356155 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.357129 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.363666 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-96nf9" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.367910 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.367990 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.368019 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrkxv\" (UniqueName: \"kubernetes.io/projected/03f30036-f34a-4751-a133-4ed9cd390ac9-kube-api-access-nrkxv\") pod \"watcher-operator-controller-manager-bccc79885-njfb8\" (UID: \"03f30036-f34a-4751-a133-4ed9cd390ac9\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.368026 4978 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.368095 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert podName:6aa1040d-b0e6-412b-ad79-67e0d2e90b63 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:17.368068112 +0000 UTC m=+1150.807324661 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert") pod "infra-operator-controller-manager-79d975b745-8k5bx" (UID: "6aa1040d-b0e6-412b-ad79-67e0d2e90b63") : secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.368122 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6xqs\" (UniqueName: \"kubernetes.io/projected/5d604e16-a85f-4bca-a322-218306711870-kube-api-access-g6xqs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.368159 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz7dx\" (UniqueName: \"kubernetes.io/projected/af52fa17-14f1-4ff3-885e-12fa43af933d-kube-api-access-bz7dx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-77v5j\" (UID: \"af52fa17-14f1-4ff3-885e-12fa43af933d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.368199 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.388720 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.411227 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrkxv\" (UniqueName: \"kubernetes.io/projected/03f30036-f34a-4751-a133-4ed9cd390ac9-kube-api-access-nrkxv\") pod \"watcher-operator-controller-manager-bccc79885-njfb8\" (UID: \"03f30036-f34a-4751-a133-4ed9cd390ac9\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.422926 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.432699 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.469528 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.469605 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6xqs\" (UniqueName: \"kubernetes.io/projected/5d604e16-a85f-4bca-a322-218306711870-kube-api-access-g6xqs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.469635 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz7dx\" (UniqueName: \"kubernetes.io/projected/af52fa17-14f1-4ff3-885e-12fa43af933d-kube-api-access-bz7dx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-77v5j\" (UID: \"af52fa17-14f1-4ff3-885e-12fa43af933d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.469660 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.469803 4978 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.469854 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:16.969837907 +0000 UTC m=+1150.409094356 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "webhook-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.470207 4978 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.470280 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:16.970263951 +0000 UTC m=+1150.409520410 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "metrics-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.491747 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6xqs\" (UniqueName: \"kubernetes.io/projected/5d604e16-a85f-4bca-a322-218306711870-kube-api-access-g6xqs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.493121 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz7dx\" (UniqueName: \"kubernetes.io/projected/af52fa17-14f1-4ff3-885e-12fa43af933d-kube-api-access-bz7dx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-77v5j\" (UID: \"af52fa17-14f1-4ff3-885e-12fa43af933d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.540172 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.540236 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.575269 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.575495 4978 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.575952 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert podName:fc4de411-fd98-4a38-b7a9-1137da7e8064 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:17.575927948 +0000 UTC m=+1151.015184407 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert") pod "openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" (UID: "fc4de411-fd98-4a38-b7a9-1137da7e8064") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.623551 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.655489 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4"] Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.766109 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.984169 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:16 crc kubenswrapper[4978]: I0225 07:04:16.984277 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.984467 4978 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.984528 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:17.984510225 +0000 UTC m=+1151.423766684 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "metrics-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.984864 4978 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 07:04:16 crc kubenswrapper[4978]: E0225 07:04:16.984888 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:17.984880657 +0000 UTC m=+1151.424137116 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "webhook-server-cert" not found Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.064657 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.103074 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.109520 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv"] Feb 25 07:04:17 crc kubenswrapper[4978]: W0225 07:04:17.120587 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02395a30_3d75_4d54_b057_422d43427f1c.slice/crio-a43cb401a2e74dec9a49c15e7844914eb39e363894fe5c72c55caa308b3bd034 WatchSource:0}: Error finding container a43cb401a2e74dec9a49c15e7844914eb39e363894fe5c72c55caa308b3bd034: Status 404 returned error can't find the container with id a43cb401a2e74dec9a49c15e7844914eb39e363894fe5c72c55caa308b3bd034 Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.157092 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.173276 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.179292 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw"] Feb 25 07:04:17 crc kubenswrapper[4978]: W0225 07:04:17.182197 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72b3dc0a_de07_42e9_b0b6_0897aa5c17c0.slice/crio-ef46182ea0c6dd405e49c3e69690b64973b45c0a8626b991222e83bb82befae7 WatchSource:0}: Error finding container ef46182ea0c6dd405e49c3e69690b64973b45c0a8626b991222e83bb82befae7: Status 404 returned error can't find the container with id ef46182ea0c6dd405e49c3e69690b64973b45c0a8626b991222e83bb82befae7 Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.192771 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.201929 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.325507 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv" event={"ID":"3f1862db-5ef7-4749-b7a2-1ee2d65dd3c6","Type":"ContainerStarted","Data":"6fb8129f5a6bceefe8c254a541e37f943b964b95e203e7483b2deeec53184623"} Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.326248 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6" event={"ID":"2baccbb6-56c4-49ff-aa9b-d5f08b1c6b48","Type":"ContainerStarted","Data":"be8c1dcb251f14226f7bf6271f4a90a964c39bfca819f676856066f88b95f6ab"} Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.339582 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf" event={"ID":"72b3dc0a-de07-42e9-b0b6-0897aa5c17c0","Type":"ContainerStarted","Data":"ef46182ea0c6dd405e49c3e69690b64973b45c0a8626b991222e83bb82befae7"} Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.339603 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4" event={"ID":"6299ec19-6e59-472a-a09c-a698bbc547e8","Type":"ContainerStarted","Data":"2533e77f5eef628a8f9d4a03c8308f5763376b2b8d2b8e8936f8cd9384fac62f"} Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.339613 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4" event={"ID":"b81c87fe-0c25-4ff3-b962-d1bab3e3a91a","Type":"ContainerStarted","Data":"4dbfcc37c2450c566a92ff306996480feaf78753663913be88cbc67833719fe1"} Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.339623 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg" event={"ID":"62fb8662-3720-4e7c-a556-08d9b441d1fe","Type":"ContainerStarted","Data":"cf7fecf7ba32b561d0282b7328be3ddd22e7cdba8c8cef2e07a04e5a96679a6c"} Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.339633 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw" event={"ID":"02395a30-3d75-4d54-b057-422d43427f1c","Type":"ContainerStarted","Data":"a43cb401a2e74dec9a49c15e7844914eb39e363894fe5c72c55caa308b3bd034"} Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.339644 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj" event={"ID":"19b7fc05-08e1-4982-9248-4f380e7733e6","Type":"ContainerStarted","Data":"bd4c1eab9c006ac9ccf201b9a02f4a73069ac5d2850e6223d37a536885f6892b"} Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.339898 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g" event={"ID":"43011269-2605-4cc7-9475-46996e0210e3","Type":"ContainerStarted","Data":"c7600f3db1eced36b092ca10e09c6f4ed682206d77a8b46991f7678d5eac614f"} Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.340637 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw" event={"ID":"3a6484c8-775f-4c06-817b-2f89501e1c71","Type":"ContainerStarted","Data":"d5b2d03fec48eeb2793401e38502a33da028a7e9140436700406337d9e2addaa"} Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.389240 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.389452 4978 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.389493 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert podName:6aa1040d-b0e6-412b-ad79-67e0d2e90b63 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:19.389481408 +0000 UTC m=+1152.828737867 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert") pod "infra-operator-controller-manager-79d975b745-8k5bx" (UID: "6aa1040d-b0e6-412b-ad79-67e0d2e90b63") : secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.561891 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.563490 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.566987 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.592519 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.592642 4978 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.592684 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert podName:fc4de411-fd98-4a38-b7a9-1137da7e8064 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:19.592671708 +0000 UTC m=+1153.031928167 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert") pod "openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" (UID: "fc4de411-fd98-4a38-b7a9-1137da7e8064") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.597425 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.605054 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-9752r"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.609488 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk"] Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.614074 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s"] Feb 25 07:04:17 crc kubenswrapper[4978]: W0225 07:04:17.616590 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5de9178c_622e_4c57_93a8_dcda89354d5e.slice/crio-6600adba0ca7650bc94cdf9ebbdcb292d9101e46ab221ca86c6c26b35eb1b46d WatchSource:0}: Error finding container 6600adba0ca7650bc94cdf9ebbdcb292d9101e46ab221ca86c6c26b35eb1b46d: Status 404 returned error can't find the container with id 6600adba0ca7650bc94cdf9ebbdcb292d9101e46ab221ca86c6c26b35eb1b46d Feb 25 07:04:17 crc kubenswrapper[4978]: W0225 07:04:17.617652 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb17d0775_1ea1_43b3_9813_a58f12645b64.slice/crio-edd6488a0b9a263e2876bc194c55e51481115d5e29cfe2245095039a78c31e56 WatchSource:0}: Error finding container edd6488a0b9a263e2876bc194c55e51481115d5e29cfe2245095039a78c31e56: Status 404 returned error can't find the container with id edd6488a0b9a263e2876bc194c55e51481115d5e29cfe2245095039a78c31e56 Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.618452 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d"] Feb 25 07:04:17 crc kubenswrapper[4978]: W0225 07:04:17.619182 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6dc1c1a6_79f3_48bd_8eab_574e65c9152f.slice/crio-8239abd8a7cd2ce3ab67b71dc76cd93ba8b58d613af2bd8afeae82d78cfceae0 WatchSource:0}: Error finding container 8239abd8a7cd2ce3ab67b71dc76cd93ba8b58d613af2bd8afeae82d78cfceae0: Status 404 returned error can't find the container with id 8239abd8a7cd2ce3ab67b71dc76cd93ba8b58d613af2bd8afeae82d78cfceae0 Feb 25 07:04:17 crc kubenswrapper[4978]: W0225 07:04:17.623663 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod555963fa_06fb_4dd4_854c_9dc86a6dec12.slice/crio-98088021ba54f909a90af41b276f3066febc2ae409e9ce41bbd1349de4195ecc WatchSource:0}: Error finding container 98088021ba54f909a90af41b276f3066febc2ae409e9ce41bbd1349de4195ecc: Status 404 returned error can't find the container with id 98088021ba54f909a90af41b276f3066febc2ae409e9ce41bbd1349de4195ecc Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.624469 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qc6t5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-6d8bf5c495-jwrjk_openstack-operators(6dc1c1a6-79f3-48bd-8eab-574e65c9152f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.624472 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qq2zd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8497b45c89-9752r_openstack-operators(29e606cc-8ae3-4065-8cf1-182f40ab4b98): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.624522 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j"] Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.624501 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:4eb8fab5530a08915d3ab3e11e2808aeae16c8a220ed34ee04a186b2ae2303dc,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ntbzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-589c568786-hpd2q_openstack-operators(b17d0775-1ea1-43b3-9813-a58f12645b64): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.625542 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" podUID="6dc1c1a6-79f3-48bd-8eab-574e65c9152f" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.625598 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" podUID="b17d0775-1ea1-43b3-9813-a58f12645b64" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.625652 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" podUID="29e606cc-8ae3-4065-8cf1-182f40ab4b98" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.627229 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f6vhv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-qjw8s_openstack-operators(555963fa-06fb-4dd4-854c-9dc86a6dec12): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.628748 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" podUID="555963fa-06fb-4dd4-854c-9dc86a6dec12" Feb 25 07:04:17 crc kubenswrapper[4978]: W0225 07:04:17.636416 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03f30036_f34a_4751_a133_4ed9cd390ac9.slice/crio-d32853cbede2522d6037dada4b50dc714f226897f6e3729237a0fe94030d3b3b WatchSource:0}: Error finding container d32853cbede2522d6037dada4b50dc714f226897f6e3729237a0fe94030d3b3b: Status 404 returned error can't find the container with id d32853cbede2522d6037dada4b50dc714f226897f6e3729237a0fe94030d3b3b Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.636957 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8"] Feb 25 07:04:17 crc kubenswrapper[4978]: W0225 07:04:17.639487 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf52fa17_14f1_4ff3_885e_12fa43af933d.slice/crio-5e611c67536e2018ba41623ab7d52b4154f3e8d5157242de55538574eb787964 WatchSource:0}: Error finding container 5e611c67536e2018ba41623ab7d52b4154f3e8d5157242de55538574eb787964: Status 404 returned error can't find the container with id 5e611c67536e2018ba41623ab7d52b4154f3e8d5157242de55538574eb787964 Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.641603 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nrkxv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-bccc79885-njfb8_openstack-operators(03f30036-f34a-4751-a133-4ed9cd390ac9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.641864 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bz7dx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-77v5j_openstack-operators(af52fa17-14f1-4ff3-885e-12fa43af933d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.642776 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" podUID="03f30036-f34a-4751-a133-4ed9cd390ac9" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.643010 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" podUID="af52fa17-14f1-4ff3-885e-12fa43af933d" Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.998590 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.998774 4978 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.999025 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:19.999005995 +0000 UTC m=+1153.438262454 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "webhook-server-cert" not found Feb 25 07:04:17 crc kubenswrapper[4978]: I0225 07:04:17.998953 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.999035 4978 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 07:04:17 crc kubenswrapper[4978]: E0225 07:04:17.999273 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:19.999240992 +0000 UTC m=+1153.438497451 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "metrics-server-cert" not found Feb 25 07:04:18 crc kubenswrapper[4978]: I0225 07:04:18.358978 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" event={"ID":"af52fa17-14f1-4ff3-885e-12fa43af933d","Type":"ContainerStarted","Data":"5e611c67536e2018ba41623ab7d52b4154f3e8d5157242de55538574eb787964"} Feb 25 07:04:18 crc kubenswrapper[4978]: E0225 07:04:18.360627 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" podUID="af52fa17-14f1-4ff3-885e-12fa43af933d" Feb 25 07:04:18 crc kubenswrapper[4978]: I0225 07:04:18.361556 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" event={"ID":"b17d0775-1ea1-43b3-9813-a58f12645b64","Type":"ContainerStarted","Data":"edd6488a0b9a263e2876bc194c55e51481115d5e29cfe2245095039a78c31e56"} Feb 25 07:04:18 crc kubenswrapper[4978]: I0225 07:04:18.364566 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" event={"ID":"6dc1c1a6-79f3-48bd-8eab-574e65c9152f","Type":"ContainerStarted","Data":"8239abd8a7cd2ce3ab67b71dc76cd93ba8b58d613af2bd8afeae82d78cfceae0"} Feb 25 07:04:18 crc kubenswrapper[4978]: E0225 07:04:18.364920 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:4eb8fab5530a08915d3ab3e11e2808aeae16c8a220ed34ee04a186b2ae2303dc\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" podUID="b17d0775-1ea1-43b3-9813-a58f12645b64" Feb 25 07:04:18 crc kubenswrapper[4978]: E0225 07:04:18.366119 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642\\\"\"" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" podUID="6dc1c1a6-79f3-48bd-8eab-574e65c9152f" Feb 25 07:04:18 crc kubenswrapper[4978]: I0225 07:04:18.366317 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6" event={"ID":"d1a081d5-4069-4ee0-9f2f-66460514ebe6","Type":"ContainerStarted","Data":"f8fac429da94e653ef327a73b4f93b16fdf7780a1477ffb75f30d84da8c0abaa"} Feb 25 07:04:18 crc kubenswrapper[4978]: I0225 07:04:18.369538 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d" event={"ID":"5de9178c-622e-4c57-93a8-dcda89354d5e","Type":"ContainerStarted","Data":"6600adba0ca7650bc94cdf9ebbdcb292d9101e46ab221ca86c6c26b35eb1b46d"} Feb 25 07:04:18 crc kubenswrapper[4978]: I0225 07:04:18.374253 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k" event={"ID":"2a48bde4-1ad7-41fd-937e-223509b1b712","Type":"ContainerStarted","Data":"0af0263e6c7fd79bf418fb3feb8353b76989a7969a21cfdc8e11607814607662"} Feb 25 07:04:18 crc kubenswrapper[4978]: I0225 07:04:18.377275 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" event={"ID":"03f30036-f34a-4751-a133-4ed9cd390ac9","Type":"ContainerStarted","Data":"d32853cbede2522d6037dada4b50dc714f226897f6e3729237a0fe94030d3b3b"} Feb 25 07:04:18 crc kubenswrapper[4978]: I0225 07:04:18.379840 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd" event={"ID":"42305c1a-c5e6-467b-9f92-75b822ede523","Type":"ContainerStarted","Data":"f01ee0da013a43ad6fb77228dc1c51b4685d5ab29a7545165d15f8e0a0f1f904"} Feb 25 07:04:18 crc kubenswrapper[4978]: I0225 07:04:18.381275 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" event={"ID":"29e606cc-8ae3-4065-8cf1-182f40ab4b98","Type":"ContainerStarted","Data":"68996cf1426985c479262e61300d5151b2f6fc3c52b0f42890ab8a66c67f4244"} Feb 25 07:04:18 crc kubenswrapper[4978]: E0225 07:04:18.381473 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" podUID="03f30036-f34a-4751-a133-4ed9cd390ac9" Feb 25 07:04:18 crc kubenswrapper[4978]: I0225 07:04:18.383395 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" event={"ID":"555963fa-06fb-4dd4-854c-9dc86a6dec12","Type":"ContainerStarted","Data":"98088021ba54f909a90af41b276f3066febc2ae409e9ce41bbd1349de4195ecc"} Feb 25 07:04:18 crc kubenswrapper[4978]: E0225 07:04:18.384622 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" podUID="29e606cc-8ae3-4065-8cf1-182f40ab4b98" Feb 25 07:04:18 crc kubenswrapper[4978]: E0225 07:04:18.390207 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" podUID="555963fa-06fb-4dd4-854c-9dc86a6dec12" Feb 25 07:04:19 crc kubenswrapper[4978]: E0225 07:04:19.395586 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642\\\"\"" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" podUID="6dc1c1a6-79f3-48bd-8eab-574e65c9152f" Feb 25 07:04:19 crc kubenswrapper[4978]: E0225 07:04:19.395594 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:4eb8fab5530a08915d3ab3e11e2808aeae16c8a220ed34ee04a186b2ae2303dc\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" podUID="b17d0775-1ea1-43b3-9813-a58f12645b64" Feb 25 07:04:19 crc kubenswrapper[4978]: E0225 07:04:19.395711 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" podUID="af52fa17-14f1-4ff3-885e-12fa43af933d" Feb 25 07:04:19 crc kubenswrapper[4978]: E0225 07:04:19.395745 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" podUID="03f30036-f34a-4751-a133-4ed9cd390ac9" Feb 25 07:04:19 crc kubenswrapper[4978]: E0225 07:04:19.395907 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" podUID="29e606cc-8ae3-4065-8cf1-182f40ab4b98" Feb 25 07:04:19 crc kubenswrapper[4978]: E0225 07:04:19.412605 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" podUID="555963fa-06fb-4dd4-854c-9dc86a6dec12" Feb 25 07:04:19 crc kubenswrapper[4978]: I0225 07:04:19.439958 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:19 crc kubenswrapper[4978]: E0225 07:04:19.441100 4978 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:19 crc kubenswrapper[4978]: E0225 07:04:19.441262 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert podName:6aa1040d-b0e6-412b-ad79-67e0d2e90b63 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:23.441227011 +0000 UTC m=+1156.880483460 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert") pod "infra-operator-controller-manager-79d975b745-8k5bx" (UID: "6aa1040d-b0e6-412b-ad79-67e0d2e90b63") : secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:19 crc kubenswrapper[4978]: I0225 07:04:19.642144 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:19 crc kubenswrapper[4978]: E0225 07:04:19.642336 4978 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:19 crc kubenswrapper[4978]: E0225 07:04:19.642395 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert podName:fc4de411-fd98-4a38-b7a9-1137da7e8064 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:23.642380957 +0000 UTC m=+1157.081637416 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert") pod "openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" (UID: "fc4de411-fd98-4a38-b7a9-1137da7e8064") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:20 crc kubenswrapper[4978]: I0225 07:04:20.054444 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:20 crc kubenswrapper[4978]: E0225 07:04:20.054650 4978 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 07:04:20 crc kubenswrapper[4978]: E0225 07:04:20.054725 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:24.054705681 +0000 UTC m=+1157.493962140 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "metrics-server-cert" not found Feb 25 07:04:20 crc kubenswrapper[4978]: I0225 07:04:20.055399 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:20 crc kubenswrapper[4978]: E0225 07:04:20.055478 4978 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 07:04:20 crc kubenswrapper[4978]: E0225 07:04:20.055512 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:24.055503516 +0000 UTC m=+1157.494759975 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "webhook-server-cert" not found Feb 25 07:04:23 crc kubenswrapper[4978]: I0225 07:04:23.510770 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:23 crc kubenswrapper[4978]: E0225 07:04:23.511296 4978 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:23 crc kubenswrapper[4978]: E0225 07:04:23.511449 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert podName:6aa1040d-b0e6-412b-ad79-67e0d2e90b63 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:31.511412462 +0000 UTC m=+1164.950668961 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert") pod "infra-operator-controller-manager-79d975b745-8k5bx" (UID: "6aa1040d-b0e6-412b-ad79-67e0d2e90b63") : secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:23 crc kubenswrapper[4978]: I0225 07:04:23.714529 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:23 crc kubenswrapper[4978]: E0225 07:04:23.714704 4978 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:23 crc kubenswrapper[4978]: E0225 07:04:23.714789 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert podName:fc4de411-fd98-4a38-b7a9-1137da7e8064 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:31.714764906 +0000 UTC m=+1165.154021455 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert") pod "openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" (UID: "fc4de411-fd98-4a38-b7a9-1137da7e8064") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:24 crc kubenswrapper[4978]: I0225 07:04:24.118236 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:24 crc kubenswrapper[4978]: I0225 07:04:24.118348 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:24 crc kubenswrapper[4978]: E0225 07:04:24.118350 4978 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 07:04:24 crc kubenswrapper[4978]: E0225 07:04:24.118432 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:32.118414009 +0000 UTC m=+1165.557670468 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "webhook-server-cert" not found Feb 25 07:04:24 crc kubenswrapper[4978]: E0225 07:04:24.118495 4978 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 07:04:24 crc kubenswrapper[4978]: E0225 07:04:24.118532 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:32.118521772 +0000 UTC m=+1165.557778231 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "metrics-server-cert" not found Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.329071 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.523734 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv" event={"ID":"3f1862db-5ef7-4749-b7a2-1ee2d65dd3c6","Type":"ContainerStarted","Data":"39bf69daab09b6480f982c1cb354bfc95c95022cb9739ea85bbc07b8719d5b74"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.523847 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.525185 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd" event={"ID":"42305c1a-c5e6-467b-9f92-75b822ede523","Type":"ContainerStarted","Data":"5977e5c73387644bd37a1c12ce6475c172450f040a644a7b43270648bb3d0b6d"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.525307 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.526184 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6" event={"ID":"d1a081d5-4069-4ee0-9f2f-66460514ebe6","Type":"ContainerStarted","Data":"a0019bb3d2792c543778e87bf5999bd72a2298a1b263ad6221f490eb02fe522c"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.526600 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.527553 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg" event={"ID":"62fb8662-3720-4e7c-a556-08d9b441d1fe","Type":"ContainerStarted","Data":"874b45f15298480f02f11596077b38704cb71c7369409311b9a41948781f5e8d"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.528088 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.541758 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4" event={"ID":"b81c87fe-0c25-4ff3-b962-d1bab3e3a91a","Type":"ContainerStarted","Data":"fb43d6d20c61f7c87e3fb5a25ffeaa67194da5d277ef1df8e7bd60ea4954f331"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.542823 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.552837 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw" event={"ID":"3a6484c8-775f-4c06-817b-2f89501e1c71","Type":"ContainerStarted","Data":"94c6fc12d7c7f25977e6342c6c37c088dae0558268070f402c1623cd3f7bc022"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.553565 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.554125 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4" event={"ID":"6299ec19-6e59-472a-a09c-a698bbc547e8","Type":"ContainerStarted","Data":"f2973e74a236b9e68390ec82f2d223d9e9520b95107089b9fc31ff3a74db8704"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.554294 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.555322 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj" event={"ID":"19b7fc05-08e1-4982-9248-4f380e7733e6","Type":"ContainerStarted","Data":"76fd1f265fd31a4c7717ce4a7ae7847df86cc301b780b252933f0a604363678e"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.558628 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g" event={"ID":"43011269-2605-4cc7-9475-46996e0210e3","Type":"ContainerStarted","Data":"3b426ed21d439e05e1900fe376a39ee48d90dd0d6aecce365a26a595de0abf65"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.559019 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.559857 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.571595 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d" event={"ID":"5de9178c-622e-4c57-93a8-dcda89354d5e","Type":"ContainerStarted","Data":"22404248cb67f811a4c0cbdcef3b2665e4a3952bc985bbecf321d6f5dfde7fe4"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.572815 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.587495 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k" event={"ID":"2a48bde4-1ad7-41fd-937e-223509b1b712","Type":"ContainerStarted","Data":"eb62ad87d3c243a08650d5e77ed78268616bc37e2d31881e18bb26ce82fc08ca"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.587639 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.590206 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv" podStartSLOduration=3.506103514 podStartE2EDuration="15.590172106s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.126622486 +0000 UTC m=+1150.565878945" lastFinishedPulling="2026-02-25 07:04:29.210691068 +0000 UTC m=+1162.649947537" observedRunningTime="2026-02-25 07:04:30.562654502 +0000 UTC m=+1164.001910961" watchObservedRunningTime="2026-02-25 07:04:30.590172106 +0000 UTC m=+1164.029428565" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.601105 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw" event={"ID":"02395a30-3d75-4d54-b057-422d43427f1c","Type":"ContainerStarted","Data":"482a98ecba96a4c28bc09f7b01e1a3288caf9b0aadb7d91df92a1aebab1611bc"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.602080 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.609541 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg" podStartSLOduration=3.57924066 podStartE2EDuration="15.609510373s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.178487285 +0000 UTC m=+1150.617743754" lastFinishedPulling="2026-02-25 07:04:29.208756998 +0000 UTC m=+1162.648013467" observedRunningTime="2026-02-25 07:04:30.587343967 +0000 UTC m=+1164.026600426" watchObservedRunningTime="2026-02-25 07:04:30.609510373 +0000 UTC m=+1164.048766832" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.609763 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6" event={"ID":"2baccbb6-56c4-49ff-aa9b-d5f08b1c6b48","Type":"ContainerStarted","Data":"3b401ac493a720c60de7a04dbf4f3f917fb5d1f61c7a2c049c3c82cf3f4d9d0e"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.610309 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.612332 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf" event={"ID":"72b3dc0a-de07-42e9-b0b6-0897aa5c17c0","Type":"ContainerStarted","Data":"5cf18176975ea70ff7bf59aed115abaa43111727d4f95ed771615bfc60bfff3c"} Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.612753 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.668835 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4" podStartSLOduration=3.247291238 podStartE2EDuration="15.668804635s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:16.714493957 +0000 UTC m=+1150.153750416" lastFinishedPulling="2026-02-25 07:04:29.136007354 +0000 UTC m=+1162.575263813" observedRunningTime="2026-02-25 07:04:30.637645326 +0000 UTC m=+1164.076901785" watchObservedRunningTime="2026-02-25 07:04:30.668804635 +0000 UTC m=+1164.108061094" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.672151 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd" podStartSLOduration=4.052778417 podStartE2EDuration="15.6721409s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.590766747 +0000 UTC m=+1151.030023206" lastFinishedPulling="2026-02-25 07:04:29.2101292 +0000 UTC m=+1162.649385689" observedRunningTime="2026-02-25 07:04:30.663738806 +0000 UTC m=+1164.102995265" watchObservedRunningTime="2026-02-25 07:04:30.6721409 +0000 UTC m=+1164.111397359" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.682257 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj" podStartSLOduration=3.689547203 podStartE2EDuration="15.682234406s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.189957915 +0000 UTC m=+1150.629214374" lastFinishedPulling="2026-02-25 07:04:29.182645078 +0000 UTC m=+1162.621901577" observedRunningTime="2026-02-25 07:04:30.680868914 +0000 UTC m=+1164.120125373" watchObservedRunningTime="2026-02-25 07:04:30.682234406 +0000 UTC m=+1164.121490865" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.746741 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6" podStartSLOduration=4.056498493 podStartE2EDuration="15.74671682s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.588734794 +0000 UTC m=+1151.027991253" lastFinishedPulling="2026-02-25 07:04:29.278953081 +0000 UTC m=+1162.718209580" observedRunningTime="2026-02-25 07:04:30.727150306 +0000 UTC m=+1164.166406765" watchObservedRunningTime="2026-02-25 07:04:30.74671682 +0000 UTC m=+1164.185973279" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.748920 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g" podStartSLOduration=3.71939776 podStartE2EDuration="15.748913729s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.1805669 +0000 UTC m=+1150.619823359" lastFinishedPulling="2026-02-25 07:04:29.210082839 +0000 UTC m=+1162.649339328" observedRunningTime="2026-02-25 07:04:30.742012013 +0000 UTC m=+1164.181268472" watchObservedRunningTime="2026-02-25 07:04:30.748913729 +0000 UTC m=+1164.188170188" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.758301 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4" podStartSLOduration=3.240503056 podStartE2EDuration="15.758287844s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:16.689689129 +0000 UTC m=+1150.128945588" lastFinishedPulling="2026-02-25 07:04:29.207473897 +0000 UTC m=+1162.646730376" observedRunningTime="2026-02-25 07:04:30.753884186 +0000 UTC m=+1164.193140645" watchObservedRunningTime="2026-02-25 07:04:30.758287844 +0000 UTC m=+1164.197544303" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.767665 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw" podStartSLOduration=3.789863842 podStartE2EDuration="15.767642718s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.187535888 +0000 UTC m=+1150.626792347" lastFinishedPulling="2026-02-25 07:04:29.165314724 +0000 UTC m=+1162.604571223" observedRunningTime="2026-02-25 07:04:30.76643176 +0000 UTC m=+1164.205688219" watchObservedRunningTime="2026-02-25 07:04:30.767642718 +0000 UTC m=+1164.206899177" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.783057 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf" podStartSLOduration=3.75985178 podStartE2EDuration="15.783033881s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.18759971 +0000 UTC m=+1150.626856169" lastFinishedPulling="2026-02-25 07:04:29.210781771 +0000 UTC m=+1162.650038270" observedRunningTime="2026-02-25 07:04:30.781791112 +0000 UTC m=+1164.221047571" watchObservedRunningTime="2026-02-25 07:04:30.783033881 +0000 UTC m=+1164.222290340" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.811431 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d" podStartSLOduration=4.265676151 podStartE2EDuration="15.811408812s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.619605563 +0000 UTC m=+1151.058862022" lastFinishedPulling="2026-02-25 07:04:29.165338184 +0000 UTC m=+1162.604594683" observedRunningTime="2026-02-25 07:04:30.804501665 +0000 UTC m=+1164.243758124" watchObservedRunningTime="2026-02-25 07:04:30.811408812 +0000 UTC m=+1164.250665271" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.842665 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k" podStartSLOduration=4.228352908 podStartE2EDuration="15.842639292s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.593482683 +0000 UTC m=+1151.032739142" lastFinishedPulling="2026-02-25 07:04:29.207769037 +0000 UTC m=+1162.647025526" observedRunningTime="2026-02-25 07:04:30.8345896 +0000 UTC m=+1164.273846069" watchObservedRunningTime="2026-02-25 07:04:30.842639292 +0000 UTC m=+1164.281895741" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.859168 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6" podStartSLOduration=3.707262009 podStartE2EDuration="15.8591455s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.076432691 +0000 UTC m=+1150.515689150" lastFinishedPulling="2026-02-25 07:04:29.228316152 +0000 UTC m=+1162.667572641" observedRunningTime="2026-02-25 07:04:30.856628701 +0000 UTC m=+1164.295885160" watchObservedRunningTime="2026-02-25 07:04:30.8591455 +0000 UTC m=+1164.298401959" Feb 25 07:04:30 crc kubenswrapper[4978]: I0225 07:04:30.881810 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw" podStartSLOduration=3.858167547 podStartE2EDuration="15.881777931s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.122194667 +0000 UTC m=+1150.561451116" lastFinishedPulling="2026-02-25 07:04:29.145805041 +0000 UTC m=+1162.585061500" observedRunningTime="2026-02-25 07:04:30.873850272 +0000 UTC m=+1164.313106751" watchObservedRunningTime="2026-02-25 07:04:30.881777931 +0000 UTC m=+1164.321034390" Feb 25 07:04:31 crc kubenswrapper[4978]: I0225 07:04:31.531671 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:31 crc kubenswrapper[4978]: E0225 07:04:31.531874 4978 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:31 crc kubenswrapper[4978]: E0225 07:04:31.532346 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert podName:6aa1040d-b0e6-412b-ad79-67e0d2e90b63 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:47.532307864 +0000 UTC m=+1180.971564323 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert") pod "infra-operator-controller-manager-79d975b745-8k5bx" (UID: "6aa1040d-b0e6-412b-ad79-67e0d2e90b63") : secret "infra-operator-webhook-server-cert" not found Feb 25 07:04:31 crc kubenswrapper[4978]: I0225 07:04:31.735161 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:31 crc kubenswrapper[4978]: E0225 07:04:31.735318 4978 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:31 crc kubenswrapper[4978]: E0225 07:04:31.735393 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert podName:fc4de411-fd98-4a38-b7a9-1137da7e8064 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:47.735361298 +0000 UTC m=+1181.174617757 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert") pod "openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" (UID: "fc4de411-fd98-4a38-b7a9-1137da7e8064") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 25 07:04:32 crc kubenswrapper[4978]: I0225 07:04:32.140450 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:32 crc kubenswrapper[4978]: I0225 07:04:32.140550 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:32 crc kubenswrapper[4978]: E0225 07:04:32.140744 4978 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 25 07:04:32 crc kubenswrapper[4978]: E0225 07:04:32.140808 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:48.140793597 +0000 UTC m=+1181.580050056 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "webhook-server-cert" not found Feb 25 07:04:32 crc kubenswrapper[4978]: E0225 07:04:32.140804 4978 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 25 07:04:32 crc kubenswrapper[4978]: E0225 07:04:32.140933 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs podName:5d604e16-a85f-4bca-a322-218306711870 nodeName:}" failed. No retries permitted until 2026-02-25 07:04:48.14090442 +0000 UTC m=+1181.580160919 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs") pod "openstack-operator-controller-manager-5dc486cffc-nwxz6" (UID: "5d604e16-a85f-4bca-a322-218306711870") : secret "metrics-server-cert" not found Feb 25 07:04:32 crc kubenswrapper[4978]: I0225 07:04:32.632747 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" event={"ID":"03f30036-f34a-4751-a133-4ed9cd390ac9","Type":"ContainerStarted","Data":"00cc8a266e88b9a0bdabdcc83ed7c487eb6fea99958b832e7f1275b40fb44538"} Feb 25 07:04:32 crc kubenswrapper[4978]: I0225 07:04:32.658238 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" podStartSLOduration=2.881215916 podStartE2EDuration="17.658217851s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.64148987 +0000 UTC m=+1151.080746319" lastFinishedPulling="2026-02-25 07:04:32.418491795 +0000 UTC m=+1165.857748254" observedRunningTime="2026-02-25 07:04:32.650264421 +0000 UTC m=+1166.089520890" watchObservedRunningTime="2026-02-25 07:04:32.658217851 +0000 UTC m=+1166.097474310" Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.660188 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" event={"ID":"555963fa-06fb-4dd4-854c-9dc86a6dec12","Type":"ContainerStarted","Data":"e6f1b8c7a186450aceb1b10f92d7a49e4e2580d7b329279f28ba5400957bfc97"} Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.660911 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.661861 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" event={"ID":"6dc1c1a6-79f3-48bd-8eab-574e65c9152f","Type":"ContainerStarted","Data":"35083326a45342925211052aa929dbdd4e325f71cf8c21d2d79fd4f7b1740e5f"} Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.662064 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.664657 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" event={"ID":"29e606cc-8ae3-4065-8cf1-182f40ab4b98","Type":"ContainerStarted","Data":"ca0624bd0cfcf2933c09c54cf305ecb3e5798203e3aa95414f084b1d1c131ad7"} Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.665203 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.677779 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" podStartSLOduration=3.183858017 podStartE2EDuration="20.677742637s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.627081247 +0000 UTC m=+1151.066337706" lastFinishedPulling="2026-02-25 07:04:35.120965857 +0000 UTC m=+1168.560222326" observedRunningTime="2026-02-25 07:04:35.674071941 +0000 UTC m=+1169.113328410" watchObservedRunningTime="2026-02-25 07:04:35.677742637 +0000 UTC m=+1169.116999106" Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.710779 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" podStartSLOduration=3.240377892 podStartE2EDuration="20.710757463s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.623697972 +0000 UTC m=+1151.062954431" lastFinishedPulling="2026-02-25 07:04:35.094077533 +0000 UTC m=+1168.533334002" observedRunningTime="2026-02-25 07:04:35.704798486 +0000 UTC m=+1169.144054955" watchObservedRunningTime="2026-02-25 07:04:35.710757463 +0000 UTC m=+1169.150013922" Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.725770 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" podStartSLOduration=3.255943881 podStartE2EDuration="20.725751444s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.623746733 +0000 UTC m=+1151.063003192" lastFinishedPulling="2026-02-25 07:04:35.093554306 +0000 UTC m=+1168.532810755" observedRunningTime="2026-02-25 07:04:35.720916352 +0000 UTC m=+1169.160172831" watchObservedRunningTime="2026-02-25 07:04:35.725751444 +0000 UTC m=+1169.165007903" Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.814730 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-gmzk4" Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.853090 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-whqp4" Feb 25 07:04:35 crc kubenswrapper[4978]: I0225 07:04:35.903915 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-xw8bv" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.007127 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-67d996989d-bqwwg" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.007558 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-7xrm6" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.017000 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-c2vsw" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.045873 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-qk97g" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.054350 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-xr2cd" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.082146 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-rsf6k" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.136440 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-wwpmj" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.165273 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-grcw6" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.193159 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-pwgkf" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.211221 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-vdnbw" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.425868 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-8xm7d" Feb 25 07:04:36 crc kubenswrapper[4978]: I0225 07:04:36.433972 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" Feb 25 07:04:37 crc kubenswrapper[4978]: I0225 07:04:37.684890 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" event={"ID":"b17d0775-1ea1-43b3-9813-a58f12645b64","Type":"ContainerStarted","Data":"32b98b3d0aa80fee978413c29432b3de97ed3e65fc6bbdb380a648061ce8ab56"} Feb 25 07:04:37 crc kubenswrapper[4978]: I0225 07:04:37.685651 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" Feb 25 07:04:37 crc kubenswrapper[4978]: I0225 07:04:37.713843 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" podStartSLOduration=3.802468019 podStartE2EDuration="22.713814838s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.624029972 +0000 UTC m=+1151.063286431" lastFinishedPulling="2026-02-25 07:04:36.535376791 +0000 UTC m=+1169.974633250" observedRunningTime="2026-02-25 07:04:37.706512398 +0000 UTC m=+1171.145768897" watchObservedRunningTime="2026-02-25 07:04:37.713814838 +0000 UTC m=+1171.153071337" Feb 25 07:04:41 crc kubenswrapper[4978]: I0225 07:04:41.717807 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" event={"ID":"af52fa17-14f1-4ff3-885e-12fa43af933d","Type":"ContainerStarted","Data":"e8d20b439a36ad21f55d848e508cd826e0f99b56b82decdc4183692407d75ba2"} Feb 25 07:04:41 crc kubenswrapper[4978]: I0225 07:04:41.742114 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-77v5j" podStartSLOduration=2.096217926 podStartE2EDuration="25.742094803s" podCreationTimestamp="2026-02-25 07:04:16 +0000 UTC" firstStartedPulling="2026-02-25 07:04:17.64180798 +0000 UTC m=+1151.081064439" lastFinishedPulling="2026-02-25 07:04:41.287684857 +0000 UTC m=+1174.726941316" observedRunningTime="2026-02-25 07:04:41.736685403 +0000 UTC m=+1175.175941922" watchObservedRunningTime="2026-02-25 07:04:41.742094803 +0000 UTC m=+1175.181351282" Feb 25 07:04:46 crc kubenswrapper[4978]: I0225 07:04:46.069866 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-jwrjk" Feb 25 07:04:46 crc kubenswrapper[4978]: I0225 07:04:46.230580 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-qjw8s" Feb 25 07:04:46 crc kubenswrapper[4978]: I0225 07:04:46.259423 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-hpd2q" Feb 25 07:04:46 crc kubenswrapper[4978]: I0225 07:04:46.277819 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-9752r" Feb 25 07:04:46 crc kubenswrapper[4978]: I0225 07:04:46.437964 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-njfb8" Feb 25 07:04:46 crc kubenswrapper[4978]: I0225 07:04:46.540482 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:04:46 crc kubenswrapper[4978]: I0225 07:04:46.540574 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:04:46 crc kubenswrapper[4978]: I0225 07:04:46.540644 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:04:46 crc kubenswrapper[4978]: I0225 07:04:46.541541 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"03d7d67e86736ca0a34d745efcdc9b40b54189fb150b16a7aa6067c654318766"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:04:46 crc kubenswrapper[4978]: I0225 07:04:46.541643 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://03d7d67e86736ca0a34d745efcdc9b40b54189fb150b16a7aa6067c654318766" gracePeriod=600 Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:46.773177 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="03d7d67e86736ca0a34d745efcdc9b40b54189fb150b16a7aa6067c654318766" exitCode=0 Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:46.773244 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"03d7d67e86736ca0a34d745efcdc9b40b54189fb150b16a7aa6067c654318766"} Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:46.773296 4978 scope.go:117] "RemoveContainer" containerID="ea6d2ab6806f2c754878e481d2101f535bcfdd355af23df81d835564bbd06358" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:47.574328 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:47.590561 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aa1040d-b0e6-412b-ad79-67e0d2e90b63-cert\") pod \"infra-operator-controller-manager-79d975b745-8k5bx\" (UID: \"6aa1040d-b0e6-412b-ad79-67e0d2e90b63\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:47.673823 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:47.778303 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:47.791360 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fc4de411-fd98-4a38-b7a9-1137da7e8064-cert\") pod \"openstack-baremetal-operator-controller-manager-579b7786b9ggmk5\" (UID: \"fc4de411-fd98-4a38-b7a9-1137da7e8064\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:48.030559 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:48.184349 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:48.184525 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:48.192474 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-metrics-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:48.193123 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/5d604e16-a85f-4bca-a322-218306711870-webhook-certs\") pod \"openstack-operator-controller-manager-5dc486cffc-nwxz6\" (UID: \"5d604e16-a85f-4bca-a322-218306711870\") " pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:49 crc kubenswrapper[4978]: I0225 07:04:48.441753 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:50 crc kubenswrapper[4978]: I0225 07:04:50.386855 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5"] Feb 25 07:04:50 crc kubenswrapper[4978]: W0225 07:04:50.393032 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc4de411_fd98_4a38_b7a9_1137da7e8064.slice/crio-46b81031d70665880055aebb7f03b064e8e85e3e4f74e469e5e2ddba7cbfa1a0 WatchSource:0}: Error finding container 46b81031d70665880055aebb7f03b064e8e85e3e4f74e469e5e2ddba7cbfa1a0: Status 404 returned error can't find the container with id 46b81031d70665880055aebb7f03b064e8e85e3e4f74e469e5e2ddba7cbfa1a0 Feb 25 07:04:50 crc kubenswrapper[4978]: I0225 07:04:50.396914 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6"] Feb 25 07:04:50 crc kubenswrapper[4978]: I0225 07:04:50.416109 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx"] Feb 25 07:04:50 crc kubenswrapper[4978]: I0225 07:04:50.826430 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" event={"ID":"6aa1040d-b0e6-412b-ad79-67e0d2e90b63","Type":"ContainerStarted","Data":"eff2ec7ab15baacb78d3d64a49f98458da182886379f4ef11842df97ad516655"} Feb 25 07:04:50 crc kubenswrapper[4978]: I0225 07:04:50.829113 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" event={"ID":"fc4de411-fd98-4a38-b7a9-1137da7e8064","Type":"ContainerStarted","Data":"46b81031d70665880055aebb7f03b064e8e85e3e4f74e469e5e2ddba7cbfa1a0"} Feb 25 07:04:50 crc kubenswrapper[4978]: I0225 07:04:50.831010 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" event={"ID":"5d604e16-a85f-4bca-a322-218306711870","Type":"ContainerStarted","Data":"211085049c41a4be809b8f34d2c60978434e16dbcacb3059473abbb88c978365"} Feb 25 07:04:51 crc kubenswrapper[4978]: I0225 07:04:51.849321 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" event={"ID":"5d604e16-a85f-4bca-a322-218306711870","Type":"ContainerStarted","Data":"fbbebcb3df80ffc50f936791c47b327e4adfedb51187c1c98d15abc16bbd803d"} Feb 25 07:04:51 crc kubenswrapper[4978]: I0225 07:04:51.849889 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:04:51 crc kubenswrapper[4978]: I0225 07:04:51.909032 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" podStartSLOduration=35.908994935 podStartE2EDuration="35.908994935s" podCreationTimestamp="2026-02-25 07:04:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:04:51.90308669 +0000 UTC m=+1185.342343249" watchObservedRunningTime="2026-02-25 07:04:51.908994935 +0000 UTC m=+1185.348251434" Feb 25 07:04:52 crc kubenswrapper[4978]: I0225 07:04:52.863176 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"b1d31d1a63c47b381a67200dff4a026477c582f264c877adf6984b366b252ba2"} Feb 25 07:04:54 crc kubenswrapper[4978]: I0225 07:04:54.879972 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" event={"ID":"6aa1040d-b0e6-412b-ad79-67e0d2e90b63","Type":"ContainerStarted","Data":"dd5c30ec7a08b07db389ca032f58e7bee33f9b5cdbb1639695ac917602a9a190"} Feb 25 07:04:54 crc kubenswrapper[4978]: I0225 07:04:54.880515 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:04:54 crc kubenswrapper[4978]: I0225 07:04:54.882286 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" event={"ID":"fc4de411-fd98-4a38-b7a9-1137da7e8064","Type":"ContainerStarted","Data":"e52028eba7e592097ea4253f5008fd9773dd1fead032db5f16e9f2201f2f867f"} Feb 25 07:04:54 crc kubenswrapper[4978]: I0225 07:04:54.882415 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:04:54 crc kubenswrapper[4978]: I0225 07:04:54.906212 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" podStartSLOduration=36.331284088 podStartE2EDuration="39.90619707s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:50.396662016 +0000 UTC m=+1183.835918475" lastFinishedPulling="2026-02-25 07:04:53.971574958 +0000 UTC m=+1187.410831457" observedRunningTime="2026-02-25 07:04:54.901032658 +0000 UTC m=+1188.340289117" watchObservedRunningTime="2026-02-25 07:04:54.90619707 +0000 UTC m=+1188.345453529" Feb 25 07:04:54 crc kubenswrapper[4978]: I0225 07:04:54.931220 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" podStartSLOduration=36.360401664 podStartE2EDuration="39.931198056s" podCreationTimestamp="2026-02-25 07:04:15 +0000 UTC" firstStartedPulling="2026-02-25 07:04:50.395956885 +0000 UTC m=+1183.835213344" lastFinishedPulling="2026-02-25 07:04:53.966753267 +0000 UTC m=+1187.406009736" observedRunningTime="2026-02-25 07:04:54.927449348 +0000 UTC m=+1188.366705847" watchObservedRunningTime="2026-02-25 07:04:54.931198056 +0000 UTC m=+1188.370454535" Feb 25 07:04:58 crc kubenswrapper[4978]: I0225 07:04:58.452898 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5dc486cffc-nwxz6" Feb 25 07:05:07 crc kubenswrapper[4978]: I0225 07:05:07.681979 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79d975b745-8k5bx" Feb 25 07:05:08 crc kubenswrapper[4978]: I0225 07:05:08.046408 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-579b7786b9ggmk5" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.273544 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-776868475c-ch4tm"] Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.286512 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-776868475c-ch4tm"] Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.286609 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-776868475c-ch4tm" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.289957 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.290357 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.290731 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-qk4cd" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.294404 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.327615 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32d9e2ad-9750-4062-a4a4-da0a072037de-config\") pod \"dnsmasq-dns-776868475c-ch4tm\" (UID: \"32d9e2ad-9750-4062-a4a4-da0a072037de\") " pod="openstack/dnsmasq-dns-776868475c-ch4tm" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.328012 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsn6k\" (UniqueName: \"kubernetes.io/projected/32d9e2ad-9750-4062-a4a4-da0a072037de-kube-api-access-hsn6k\") pod \"dnsmasq-dns-776868475c-ch4tm\" (UID: \"32d9e2ad-9750-4062-a4a4-da0a072037de\") " pod="openstack/dnsmasq-dns-776868475c-ch4tm" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.352960 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-648df8f4f9-d8vbv"] Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.354348 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.359564 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.366825 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-648df8f4f9-d8vbv"] Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.429151 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsn6k\" (UniqueName: \"kubernetes.io/projected/32d9e2ad-9750-4062-a4a4-da0a072037de-kube-api-access-hsn6k\") pod \"dnsmasq-dns-776868475c-ch4tm\" (UID: \"32d9e2ad-9750-4062-a4a4-da0a072037de\") " pod="openstack/dnsmasq-dns-776868475c-ch4tm" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.429217 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-dns-svc\") pod \"dnsmasq-dns-648df8f4f9-d8vbv\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.429243 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32d9e2ad-9750-4062-a4a4-da0a072037de-config\") pod \"dnsmasq-dns-776868475c-ch4tm\" (UID: \"32d9e2ad-9750-4062-a4a4-da0a072037de\") " pod="openstack/dnsmasq-dns-776868475c-ch4tm" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.429434 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-config\") pod \"dnsmasq-dns-648df8f4f9-d8vbv\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.429510 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pglj\" (UniqueName: \"kubernetes.io/projected/819a9975-6fc4-4fae-a817-9cc2043dc231-kube-api-access-9pglj\") pod \"dnsmasq-dns-648df8f4f9-d8vbv\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.430547 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32d9e2ad-9750-4062-a4a4-da0a072037de-config\") pod \"dnsmasq-dns-776868475c-ch4tm\" (UID: \"32d9e2ad-9750-4062-a4a4-da0a072037de\") " pod="openstack/dnsmasq-dns-776868475c-ch4tm" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.451772 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsn6k\" (UniqueName: \"kubernetes.io/projected/32d9e2ad-9750-4062-a4a4-da0a072037de-kube-api-access-hsn6k\") pod \"dnsmasq-dns-776868475c-ch4tm\" (UID: \"32d9e2ad-9750-4062-a4a4-da0a072037de\") " pod="openstack/dnsmasq-dns-776868475c-ch4tm" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.530560 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-dns-svc\") pod \"dnsmasq-dns-648df8f4f9-d8vbv\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.530697 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-config\") pod \"dnsmasq-dns-648df8f4f9-d8vbv\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.530736 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pglj\" (UniqueName: \"kubernetes.io/projected/819a9975-6fc4-4fae-a817-9cc2043dc231-kube-api-access-9pglj\") pod \"dnsmasq-dns-648df8f4f9-d8vbv\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.531561 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-dns-svc\") pod \"dnsmasq-dns-648df8f4f9-d8vbv\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.531636 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-config\") pod \"dnsmasq-dns-648df8f4f9-d8vbv\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.552925 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pglj\" (UniqueName: \"kubernetes.io/projected/819a9975-6fc4-4fae-a817-9cc2043dc231-kube-api-access-9pglj\") pod \"dnsmasq-dns-648df8f4f9-d8vbv\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.605712 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-776868475c-ch4tm" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.670116 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.912319 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-648df8f4f9-d8vbv"] Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.948246 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9fb6c8b7-lblhx"] Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.952067 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:26 crc kubenswrapper[4978]: I0225 07:05:26.962589 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9fb6c8b7-lblhx"] Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.054570 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-dns-svc\") pod \"dnsmasq-dns-5c9fb6c8b7-lblhx\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.054659 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2chpx\" (UniqueName: \"kubernetes.io/projected/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-kube-api-access-2chpx\") pod \"dnsmasq-dns-5c9fb6c8b7-lblhx\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.054693 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-config\") pod \"dnsmasq-dns-5c9fb6c8b7-lblhx\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.127766 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-776868475c-ch4tm"] Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.156033 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2chpx\" (UniqueName: \"kubernetes.io/projected/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-kube-api-access-2chpx\") pod \"dnsmasq-dns-5c9fb6c8b7-lblhx\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.156102 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-config\") pod \"dnsmasq-dns-5c9fb6c8b7-lblhx\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.156185 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-dns-svc\") pod \"dnsmasq-dns-5c9fb6c8b7-lblhx\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.157214 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-config\") pod \"dnsmasq-dns-5c9fb6c8b7-lblhx\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.157226 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-dns-svc\") pod \"dnsmasq-dns-5c9fb6c8b7-lblhx\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.176696 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2chpx\" (UniqueName: \"kubernetes.io/projected/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-kube-api-access-2chpx\") pod \"dnsmasq-dns-5c9fb6c8b7-lblhx\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.186120 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-776868475c-ch4tm" event={"ID":"32d9e2ad-9750-4062-a4a4-da0a072037de","Type":"ContainerStarted","Data":"78e0bfc37fc8b7efe27b3f4c6a178d38875603f6b1351673423b6a7bf36d1963"} Feb 25 07:05:27 crc kubenswrapper[4978]: W0225 07:05:27.223128 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod819a9975_6fc4_4fae_a817_9cc2043dc231.slice/crio-8e12e77349447e56b4ff1ca84957877d555559793bd5a647f225229e9d80e0ae WatchSource:0}: Error finding container 8e12e77349447e56b4ff1ca84957877d555559793bd5a647f225229e9d80e0ae: Status 404 returned error can't find the container with id 8e12e77349447e56b4ff1ca84957877d555559793bd5a647f225229e9d80e0ae Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.231508 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-648df8f4f9-d8vbv"] Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.285941 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.461433 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9fb6c8b7-lblhx"] Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.507204 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6dfb8ff55f-nhh4k"] Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.512454 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.534918 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dfb8ff55f-nhh4k"] Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.569581 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9xtg\" (UniqueName: \"kubernetes.io/projected/80d0ea7a-b122-4988-bff6-e4377b79b7a0-kube-api-access-v9xtg\") pod \"dnsmasq-dns-6dfb8ff55f-nhh4k\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.569663 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-dns-svc\") pod \"dnsmasq-dns-6dfb8ff55f-nhh4k\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.569696 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-config\") pod \"dnsmasq-dns-6dfb8ff55f-nhh4k\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.672613 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9xtg\" (UniqueName: \"kubernetes.io/projected/80d0ea7a-b122-4988-bff6-e4377b79b7a0-kube-api-access-v9xtg\") pod \"dnsmasq-dns-6dfb8ff55f-nhh4k\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.672723 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-dns-svc\") pod \"dnsmasq-dns-6dfb8ff55f-nhh4k\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.672764 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-config\") pod \"dnsmasq-dns-6dfb8ff55f-nhh4k\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.673868 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-config\") pod \"dnsmasq-dns-6dfb8ff55f-nhh4k\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.675484 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-dns-svc\") pod \"dnsmasq-dns-6dfb8ff55f-nhh4k\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.709684 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9xtg\" (UniqueName: \"kubernetes.io/projected/80d0ea7a-b122-4988-bff6-e4377b79b7a0-kube-api-access-v9xtg\") pod \"dnsmasq-dns-6dfb8ff55f-nhh4k\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.839167 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:05:27 crc kubenswrapper[4978]: I0225 07:05:27.859765 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9fb6c8b7-lblhx"] Feb 25 07:05:27 crc kubenswrapper[4978]: W0225 07:05:27.877783 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99bb43c1_17a0_467f_b5fe_e0092f58d0c3.slice/crio-4eeb334f5665034710ab18b485cf5ad9a80eb6c5f9850842041ff381df65a180 WatchSource:0}: Error finding container 4eeb334f5665034710ab18b485cf5ad9a80eb6c5f9850842041ff381df65a180: Status 404 returned error can't find the container with id 4eeb334f5665034710ab18b485cf5ad9a80eb6c5f9850842041ff381df65a180 Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.074755 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.076283 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.091653 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.091762 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.092132 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.092202 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-77djh" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.092293 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.092362 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.092502 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.107244 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.203591 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" event={"ID":"819a9975-6fc4-4fae-a817-9cc2043dc231","Type":"ContainerStarted","Data":"8e12e77349447e56b4ff1ca84957877d555559793bd5a647f225229e9d80e0ae"} Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.205343 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" event={"ID":"99bb43c1-17a0-467f-b5fe-e0092f58d0c3","Type":"ContainerStarted","Data":"4eeb334f5665034710ab18b485cf5ad9a80eb6c5f9850842041ff381df65a180"} Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.222800 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/09112202-a3ef-4742-bdbe-b46c1c9a55ba-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.222835 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.222855 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.222883 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.222913 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.222943 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.222965 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvzgx\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-kube-api-access-cvzgx\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.223182 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.224736 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/09112202-a3ef-4742-bdbe-b46c1c9a55ba-pod-info\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.224813 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.225015 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-server-conf\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.329300 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.329464 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.329515 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.329543 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvzgx\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-kube-api-access-cvzgx\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.329563 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.329695 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/09112202-a3ef-4742-bdbe-b46c1c9a55ba-pod-info\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.329753 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.329881 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-server-conf\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.330019 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.330052 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/09112202-a3ef-4742-bdbe-b46c1c9a55ba-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.330152 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.330181 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.330589 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.330922 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.331977 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.334464 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.334583 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-server-conf\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.339957 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/09112202-a3ef-4742-bdbe-b46c1c9a55ba-pod-info\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.340663 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/09112202-a3ef-4742-bdbe-b46c1c9a55ba-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.343909 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.349006 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvzgx\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-kube-api-access-cvzgx\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.359093 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.363176 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.372947 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dfb8ff55f-nhh4k"] Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.408044 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.750054 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.752711 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.755329 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4hlb2" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.755658 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.755727 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.755786 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.755825 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.755872 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.756045 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.767125 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.906302 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 07:05:28 crc kubenswrapper[4978]: W0225 07:05:28.938521 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09112202_a3ef_4742_bdbe_b46c1c9a55ba.slice/crio-b5b21db6b8c3246cc8774533366c1a40f3d7d9ddd47ef7ebb01557d48baaa599 WatchSource:0}: Error finding container b5b21db6b8c3246cc8774533366c1a40f3d7d9ddd47ef7ebb01557d48baaa599: Status 404 returned error can't find the container with id b5b21db6b8c3246cc8774533366c1a40f3d7d9ddd47ef7ebb01557d48baaa599 Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.941442 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.941485 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.941511 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.941942 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43bd8325-9408-4a32-9285-d522d497205e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.942058 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.942095 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnl5k\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-kube-api-access-dnl5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.942123 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.942213 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43bd8325-9408-4a32-9285-d522d497205e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.942351 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.942421 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:28 crc kubenswrapper[4978]: I0225 07:05:28.942540 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044608 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43bd8325-9408-4a32-9285-d522d497205e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044674 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044704 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnl5k\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-kube-api-access-dnl5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044727 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044759 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43bd8325-9408-4a32-9285-d522d497205e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044788 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044807 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044832 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044867 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044892 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.044921 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.045168 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.045989 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.046531 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.046568 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.046669 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.048003 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.050684 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.050927 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.052223 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43bd8325-9408-4a32-9285-d522d497205e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.058782 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43bd8325-9408-4a32-9285-d522d497205e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.061981 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnl5k\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-kube-api-access-dnl5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.076807 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.229725 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"09112202-a3ef-4742-bdbe-b46c1c9a55ba","Type":"ContainerStarted","Data":"b5b21db6b8c3246cc8774533366c1a40f3d7d9ddd47ef7ebb01557d48baaa599"} Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.232989 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" event={"ID":"80d0ea7a-b122-4988-bff6-e4377b79b7a0","Type":"ContainerStarted","Data":"47324e916a4f0ec55346b1b03c72b5b3c7ddd14344f58639eb11bf73d6d5a2e2"} Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.378400 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.907588 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.916039 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.916214 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.922635 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.924071 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.924529 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.925126 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-wphlj" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.927411 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 25 07:05:29 crc kubenswrapper[4978]: I0225 07:05:29.927635 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.060631 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.060677 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-default\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.060738 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-operator-scripts\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.060768 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-generated\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.060796 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.060821 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.060849 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-kolla-config\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.060883 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7gmq\" (UniqueName: \"kubernetes.io/projected/315eefbb-24fb-4e45-97ed-03caf894bc08-kube-api-access-t7gmq\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.162085 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.162157 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.162186 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-kolla-config\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.162221 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7gmq\" (UniqueName: \"kubernetes.io/projected/315eefbb-24fb-4e45-97ed-03caf894bc08-kube-api-access-t7gmq\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.162248 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.162267 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-default\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.162316 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-operator-scripts\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.162340 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-generated\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.162849 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-generated\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.164049 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-default\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.165898 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-operator-scripts\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.166189 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.169856 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.179699 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-kolla-config\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.182335 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7gmq\" (UniqueName: \"kubernetes.io/projected/315eefbb-24fb-4e45-97ed-03caf894bc08-kube-api-access-t7gmq\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.219334 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.221406 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.237011 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.251814 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43bd8325-9408-4a32-9285-d522d497205e","Type":"ContainerStarted","Data":"594b702495e86182b648ba4b63933e81dc4baf29486b60b748ddee3ac22df6be"} Feb 25 07:05:30 crc kubenswrapper[4978]: I0225 07:05:30.861922 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 25 07:05:30 crc kubenswrapper[4978]: W0225 07:05:30.874807 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod315eefbb_24fb_4e45_97ed_03caf894bc08.slice/crio-3a840466a68665264ab5537ac4787284f8dc1e57ebaeb06a96ec0e50cca2b4eb WatchSource:0}: Error finding container 3a840466a68665264ab5537ac4787284f8dc1e57ebaeb06a96ec0e50cca2b4eb: Status 404 returned error can't find the container with id 3a840466a68665264ab5537ac4787284f8dc1e57ebaeb06a96ec0e50cca2b4eb Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.269692 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"315eefbb-24fb-4e45-97ed-03caf894bc08","Type":"ContainerStarted","Data":"3a840466a68665264ab5537ac4787284f8dc1e57ebaeb06a96ec0e50cca2b4eb"} Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.431573 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.432673 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.434508 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.434701 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-h465g" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.435354 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.461256 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.469459 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.589868 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.589924 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.589969 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-596qq\" (UniqueName: \"kubernetes.io/projected/c488aaac-88dd-4eb1-9020-e17749814ed9-kube-api-access-596qq\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.589993 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.590011 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.590062 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.590084 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.590103 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.692141 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.692233 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.692314 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.692354 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.692397 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.692440 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.692473 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.692528 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-596qq\" (UniqueName: \"kubernetes.io/projected/c488aaac-88dd-4eb1-9020-e17749814ed9-kube-api-access-596qq\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.692770 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.693247 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.693267 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.693447 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.694350 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.700480 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.714491 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-596qq\" (UniqueName: \"kubernetes.io/projected/c488aaac-88dd-4eb1-9020-e17749814ed9-kube-api-access-596qq\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.715706 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.729995 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.776321 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.895517 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.897736 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.899787 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.899813 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-vlvfm" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.902686 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.919743 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.997663 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-combined-ca-bundle\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.997748 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-memcached-tls-certs\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.997812 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kolla-config\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.997889 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwgww\" (UniqueName: \"kubernetes.io/projected/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kube-api-access-fwgww\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:31 crc kubenswrapper[4978]: I0225 07:05:31.997912 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-config-data\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.099163 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwgww\" (UniqueName: \"kubernetes.io/projected/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kube-api-access-fwgww\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.099222 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-config-data\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.099283 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-combined-ca-bundle\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.099340 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-memcached-tls-certs\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.099405 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kolla-config\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.100936 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-config-data\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.103038 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kolla-config\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.106641 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-combined-ca-bundle\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.107973 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-memcached-tls-certs\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.117888 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwgww\" (UniqueName: \"kubernetes.io/projected/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kube-api-access-fwgww\") pod \"memcached-0\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.239619 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 25 07:05:32 crc kubenswrapper[4978]: I0225 07:05:32.433838 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 07:05:34 crc kubenswrapper[4978]: I0225 07:05:34.143454 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:05:34 crc kubenswrapper[4978]: I0225 07:05:34.144758 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 07:05:34 crc kubenswrapper[4978]: I0225 07:05:34.146823 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-qdsnh" Feb 25 07:05:34 crc kubenswrapper[4978]: I0225 07:05:34.149927 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:05:34 crc kubenswrapper[4978]: I0225 07:05:34.238251 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjnfp\" (UniqueName: \"kubernetes.io/projected/fbf80328-4523-49ff-b390-0e8b3dd9cc02-kube-api-access-bjnfp\") pod \"kube-state-metrics-0\" (UID: \"fbf80328-4523-49ff-b390-0e8b3dd9cc02\") " pod="openstack/kube-state-metrics-0" Feb 25 07:05:34 crc kubenswrapper[4978]: I0225 07:05:34.339524 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjnfp\" (UniqueName: \"kubernetes.io/projected/fbf80328-4523-49ff-b390-0e8b3dd9cc02-kube-api-access-bjnfp\") pod \"kube-state-metrics-0\" (UID: \"fbf80328-4523-49ff-b390-0e8b3dd9cc02\") " pod="openstack/kube-state-metrics-0" Feb 25 07:05:34 crc kubenswrapper[4978]: I0225 07:05:34.357992 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjnfp\" (UniqueName: \"kubernetes.io/projected/fbf80328-4523-49ff-b390-0e8b3dd9cc02-kube-api-access-bjnfp\") pod \"kube-state-metrics-0\" (UID: \"fbf80328-4523-49ff-b390-0e8b3dd9cc02\") " pod="openstack/kube-state-metrics-0" Feb 25 07:05:34 crc kubenswrapper[4978]: I0225 07:05:34.481992 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 07:05:36 crc kubenswrapper[4978]: I0225 07:05:36.362139 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c488aaac-88dd-4eb1-9020-e17749814ed9","Type":"ContainerStarted","Data":"58b5d1d02d020975779c0d2f0eae1010d9bf42641ae62f59a5cbe5a067fc115d"} Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.577449 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-r45vg"] Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.579200 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.592052 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.592088 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.592315 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-wxxfd" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.609023 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-njndh"] Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.610645 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.613596 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-r45vg"] Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.635058 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-njndh"] Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.699889 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-combined-ca-bundle\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.699942 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-log\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700074 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run-ovn\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700158 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmr2q\" (UniqueName: \"kubernetes.io/projected/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-kube-api-access-pmr2q\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700341 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd7s4\" (UniqueName: \"kubernetes.io/projected/01eac805-ca08-46de-b79b-0ca5645f6955-kube-api-access-zd7s4\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700655 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-log-ovn\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700726 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-run\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700743 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-lib\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700763 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-scripts\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700817 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-etc-ovs\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700852 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01eac805-ca08-46de-b79b-0ca5645f6955-scripts\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700893 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.700929 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-ovn-controller-tls-certs\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802533 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01eac805-ca08-46de-b79b-0ca5645f6955-scripts\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802634 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802685 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-ovn-controller-tls-certs\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802724 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-combined-ca-bundle\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802753 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-log\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802795 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run-ovn\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802851 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmr2q\" (UniqueName: \"kubernetes.io/projected/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-kube-api-access-pmr2q\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802913 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd7s4\" (UniqueName: \"kubernetes.io/projected/01eac805-ca08-46de-b79b-0ca5645f6955-kube-api-access-zd7s4\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802940 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-log-ovn\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802970 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-run\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.802988 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-lib\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.803006 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-scripts\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.803037 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-etc-ovs\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.803351 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-log\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.803441 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-run\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.803525 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-lib\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.803589 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-etc-ovs\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.803617 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-log-ovn\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.803623 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run-ovn\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.803636 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.805616 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01eac805-ca08-46de-b79b-0ca5645f6955-scripts\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.806113 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-scripts\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.808681 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-ovn-controller-tls-certs\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.808717 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-combined-ca-bundle\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.833188 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd7s4\" (UniqueName: \"kubernetes.io/projected/01eac805-ca08-46de-b79b-0ca5645f6955-kube-api-access-zd7s4\") pod \"ovn-controller-ovs-njndh\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.840049 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmr2q\" (UniqueName: \"kubernetes.io/projected/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-kube-api-access-pmr2q\") pod \"ovn-controller-r45vg\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.895349 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r45vg" Feb 25 07:05:37 crc kubenswrapper[4978]: I0225 07:05:37.953781 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.504260 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.507662 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.516691 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.516723 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.517048 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.517158 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.517515 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.527879 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-45nc6" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.617023 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.617086 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.617113 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-config\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.617128 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.617146 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.617169 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npq46\" (UniqueName: \"kubernetes.io/projected/ed9859d2-701a-4011-b9d6-d8cf6291265a-kube-api-access-npq46\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.617194 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.617263 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.718677 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.719256 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.719297 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.719319 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-config\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.719335 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.719360 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.719398 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npq46\" (UniqueName: \"kubernetes.io/projected/ed9859d2-701a-4011-b9d6-d8cf6291265a-kube-api-access-npq46\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.719419 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.720316 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.720881 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-config\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.721578 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.721899 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.725861 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.726914 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.727928 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.739408 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npq46\" (UniqueName: \"kubernetes.io/projected/ed9859d2-701a-4011-b9d6-d8cf6291265a-kube-api-access-npq46\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.758993 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:38 crc kubenswrapper[4978]: I0225 07:05:38.843738 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.606655 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.608984 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.612158 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.612485 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.612970 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-cbfvv" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.613190 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.618306 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.680157 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdpqt\" (UniqueName: \"kubernetes.io/projected/6df7a672-b889-40d3-abe5-271e4d36dd85-kube-api-access-fdpqt\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.680218 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.680255 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.680407 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.680445 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.680532 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.680603 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.680743 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-config\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.782184 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.782270 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.782343 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-config\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.782418 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdpqt\" (UniqueName: \"kubernetes.io/projected/6df7a672-b889-40d3-abe5-271e4d36dd85-kube-api-access-fdpqt\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.782440 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.782467 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.782496 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.782518 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.782915 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.784077 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.785413 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-config\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.786000 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.791003 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.791810 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.792442 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.807282 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdpqt\" (UniqueName: \"kubernetes.io/projected/6df7a672-b889-40d3-abe5-271e4d36dd85-kube-api-access-fdpqt\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.837751 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:41 crc kubenswrapper[4978]: I0225 07:05:41.954013 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 25 07:05:48 crc kubenswrapper[4978]: E0225 07:05:48.650228 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658" Feb 25 07:05:48 crc kubenswrapper[4978]: E0225 07:05:48.651327 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7gmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(315eefbb-24fb-4e45-97ed-03caf894bc08): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 07:05:48 crc kubenswrapper[4978]: E0225 07:05:48.652882 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="315eefbb-24fb-4e45-97ed-03caf894bc08" Feb 25 07:05:49 crc kubenswrapper[4978]: E0225 07:05:49.489783 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658\\\"\"" pod="openstack/openstack-galera-0" podUID="315eefbb-24fb-4e45-97ed-03caf894bc08" Feb 25 07:05:52 crc kubenswrapper[4978]: E0225 07:05:52.526747 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:ea8869786571e9ad2388af89ff4d38d887e32bc9340186598c63fe61a561eb20" Feb 25 07:05:52 crc kubenswrapper[4978]: E0225 07:05:52.526949 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:ea8869786571e9ad2388af89ff4d38d887e32bc9340186598c63fe61a561eb20,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cvzgx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(09112202-a3ef-4742-bdbe-b46c1c9a55ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 07:05:52 crc kubenswrapper[4978]: E0225 07:05:52.528413 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="09112202-a3ef-4742-bdbe-b46c1c9a55ba" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.453792 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c21a3e6f24adda8d9f7cfdb1115a43c928c3ee0ec263e331a215d9da533bbfcd" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.454166 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c21a3e6f24adda8d9f7cfdb1115a43c928c3ee0ec263e331a215d9da533bbfcd,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v9xtg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6dfb8ff55f-nhh4k_openstack(80d0ea7a-b122-4988-bff6-e4377b79b7a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.456017 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" podUID="80d0ea7a-b122-4988-bff6-e4377b79b7a0" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.462829 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c21a3e6f24adda8d9f7cfdb1115a43c928c3ee0ec263e331a215d9da533bbfcd" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.462982 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c21a3e6f24adda8d9f7cfdb1115a43c928c3ee0ec263e331a215d9da533bbfcd,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9pglj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-648df8f4f9-d8vbv_openstack(819a9975-6fc4-4fae-a817-9cc2043dc231): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.464305 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" podUID="819a9975-6fc4-4fae-a817-9cc2043dc231" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.467572 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c21a3e6f24adda8d9f7cfdb1115a43c928c3ee0ec263e331a215d9da533bbfcd" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.467749 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c21a3e6f24adda8d9f7cfdb1115a43c928c3ee0ec263e331a215d9da533bbfcd,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2chpx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5c9fb6c8b7-lblhx_openstack(99bb43c1-17a0-467f-b5fe-e0092f58d0c3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.468926 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" podUID="99bb43c1-17a0-467f-b5fe-e0092f58d0c3" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.474083 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c21a3e6f24adda8d9f7cfdb1115a43c928c3ee0ec263e331a215d9da533bbfcd" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.474196 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c21a3e6f24adda8d9f7cfdb1115a43c928c3ee0ec263e331a215d9da533bbfcd,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hsn6k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-776868475c-ch4tm_openstack(32d9e2ad-9750-4062-a4a4-da0a072037de): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.476145 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-776868475c-ch4tm" podUID="32d9e2ad-9750-4062-a4a4-da0a072037de" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.553412 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c21a3e6f24adda8d9f7cfdb1115a43c928c3ee0ec263e331a215d9da533bbfcd\\\"\"" pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" podUID="80d0ea7a-b122-4988-bff6-e4377b79b7a0" Feb 25 07:05:53 crc kubenswrapper[4978]: E0225 07:05:53.558574 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c21a3e6f24adda8d9f7cfdb1115a43c928c3ee0ec263e331a215d9da533bbfcd\\\"\"" pod="openstack/dnsmasq-dns-776868475c-ch4tm" podUID="32d9e2ad-9750-4062-a4a4-da0a072037de" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.028614 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.121579 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.134272 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.150250 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.185048 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-r45vg"] Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.191325 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.218299 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pglj\" (UniqueName: \"kubernetes.io/projected/819a9975-6fc4-4fae-a817-9cc2043dc231-kube-api-access-9pglj\") pod \"819a9975-6fc4-4fae-a817-9cc2043dc231\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.218466 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-config\") pod \"819a9975-6fc4-4fae-a817-9cc2043dc231\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.218594 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-dns-svc\") pod \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.218724 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2chpx\" (UniqueName: \"kubernetes.io/projected/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-kube-api-access-2chpx\") pod \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.218852 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-config\") pod \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\" (UID: \"99bb43c1-17a0-467f-b5fe-e0092f58d0c3\") " Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.218921 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-config" (OuterVolumeSpecName: "config") pod "819a9975-6fc4-4fae-a817-9cc2043dc231" (UID: "819a9975-6fc4-4fae-a817-9cc2043dc231"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.218963 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "99bb43c1-17a0-467f-b5fe-e0092f58d0c3" (UID: "99bb43c1-17a0-467f-b5fe-e0092f58d0c3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.219070 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-dns-svc\") pod \"819a9975-6fc4-4fae-a817-9cc2043dc231\" (UID: \"819a9975-6fc4-4fae-a817-9cc2043dc231\") " Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.219402 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-config" (OuterVolumeSpecName: "config") pod "99bb43c1-17a0-467f-b5fe-e0092f58d0c3" (UID: "99bb43c1-17a0-467f-b5fe-e0092f58d0c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.219511 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "819a9975-6fc4-4fae-a817-9cc2043dc231" (UID: "819a9975-6fc4-4fae-a817-9cc2043dc231"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.219649 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.219668 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.221663 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.224050 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/819a9975-6fc4-4fae-a817-9cc2043dc231-kube-api-access-9pglj" (OuterVolumeSpecName: "kube-api-access-9pglj") pod "819a9975-6fc4-4fae-a817-9cc2043dc231" (UID: "819a9975-6fc4-4fae-a817-9cc2043dc231"). InnerVolumeSpecName "kube-api-access-9pglj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.225490 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-kube-api-access-2chpx" (OuterVolumeSpecName: "kube-api-access-2chpx") pod "99bb43c1-17a0-467f-b5fe-e0092f58d0c3" (UID: "99bb43c1-17a0-467f-b5fe-e0092f58d0c3"). InnerVolumeSpecName "kube-api-access-2chpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.321751 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2chpx\" (UniqueName: \"kubernetes.io/projected/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-kube-api-access-2chpx\") on node \"crc\" DevicePath \"\"" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.321777 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bb43c1-17a0-467f-b5fe-e0092f58d0c3-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.321787 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/819a9975-6fc4-4fae-a817-9cc2043dc231-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.321796 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pglj\" (UniqueName: \"kubernetes.io/projected/819a9975-6fc4-4fae-a817-9cc2043dc231-kube-api-access-9pglj\") on node \"crc\" DevicePath \"\"" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.423464 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-njndh"] Feb 25 07:05:54 crc kubenswrapper[4978]: W0225 07:05:54.431546 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01eac805_ca08_46de_b79b_0ca5645f6955.slice/crio-3e44376df0425e7f67c053a698f5548c2e2d3c3e261fe12a08955000080a6fbc WatchSource:0}: Error finding container 3e44376df0425e7f67c053a698f5548c2e2d3c3e261fe12a08955000080a6fbc: Status 404 returned error can't find the container with id 3e44376df0425e7f67c053a698f5548c2e2d3c3e261fe12a08955000080a6fbc Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.559894 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" event={"ID":"99bb43c1-17a0-467f-b5fe-e0092f58d0c3","Type":"ContainerDied","Data":"4eeb334f5665034710ab18b485cf5ad9a80eb6c5f9850842041ff381df65a180"} Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.559904 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fb6c8b7-lblhx" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.562001 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"85624b7d-192e-42c3-9c5b-18cfd91d37e9","Type":"ContainerStarted","Data":"17df30ac7bbc09d7f30340ec5bcdd268698d7d872757d6efefa7590e58e98dd9"} Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.563308 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fbf80328-4523-49ff-b390-0e8b3dd9cc02","Type":"ContainerStarted","Data":"80c3b9c65701b46a40069afdb55e90b6da48c0e1e407e3ce8974dfc61a35413c"} Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.565476 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ed9859d2-701a-4011-b9d6-d8cf6291265a","Type":"ContainerStarted","Data":"31a02b14eff08a3a2fb7cfd46bb47a4dea8b566f67dc09dede386d4454c64379"} Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.567859 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6df7a672-b889-40d3-abe5-271e4d36dd85","Type":"ContainerStarted","Data":"944d35b451cb47c9e602257957eeb480054d54b2e1346c5e1d27f74e82490a9c"} Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.569500 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-njndh" event={"ID":"01eac805-ca08-46de-b79b-0ca5645f6955","Type":"ContainerStarted","Data":"3e44376df0425e7f67c053a698f5548c2e2d3c3e261fe12a08955000080a6fbc"} Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.571074 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" event={"ID":"819a9975-6fc4-4fae-a817-9cc2043dc231","Type":"ContainerDied","Data":"8e12e77349447e56b4ff1ca84957877d555559793bd5a647f225229e9d80e0ae"} Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.571200 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648df8f4f9-d8vbv" Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.575973 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r45vg" event={"ID":"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb","Type":"ContainerStarted","Data":"df64dd81561667b23fde5ead3623cad904eb563e51098c9a81b12150ce28c753"} Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.578284 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c488aaac-88dd-4eb1-9020-e17749814ed9","Type":"ContainerStarted","Data":"3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658"} Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.653622 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9fb6c8b7-lblhx"] Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.667050 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9fb6c8b7-lblhx"] Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.682313 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-648df8f4f9-d8vbv"] Feb 25 07:05:54 crc kubenswrapper[4978]: I0225 07:05:54.692744 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-648df8f4f9-d8vbv"] Feb 25 07:05:55 crc kubenswrapper[4978]: I0225 07:05:55.336965 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="819a9975-6fc4-4fae-a817-9cc2043dc231" path="/var/lib/kubelet/pods/819a9975-6fc4-4fae-a817-9cc2043dc231/volumes" Feb 25 07:05:55 crc kubenswrapper[4978]: I0225 07:05:55.338745 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99bb43c1-17a0-467f-b5fe-e0092f58d0c3" path="/var/lib/kubelet/pods/99bb43c1-17a0-467f-b5fe-e0092f58d0c3/volumes" Feb 25 07:05:55 crc kubenswrapper[4978]: I0225 07:05:55.588278 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"09112202-a3ef-4742-bdbe-b46c1c9a55ba","Type":"ContainerStarted","Data":"62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732"} Feb 25 07:05:55 crc kubenswrapper[4978]: I0225 07:05:55.590435 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43bd8325-9408-4a32-9285-d522d497205e","Type":"ContainerStarted","Data":"765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8"} Feb 25 07:05:58 crc kubenswrapper[4978]: I0225 07:05:58.615006 4978 generic.go:334] "Generic (PLEG): container finished" podID="c488aaac-88dd-4eb1-9020-e17749814ed9" containerID="3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658" exitCode=0 Feb 25 07:05:58 crc kubenswrapper[4978]: I0225 07:05:58.615088 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c488aaac-88dd-4eb1-9020-e17749814ed9","Type":"ContainerDied","Data":"3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658"} Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.135724 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533386-94rxl"] Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.139525 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533386-94rxl" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.141862 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.142105 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.145268 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.149788 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533386-94rxl"] Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.225650 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxl2c\" (UniqueName: \"kubernetes.io/projected/fff273d8-333d-435b-86cc-613bb1c53b7c-kube-api-access-zxl2c\") pod \"auto-csr-approver-29533386-94rxl\" (UID: \"fff273d8-333d-435b-86cc-613bb1c53b7c\") " pod="openshift-infra/auto-csr-approver-29533386-94rxl" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.327791 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxl2c\" (UniqueName: \"kubernetes.io/projected/fff273d8-333d-435b-86cc-613bb1c53b7c-kube-api-access-zxl2c\") pod \"auto-csr-approver-29533386-94rxl\" (UID: \"fff273d8-333d-435b-86cc-613bb1c53b7c\") " pod="openshift-infra/auto-csr-approver-29533386-94rxl" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.348568 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxl2c\" (UniqueName: \"kubernetes.io/projected/fff273d8-333d-435b-86cc-613bb1c53b7c-kube-api-access-zxl2c\") pod \"auto-csr-approver-29533386-94rxl\" (UID: \"fff273d8-333d-435b-86cc-613bb1c53b7c\") " pod="openshift-infra/auto-csr-approver-29533386-94rxl" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.457955 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533386-94rxl" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.637239 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"85624b7d-192e-42c3-9c5b-18cfd91d37e9","Type":"ContainerStarted","Data":"22db8e67ef3c374833a5a7d3778dc19f3cb69c3620c3009d7badbe8c31ef7006"} Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.638020 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.645829 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fbf80328-4523-49ff-b390-0e8b3dd9cc02","Type":"ContainerStarted","Data":"037e82b5ff3e20c28f0050cfb9295e40e3a11fad5596b116d220432860b85e2d"} Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.646702 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.656993 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=24.706662737 podStartE2EDuration="29.656975947s" podCreationTimestamp="2026-02-25 07:05:31 +0000 UTC" firstStartedPulling="2026-02-25 07:05:54.171750334 +0000 UTC m=+1247.611006793" lastFinishedPulling="2026-02-25 07:05:59.122063514 +0000 UTC m=+1252.561320003" observedRunningTime="2026-02-25 07:06:00.656405241 +0000 UTC m=+1254.095661740" watchObservedRunningTime="2026-02-25 07:06:00.656975947 +0000 UTC m=+1254.096232406" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.658519 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c488aaac-88dd-4eb1-9020-e17749814ed9","Type":"ContainerStarted","Data":"df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79"} Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.661490 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ed9859d2-701a-4011-b9d6-d8cf6291265a","Type":"ContainerStarted","Data":"8baa6e062d410fe1594d058fc3924f5a1ed85894951c8f77a4918acff9d1e1bd"} Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.664224 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6df7a672-b889-40d3-abe5-271e4d36dd85","Type":"ContainerStarted","Data":"c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7"} Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.665743 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r45vg" event={"ID":"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb","Type":"ContainerStarted","Data":"aefb1dba292711435eac1e73e9e529ab2e03a7b4863f808c55d3071e985aa85e"} Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.666735 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-r45vg" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.669558 4978 generic.go:334] "Generic (PLEG): container finished" podID="01eac805-ca08-46de-b79b-0ca5645f6955" containerID="6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8" exitCode=0 Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.669593 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-njndh" event={"ID":"01eac805-ca08-46de-b79b-0ca5645f6955","Type":"ContainerDied","Data":"6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8"} Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.677360 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=21.062069901 podStartE2EDuration="26.677344271s" podCreationTimestamp="2026-02-25 07:05:34 +0000 UTC" firstStartedPulling="2026-02-25 07:05:54.189927491 +0000 UTC m=+1247.629183950" lastFinishedPulling="2026-02-25 07:05:59.805201851 +0000 UTC m=+1253.244458320" observedRunningTime="2026-02-25 07:06:00.671563057 +0000 UTC m=+1254.110819536" watchObservedRunningTime="2026-02-25 07:06:00.677344271 +0000 UTC m=+1254.116600730" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.704210 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=13.284879045 podStartE2EDuration="30.70419108s" podCreationTimestamp="2026-02-25 07:05:30 +0000 UTC" firstStartedPulling="2026-02-25 07:05:36.007315024 +0000 UTC m=+1229.446571523" lastFinishedPulling="2026-02-25 07:05:53.426627099 +0000 UTC m=+1246.865883558" observedRunningTime="2026-02-25 07:06:00.699303822 +0000 UTC m=+1254.138560291" watchObservedRunningTime="2026-02-25 07:06:00.70419108 +0000 UTC m=+1254.143447549" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.717470 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-r45vg" podStartSLOduration=18.976811064 podStartE2EDuration="23.717419008s" podCreationTimestamp="2026-02-25 07:05:37 +0000 UTC" firstStartedPulling="2026-02-25 07:05:54.189968832 +0000 UTC m=+1247.629225291" lastFinishedPulling="2026-02-25 07:05:58.930576746 +0000 UTC m=+1252.369833235" observedRunningTime="2026-02-25 07:06:00.715246513 +0000 UTC m=+1254.154502982" watchObservedRunningTime="2026-02-25 07:06:00.717419008 +0000 UTC m=+1254.156675467" Feb 25 07:06:00 crc kubenswrapper[4978]: I0225 07:06:00.907105 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533386-94rxl"] Feb 25 07:06:00 crc kubenswrapper[4978]: W0225 07:06:00.909262 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfff273d8_333d_435b_86cc_613bb1c53b7c.slice/crio-80caf51c1ad1b0c11d8ef9a1bad78db4142507a2480149c0710105461737971c WatchSource:0}: Error finding container 80caf51c1ad1b0c11d8ef9a1bad78db4142507a2480149c0710105461737971c: Status 404 returned error can't find the container with id 80caf51c1ad1b0c11d8ef9a1bad78db4142507a2480149c0710105461737971c Feb 25 07:06:01 crc kubenswrapper[4978]: I0225 07:06:01.684687 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533386-94rxl" event={"ID":"fff273d8-333d-435b-86cc-613bb1c53b7c","Type":"ContainerStarted","Data":"80caf51c1ad1b0c11d8ef9a1bad78db4142507a2480149c0710105461737971c"} Feb 25 07:06:01 crc kubenswrapper[4978]: I0225 07:06:01.691531 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-njndh" event={"ID":"01eac805-ca08-46de-b79b-0ca5645f6955","Type":"ContainerStarted","Data":"f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99"} Feb 25 07:06:01 crc kubenswrapper[4978]: I0225 07:06:01.778531 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 25 07:06:01 crc kubenswrapper[4978]: I0225 07:06:01.778637 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.701203 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-njndh" event={"ID":"01eac805-ca08-46de-b79b-0ca5645f6955","Type":"ContainerStarted","Data":"02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b"} Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.702139 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.702158 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.703649 4978 generic.go:334] "Generic (PLEG): container finished" podID="fff273d8-333d-435b-86cc-613bb1c53b7c" containerID="ea0f40bdc05d6dee36735b9c80c86d91e506b9d4f91466ebf56de31a28074aa5" exitCode=0 Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.703724 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533386-94rxl" event={"ID":"fff273d8-333d-435b-86cc-613bb1c53b7c","Type":"ContainerDied","Data":"ea0f40bdc05d6dee36735b9c80c86d91e506b9d4f91466ebf56de31a28074aa5"} Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.705757 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"315eefbb-24fb-4e45-97ed-03caf894bc08","Type":"ContainerStarted","Data":"414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b"} Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.707804 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ed9859d2-701a-4011-b9d6-d8cf6291265a","Type":"ContainerStarted","Data":"a15f02424a1013ac6691c287f99c51cc3e6aff1c7146b986bb694436609fc027"} Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.709970 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6df7a672-b889-40d3-abe5-271e4d36dd85","Type":"ContainerStarted","Data":"c227278c28e8d88bb7c9ddfa3eaa4e8de5003b5ef2d93630c76d010d8ca314e0"} Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.727850 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-njndh" podStartSLOduration=21.338360178 podStartE2EDuration="25.727829634s" podCreationTimestamp="2026-02-25 07:05:37 +0000 UTC" firstStartedPulling="2026-02-25 07:05:54.433082466 +0000 UTC m=+1247.872338925" lastFinishedPulling="2026-02-25 07:05:58.822551922 +0000 UTC m=+1252.261808381" observedRunningTime="2026-02-25 07:06:02.724284417 +0000 UTC m=+1256.163540866" watchObservedRunningTime="2026-02-25 07:06:02.727829634 +0000 UTC m=+1256.167086093" Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.768579 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=15.031973154 podStartE2EDuration="22.76856181s" podCreationTimestamp="2026-02-25 07:05:40 +0000 UTC" firstStartedPulling="2026-02-25 07:05:54.160520625 +0000 UTC m=+1247.599777084" lastFinishedPulling="2026-02-25 07:06:01.897109281 +0000 UTC m=+1255.336365740" observedRunningTime="2026-02-25 07:06:02.764801837 +0000 UTC m=+1256.204058306" watchObservedRunningTime="2026-02-25 07:06:02.76856181 +0000 UTC m=+1256.207818269" Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.793683 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=18.20283475 podStartE2EDuration="25.793661236s" podCreationTimestamp="2026-02-25 07:05:37 +0000 UTC" firstStartedPulling="2026-02-25 07:05:54.291299664 +0000 UTC m=+1247.730556123" lastFinishedPulling="2026-02-25 07:06:01.88212613 +0000 UTC m=+1255.321382609" observedRunningTime="2026-02-25 07:06:02.788083398 +0000 UTC m=+1256.227339887" watchObservedRunningTime="2026-02-25 07:06:02.793661236 +0000 UTC m=+1256.232917705" Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.844903 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.906237 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 25 07:06:02 crc kubenswrapper[4978]: I0225 07:06:02.954929 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 25 07:06:03 crc kubenswrapper[4978]: I0225 07:06:03.721091 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 25 07:06:04 crc kubenswrapper[4978]: I0225 07:06:04.134207 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533386-94rxl" Feb 25 07:06:04 crc kubenswrapper[4978]: I0225 07:06:04.199106 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxl2c\" (UniqueName: \"kubernetes.io/projected/fff273d8-333d-435b-86cc-613bb1c53b7c-kube-api-access-zxl2c\") pod \"fff273d8-333d-435b-86cc-613bb1c53b7c\" (UID: \"fff273d8-333d-435b-86cc-613bb1c53b7c\") " Feb 25 07:06:04 crc kubenswrapper[4978]: I0225 07:06:04.205054 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fff273d8-333d-435b-86cc-613bb1c53b7c-kube-api-access-zxl2c" (OuterVolumeSpecName: "kube-api-access-zxl2c") pod "fff273d8-333d-435b-86cc-613bb1c53b7c" (UID: "fff273d8-333d-435b-86cc-613bb1c53b7c"). InnerVolumeSpecName "kube-api-access-zxl2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:04 crc kubenswrapper[4978]: I0225 07:06:04.302167 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxl2c\" (UniqueName: \"kubernetes.io/projected/fff273d8-333d-435b-86cc-613bb1c53b7c-kube-api-access-zxl2c\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:04 crc kubenswrapper[4978]: I0225 07:06:04.738996 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533386-94rxl" event={"ID":"fff273d8-333d-435b-86cc-613bb1c53b7c","Type":"ContainerDied","Data":"80caf51c1ad1b0c11d8ef9a1bad78db4142507a2480149c0710105461737971c"} Feb 25 07:06:04 crc kubenswrapper[4978]: I0225 07:06:04.739062 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80caf51c1ad1b0c11d8ef9a1bad78db4142507a2480149c0710105461737971c" Feb 25 07:06:04 crc kubenswrapper[4978]: I0225 07:06:04.739430 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533386-94rxl" Feb 25 07:06:04 crc kubenswrapper[4978]: I0225 07:06:04.791128 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.088417 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-776868475c-ch4tm"] Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.124516 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-649c9df595-qpm9m"] Feb 25 07:06:05 crc kubenswrapper[4978]: E0225 07:06:05.125095 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fff273d8-333d-435b-86cc-613bb1c53b7c" containerName="oc" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.125107 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="fff273d8-333d-435b-86cc-613bb1c53b7c" containerName="oc" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.125247 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="fff273d8-333d-435b-86cc-613bb1c53b7c" containerName="oc" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.133838 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-649c9df595-qpm9m"] Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.133959 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.136595 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.151614 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-4z765"] Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.152661 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.155338 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.174738 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4z765"] Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.218364 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533380-mbpf9"] Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.227796 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533380-mbpf9"] Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.247031 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5s5p\" (UniqueName: \"kubernetes.io/projected/2ab08149-16b3-42ee-be06-5e663fad2df0-kube-api-access-k5s5p\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.247103 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-config\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.247127 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-dns-svc\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.247195 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-ovsdbserver-nb\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.338126 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87f4f3a6-8318-45cd-8942-b3a58210c930" path="/var/lib/kubelet/pods/87f4f3a6-8318-45cd-8942-b3a58210c930/volumes" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.348758 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovn-rundir\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.348819 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovs-rundir\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.348840 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.348882 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-combined-ca-bundle\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.348908 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-ovsdbserver-nb\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.348978 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-config\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.349002 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5s5p\" (UniqueName: \"kubernetes.io/projected/2ab08149-16b3-42ee-be06-5e663fad2df0-kube-api-access-k5s5p\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.349054 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-config\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.349073 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-dns-svc\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.349098 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnbv5\" (UniqueName: \"kubernetes.io/projected/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-kube-api-access-jnbv5\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.352888 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-ovsdbserver-nb\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.354866 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-dns-svc\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.355289 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-config\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.377654 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5s5p\" (UniqueName: \"kubernetes.io/projected/2ab08149-16b3-42ee-be06-5e663fad2df0-kube-api-access-k5s5p\") pod \"dnsmasq-dns-649c9df595-qpm9m\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.450508 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovs-rundir\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.450557 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.450588 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-combined-ca-bundle\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.450664 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-config\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.450742 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnbv5\" (UniqueName: \"kubernetes.io/projected/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-kube-api-access-jnbv5\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.450775 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovn-rundir\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.450801 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovs-rundir\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.451992 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovn-rundir\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.452484 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-config\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.454847 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.456305 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-combined-ca-bundle\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.477957 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnbv5\" (UniqueName: \"kubernetes.io/projected/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-kube-api-access-jnbv5\") pod \"ovn-controller-metrics-4z765\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.485559 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dfb8ff55f-nhh4k"] Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.486595 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.507505 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.518838 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f9bcc5599-c5twm"] Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.520010 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.521220 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.527328 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f9bcc5599-c5twm"] Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.590238 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-776868475c-ch4tm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.653815 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-config\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.653871 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-dns-svc\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.653909 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmpjw\" (UniqueName: \"kubernetes.io/projected/3115eb4c-6af8-41df-9d7e-4db587e37ac5-kube-api-access-zmpjw\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.653966 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.653988 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.748793 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-776868475c-ch4tm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.748847 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-776868475c-ch4tm" event={"ID":"32d9e2ad-9750-4062-a4a4-da0a072037de","Type":"ContainerDied","Data":"78e0bfc37fc8b7efe27b3f4c6a178d38875603f6b1351673423b6a7bf36d1963"} Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.754960 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32d9e2ad-9750-4062-a4a4-da0a072037de-config\") pod \"32d9e2ad-9750-4062-a4a4-da0a072037de\" (UID: \"32d9e2ad-9750-4062-a4a4-da0a072037de\") " Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.755015 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsn6k\" (UniqueName: \"kubernetes.io/projected/32d9e2ad-9750-4062-a4a4-da0a072037de-kube-api-access-hsn6k\") pod \"32d9e2ad-9750-4062-a4a4-da0a072037de\" (UID: \"32d9e2ad-9750-4062-a4a4-da0a072037de\") " Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.755454 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.755479 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.755544 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-config\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.755597 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-dns-svc\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.755641 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmpjw\" (UniqueName: \"kubernetes.io/projected/3115eb4c-6af8-41df-9d7e-4db587e37ac5-kube-api-access-zmpjw\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.756196 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.756747 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-config\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.757631 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-dns-svc\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.757898 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.757982 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32d9e2ad-9750-4062-a4a4-da0a072037de-config" (OuterVolumeSpecName: "config") pod "32d9e2ad-9750-4062-a4a4-da0a072037de" (UID: "32d9e2ad-9750-4062-a4a4-da0a072037de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.759952 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32d9e2ad-9750-4062-a4a4-da0a072037de-kube-api-access-hsn6k" (OuterVolumeSpecName: "kube-api-access-hsn6k") pod "32d9e2ad-9750-4062-a4a4-da0a072037de" (UID: "32d9e2ad-9750-4062-a4a4-da0a072037de"). InnerVolumeSpecName "kube-api-access-hsn6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.773055 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmpjw\" (UniqueName: \"kubernetes.io/projected/3115eb4c-6af8-41df-9d7e-4db587e37ac5-kube-api-access-zmpjw\") pod \"dnsmasq-dns-6f9bcc5599-c5twm\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.856730 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32d9e2ad-9750-4062-a4a4-da0a072037de-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.856765 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsn6k\" (UniqueName: \"kubernetes.io/projected/32d9e2ad-9750-4062-a4a4-da0a072037de-kube-api-access-hsn6k\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.906481 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:05 crc kubenswrapper[4978]: I0225 07:06:05.952999 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4z765"] Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.028099 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.045607 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-649c9df595-qpm9m"] Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.045647 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.164157 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.224927 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-776868475c-ch4tm"] Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.262193 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-776868475c-ch4tm"] Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.320933 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.400453 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.401775 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.412813 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.412987 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.413094 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-29jpp" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.413199 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.414783 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.465491 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.471245 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.471278 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-scripts\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.471311 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-config\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.471343 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg4nd\" (UniqueName: \"kubernetes.io/projected/27864863-0a60-43ec-ac3e-f86d9f68c7d2-kube-api-access-fg4nd\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.471364 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.471415 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.471837 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.511577 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f9bcc5599-c5twm"] Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.573891 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.573970 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.573992 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-scripts\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.574011 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-config\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.574035 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg4nd\" (UniqueName: \"kubernetes.io/projected/27864863-0a60-43ec-ac3e-f86d9f68c7d2-kube-api-access-fg4nd\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.574059 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.574087 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.579625 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-scripts\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.580630 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.580678 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.580993 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-config\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.581826 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.589112 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.595545 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg4nd\" (UniqueName: \"kubernetes.io/projected/27864863-0a60-43ec-ac3e-f86d9f68c7d2-kube-api-access-fg4nd\") pod \"ovn-northd-0\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.743095 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.759320 4978 generic.go:334] "Generic (PLEG): container finished" podID="315eefbb-24fb-4e45-97ed-03caf894bc08" containerID="414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b" exitCode=0 Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.759410 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"315eefbb-24fb-4e45-97ed-03caf894bc08","Type":"ContainerDied","Data":"414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b"} Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.762793 4978 generic.go:334] "Generic (PLEG): container finished" podID="2ab08149-16b3-42ee-be06-5e663fad2df0" containerID="be274f99b956d30f11bd4b5e8b166fe685a787c558c954c8d45787aaf386cdcf" exitCode=0 Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.762893 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" event={"ID":"2ab08149-16b3-42ee-be06-5e663fad2df0","Type":"ContainerDied","Data":"be274f99b956d30f11bd4b5e8b166fe685a787c558c954c8d45787aaf386cdcf"} Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.762933 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" event={"ID":"2ab08149-16b3-42ee-be06-5e663fad2df0","Type":"ContainerStarted","Data":"c2f14279704516a369463eb86c7376a81c146ff90150417ef9e703357bde568d"} Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.765640 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4z765" event={"ID":"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682","Type":"ContainerStarted","Data":"986d5b46fb6a179bd4f7544f78f8396ed0327bcfeea74164711b40b9816145da"} Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.765685 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4z765" event={"ID":"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682","Type":"ContainerStarted","Data":"440ba6ed4d8946c310454eaecfb11414cacb380bf48e6d2fa2c5424f41b036e0"} Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.768440 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" event={"ID":"3115eb4c-6af8-41df-9d7e-4db587e37ac5","Type":"ContainerStarted","Data":"ef4f3364dc9f05a145615387e50a528356278511ff4066da67e7d2bfe3e3f146"} Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.771163 4978 generic.go:334] "Generic (PLEG): container finished" podID="80d0ea7a-b122-4988-bff6-e4377b79b7a0" containerID="729d373bb8eef510e794c6bb8d8a6d91c30aa9a6130ab5cc12c98ccc1808ac35" exitCode=0 Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.772800 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" event={"ID":"80d0ea7a-b122-4988-bff6-e4377b79b7a0","Type":"ContainerDied","Data":"729d373bb8eef510e794c6bb8d8a6d91c30aa9a6130ab5cc12c98ccc1808ac35"} Feb 25 07:06:06 crc kubenswrapper[4978]: I0225 07:06:06.863064 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-4z765" podStartSLOduration=1.8630422009999998 podStartE2EDuration="1.863042201s" podCreationTimestamp="2026-02-25 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:06.857727791 +0000 UTC m=+1260.296984270" watchObservedRunningTime="2026-02-25 07:06:06.863042201 +0000 UTC m=+1260.302298660" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.168123 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.241592 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.290222 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-dns-svc\") pod \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.290304 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9xtg\" (UniqueName: \"kubernetes.io/projected/80d0ea7a-b122-4988-bff6-e4377b79b7a0-kube-api-access-v9xtg\") pod \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.290486 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-config\") pod \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\" (UID: \"80d0ea7a-b122-4988-bff6-e4377b79b7a0\") " Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.297330 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80d0ea7a-b122-4988-bff6-e4377b79b7a0-kube-api-access-v9xtg" (OuterVolumeSpecName: "kube-api-access-v9xtg") pod "80d0ea7a-b122-4988-bff6-e4377b79b7a0" (UID: "80d0ea7a-b122-4988-bff6-e4377b79b7a0"). InnerVolumeSpecName "kube-api-access-v9xtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.312129 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-config" (OuterVolumeSpecName: "config") pod "80d0ea7a-b122-4988-bff6-e4377b79b7a0" (UID: "80d0ea7a-b122-4988-bff6-e4377b79b7a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.316907 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "80d0ea7a-b122-4988-bff6-e4377b79b7a0" (UID: "80d0ea7a-b122-4988-bff6-e4377b79b7a0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.341490 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32d9e2ad-9750-4062-a4a4-da0a072037de" path="/var/lib/kubelet/pods/32d9e2ad-9750-4062-a4a4-da0a072037de/volumes" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.373686 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.393439 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.393473 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9xtg\" (UniqueName: \"kubernetes.io/projected/80d0ea7a-b122-4988-bff6-e4377b79b7a0-kube-api-access-v9xtg\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.393486 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80d0ea7a-b122-4988-bff6-e4377b79b7a0-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.795449 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" event={"ID":"2ab08149-16b3-42ee-be06-5e663fad2df0","Type":"ContainerStarted","Data":"5deb0f5164203d2b369cebbc143c949a54c18541d08040ce69321a3538515667"} Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.795968 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.798175 4978 generic.go:334] "Generic (PLEG): container finished" podID="3115eb4c-6af8-41df-9d7e-4db587e37ac5" containerID="f90d153c5268ec061775ca86fed36beeeb119b85319aded398fd1612371d491c" exitCode=0 Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.798310 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" event={"ID":"3115eb4c-6af8-41df-9d7e-4db587e37ac5","Type":"ContainerDied","Data":"f90d153c5268ec061775ca86fed36beeeb119b85319aded398fd1612371d491c"} Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.804312 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" event={"ID":"80d0ea7a-b122-4988-bff6-e4377b79b7a0","Type":"ContainerDied","Data":"47324e916a4f0ec55346b1b03c72b5b3c7ddd14344f58639eb11bf73d6d5a2e2"} Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.804358 4978 scope.go:117] "RemoveContainer" containerID="729d373bb8eef510e794c6bb8d8a6d91c30aa9a6130ab5cc12c98ccc1808ac35" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.804468 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dfb8ff55f-nhh4k" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.810151 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"27864863-0a60-43ec-ac3e-f86d9f68c7d2","Type":"ContainerStarted","Data":"b2bcc3e0dfa7ab325a01e93711b4a10efbcd7ee43e7e9c52abf8b0b46867fb7e"} Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.816432 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"315eefbb-24fb-4e45-97ed-03caf894bc08","Type":"ContainerStarted","Data":"4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f"} Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.832753 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" podStartSLOduration=2.832735939 podStartE2EDuration="2.832735939s" podCreationTimestamp="2026-02-25 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:07.831360488 +0000 UTC m=+1261.270616987" watchObservedRunningTime="2026-02-25 07:06:07.832735939 +0000 UTC m=+1261.271992398" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.859757 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371996.995037 podStartE2EDuration="39.859739173s" podCreationTimestamp="2026-02-25 07:05:28 +0000 UTC" firstStartedPulling="2026-02-25 07:05:30.885649403 +0000 UTC m=+1224.324905862" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:07.856923798 +0000 UTC m=+1261.296180287" watchObservedRunningTime="2026-02-25 07:06:07.859739173 +0000 UTC m=+1261.298995632" Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.940101 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dfb8ff55f-nhh4k"] Feb 25 07:06:07 crc kubenswrapper[4978]: I0225 07:06:07.948059 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6dfb8ff55f-nhh4k"] Feb 25 07:06:08 crc kubenswrapper[4978]: I0225 07:06:08.846074 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" event={"ID":"3115eb4c-6af8-41df-9d7e-4db587e37ac5","Type":"ContainerStarted","Data":"bc7353a13e461bf5e8472c1a0de787f88ed960bbac64db951a039c2f381f9ce5"} Feb 25 07:06:08 crc kubenswrapper[4978]: I0225 07:06:08.846608 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:08 crc kubenswrapper[4978]: I0225 07:06:08.860324 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"27864863-0a60-43ec-ac3e-f86d9f68c7d2","Type":"ContainerStarted","Data":"c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d"} Feb 25 07:06:08 crc kubenswrapper[4978]: I0225 07:06:08.860467 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"27864863-0a60-43ec-ac3e-f86d9f68c7d2","Type":"ContainerStarted","Data":"6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98"} Feb 25 07:06:08 crc kubenswrapper[4978]: I0225 07:06:08.860732 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 25 07:06:08 crc kubenswrapper[4978]: I0225 07:06:08.869002 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" podStartSLOduration=3.868986132 podStartE2EDuration="3.868986132s" podCreationTimestamp="2026-02-25 07:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:08.866215249 +0000 UTC m=+1262.305471758" watchObservedRunningTime="2026-02-25 07:06:08.868986132 +0000 UTC m=+1262.308242591" Feb 25 07:06:08 crc kubenswrapper[4978]: I0225 07:06:08.891038 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.910200472 podStartE2EDuration="2.891017736s" podCreationTimestamp="2026-02-25 07:06:06 +0000 UTC" firstStartedPulling="2026-02-25 07:06:07.381835377 +0000 UTC m=+1260.821091846" lastFinishedPulling="2026-02-25 07:06:08.362652651 +0000 UTC m=+1261.801909110" observedRunningTime="2026-02-25 07:06:08.886033826 +0000 UTC m=+1262.325290295" watchObservedRunningTime="2026-02-25 07:06:08.891017736 +0000 UTC m=+1262.330274195" Feb 25 07:06:09 crc kubenswrapper[4978]: I0225 07:06:09.338845 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80d0ea7a-b122-4988-bff6-e4377b79b7a0" path="/var/lib/kubelet/pods/80d0ea7a-b122-4988-bff6-e4377b79b7a0/volumes" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.238019 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.238087 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.514881 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-4lv9d"] Feb 25 07:06:10 crc kubenswrapper[4978]: E0225 07:06:10.515512 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80d0ea7a-b122-4988-bff6-e4377b79b7a0" containerName="init" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.515532 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="80d0ea7a-b122-4988-bff6-e4377b79b7a0" containerName="init" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.515708 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="80d0ea7a-b122-4988-bff6-e4377b79b7a0" containerName="init" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.516286 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4lv9d" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.519055 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.546677 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-4lv9d"] Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.601757 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/889b0ea1-aa03-4331-8247-d29684abdcdc-operator-scripts\") pod \"root-account-create-update-4lv9d\" (UID: \"889b0ea1-aa03-4331-8247-d29684abdcdc\") " pod="openstack/root-account-create-update-4lv9d" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.601888 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8cnl\" (UniqueName: \"kubernetes.io/projected/889b0ea1-aa03-4331-8247-d29684abdcdc-kube-api-access-w8cnl\") pod \"root-account-create-update-4lv9d\" (UID: \"889b0ea1-aa03-4331-8247-d29684abdcdc\") " pod="openstack/root-account-create-update-4lv9d" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.703361 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/889b0ea1-aa03-4331-8247-d29684abdcdc-operator-scripts\") pod \"root-account-create-update-4lv9d\" (UID: \"889b0ea1-aa03-4331-8247-d29684abdcdc\") " pod="openstack/root-account-create-update-4lv9d" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.703523 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8cnl\" (UniqueName: \"kubernetes.io/projected/889b0ea1-aa03-4331-8247-d29684abdcdc-kube-api-access-w8cnl\") pod \"root-account-create-update-4lv9d\" (UID: \"889b0ea1-aa03-4331-8247-d29684abdcdc\") " pod="openstack/root-account-create-update-4lv9d" Feb 25 07:06:10 crc kubenswrapper[4978]: I0225 07:06:10.704643 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/889b0ea1-aa03-4331-8247-d29684abdcdc-operator-scripts\") pod \"root-account-create-update-4lv9d\" (UID: \"889b0ea1-aa03-4331-8247-d29684abdcdc\") " pod="openstack/root-account-create-update-4lv9d" Feb 25 07:06:11 crc kubenswrapper[4978]: I0225 07:06:11.003396 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8cnl\" (UniqueName: \"kubernetes.io/projected/889b0ea1-aa03-4331-8247-d29684abdcdc-kube-api-access-w8cnl\") pod \"root-account-create-update-4lv9d\" (UID: \"889b0ea1-aa03-4331-8247-d29684abdcdc\") " pod="openstack/root-account-create-update-4lv9d" Feb 25 07:06:11 crc kubenswrapper[4978]: I0225 07:06:11.137430 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4lv9d" Feb 25 07:06:11 crc kubenswrapper[4978]: I0225 07:06:11.634298 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-4lv9d"] Feb 25 07:06:11 crc kubenswrapper[4978]: W0225 07:06:11.651595 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod889b0ea1_aa03_4331_8247_d29684abdcdc.slice/crio-61adfa5198579d9fa832be2066afb9821fa0712d320e73edcccc2bf4110179d5 WatchSource:0}: Error finding container 61adfa5198579d9fa832be2066afb9821fa0712d320e73edcccc2bf4110179d5: Status 404 returned error can't find the container with id 61adfa5198579d9fa832be2066afb9821fa0712d320e73edcccc2bf4110179d5 Feb 25 07:06:12 crc kubenswrapper[4978]: I0225 07:06:12.019566 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4lv9d" event={"ID":"889b0ea1-aa03-4331-8247-d29684abdcdc","Type":"ContainerStarted","Data":"8989205d360d992ab99ca3347077abbe29a1d3dcc44b9b0e61af2f88c88568b8"} Feb 25 07:06:12 crc kubenswrapper[4978]: I0225 07:06:12.019927 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4lv9d" event={"ID":"889b0ea1-aa03-4331-8247-d29684abdcdc","Type":"ContainerStarted","Data":"61adfa5198579d9fa832be2066afb9821fa0712d320e73edcccc2bf4110179d5"} Feb 25 07:06:12 crc kubenswrapper[4978]: I0225 07:06:12.047837 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-4lv9d" podStartSLOduration=2.047812173 podStartE2EDuration="2.047812173s" podCreationTimestamp="2026-02-25 07:06:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:12.037212722 +0000 UTC m=+1265.476469191" watchObservedRunningTime="2026-02-25 07:06:12.047812173 +0000 UTC m=+1265.487068642" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.029606 4978 generic.go:334] "Generic (PLEG): container finished" podID="889b0ea1-aa03-4331-8247-d29684abdcdc" containerID="8989205d360d992ab99ca3347077abbe29a1d3dcc44b9b0e61af2f88c88568b8" exitCode=0 Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.029888 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4lv9d" event={"ID":"889b0ea1-aa03-4331-8247-d29684abdcdc","Type":"ContainerDied","Data":"8989205d360d992ab99ca3347077abbe29a1d3dcc44b9b0e61af2f88c88568b8"} Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.049068 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.159855 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.367116 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-d4b94"] Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.368116 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4b94" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.377717 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-cb34-account-create-update-k6dk6"] Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.378812 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cb34-account-create-update-k6dk6" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.384246 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.386469 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-d4b94"] Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.394697 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cb34-account-create-update-k6dk6"] Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.441051 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgvkx\" (UniqueName: \"kubernetes.io/projected/c183deb4-f193-4873-9617-5447f5c36e35-kube-api-access-kgvkx\") pod \"placement-db-create-d4b94\" (UID: \"c183deb4-f193-4873-9617-5447f5c36e35\") " pod="openstack/placement-db-create-d4b94" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.441148 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608f63c7-1251-42ea-996d-c7629a992f4a-operator-scripts\") pod \"placement-cb34-account-create-update-k6dk6\" (UID: \"608f63c7-1251-42ea-996d-c7629a992f4a\") " pod="openstack/placement-cb34-account-create-update-k6dk6" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.441357 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c183deb4-f193-4873-9617-5447f5c36e35-operator-scripts\") pod \"placement-db-create-d4b94\" (UID: \"c183deb4-f193-4873-9617-5447f5c36e35\") " pod="openstack/placement-db-create-d4b94" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.441486 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfnhv\" (UniqueName: \"kubernetes.io/projected/608f63c7-1251-42ea-996d-c7629a992f4a-kube-api-access-lfnhv\") pod \"placement-cb34-account-create-update-k6dk6\" (UID: \"608f63c7-1251-42ea-996d-c7629a992f4a\") " pod="openstack/placement-cb34-account-create-update-k6dk6" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.542502 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c183deb4-f193-4873-9617-5447f5c36e35-operator-scripts\") pod \"placement-db-create-d4b94\" (UID: \"c183deb4-f193-4873-9617-5447f5c36e35\") " pod="openstack/placement-db-create-d4b94" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.542551 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfnhv\" (UniqueName: \"kubernetes.io/projected/608f63c7-1251-42ea-996d-c7629a992f4a-kube-api-access-lfnhv\") pod \"placement-cb34-account-create-update-k6dk6\" (UID: \"608f63c7-1251-42ea-996d-c7629a992f4a\") " pod="openstack/placement-cb34-account-create-update-k6dk6" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.542614 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgvkx\" (UniqueName: \"kubernetes.io/projected/c183deb4-f193-4873-9617-5447f5c36e35-kube-api-access-kgvkx\") pod \"placement-db-create-d4b94\" (UID: \"c183deb4-f193-4873-9617-5447f5c36e35\") " pod="openstack/placement-db-create-d4b94" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.542661 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608f63c7-1251-42ea-996d-c7629a992f4a-operator-scripts\") pod \"placement-cb34-account-create-update-k6dk6\" (UID: \"608f63c7-1251-42ea-996d-c7629a992f4a\") " pod="openstack/placement-cb34-account-create-update-k6dk6" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.543455 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608f63c7-1251-42ea-996d-c7629a992f4a-operator-scripts\") pod \"placement-cb34-account-create-update-k6dk6\" (UID: \"608f63c7-1251-42ea-996d-c7629a992f4a\") " pod="openstack/placement-cb34-account-create-update-k6dk6" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.543537 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c183deb4-f193-4873-9617-5447f5c36e35-operator-scripts\") pod \"placement-db-create-d4b94\" (UID: \"c183deb4-f193-4873-9617-5447f5c36e35\") " pod="openstack/placement-db-create-d4b94" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.565277 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgvkx\" (UniqueName: \"kubernetes.io/projected/c183deb4-f193-4873-9617-5447f5c36e35-kube-api-access-kgvkx\") pod \"placement-db-create-d4b94\" (UID: \"c183deb4-f193-4873-9617-5447f5c36e35\") " pod="openstack/placement-db-create-d4b94" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.567436 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfnhv\" (UniqueName: \"kubernetes.io/projected/608f63c7-1251-42ea-996d-c7629a992f4a-kube-api-access-lfnhv\") pod \"placement-cb34-account-create-update-k6dk6\" (UID: \"608f63c7-1251-42ea-996d-c7629a992f4a\") " pod="openstack/placement-cb34-account-create-update-k6dk6" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.690947 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4b94" Feb 25 07:06:13 crc kubenswrapper[4978]: I0225 07:06:13.699391 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cb34-account-create-update-k6dk6" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.174113 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-d4b94"] Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.294812 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cb34-account-create-update-k6dk6"] Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.446015 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4lv9d" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.473118 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/889b0ea1-aa03-4331-8247-d29684abdcdc-operator-scripts\") pod \"889b0ea1-aa03-4331-8247-d29684abdcdc\" (UID: \"889b0ea1-aa03-4331-8247-d29684abdcdc\") " Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.473171 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8cnl\" (UniqueName: \"kubernetes.io/projected/889b0ea1-aa03-4331-8247-d29684abdcdc-kube-api-access-w8cnl\") pod \"889b0ea1-aa03-4331-8247-d29684abdcdc\" (UID: \"889b0ea1-aa03-4331-8247-d29684abdcdc\") " Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.475975 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/889b0ea1-aa03-4331-8247-d29684abdcdc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "889b0ea1-aa03-4331-8247-d29684abdcdc" (UID: "889b0ea1-aa03-4331-8247-d29684abdcdc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.501345 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/889b0ea1-aa03-4331-8247-d29684abdcdc-kube-api-access-w8cnl" (OuterVolumeSpecName: "kube-api-access-w8cnl") pod "889b0ea1-aa03-4331-8247-d29684abdcdc" (UID: "889b0ea1-aa03-4331-8247-d29684abdcdc"). InnerVolumeSpecName "kube-api-access-w8cnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.507906 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.542446 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-649c9df595-qpm9m"] Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.542727 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" podUID="2ab08149-16b3-42ee-be06-5e663fad2df0" containerName="dnsmasq-dns" containerID="cri-o://5deb0f5164203d2b369cebbc143c949a54c18541d08040ce69321a3538515667" gracePeriod=10 Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.544823 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.600708 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/889b0ea1-aa03-4331-8247-d29684abdcdc-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.600742 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8cnl\" (UniqueName: \"kubernetes.io/projected/889b0ea1-aa03-4331-8247-d29684abdcdc-kube-api-access-w8cnl\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.608447 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77cf9b784c-524z4"] Feb 25 07:06:14 crc kubenswrapper[4978]: E0225 07:06:14.609088 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="889b0ea1-aa03-4331-8247-d29684abdcdc" containerName="mariadb-account-create-update" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.609105 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="889b0ea1-aa03-4331-8247-d29684abdcdc" containerName="mariadb-account-create-update" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.609297 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="889b0ea1-aa03-4331-8247-d29684abdcdc" containerName="mariadb-account-create-update" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.610156 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.633390 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77cf9b784c-524z4"] Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.706949 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-config\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.706992 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-dns-svc\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.707048 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-nb\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.707077 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-sb\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.707140 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2d9b\" (UniqueName: \"kubernetes.io/projected/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-kube-api-access-z2d9b\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.838270 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-nb\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.839745 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-nb\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.840164 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-sb\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.841270 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-sb\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.841478 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2d9b\" (UniqueName: \"kubernetes.io/projected/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-kube-api-access-z2d9b\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.841538 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-config\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.841571 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-dns-svc\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.843182 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-config\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.843200 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-dns-svc\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.864209 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2d9b\" (UniqueName: \"kubernetes.io/projected/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-kube-api-access-z2d9b\") pod \"dnsmasq-dns-77cf9b784c-524z4\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:14 crc kubenswrapper[4978]: I0225 07:06:14.957944 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.040448 4978 scope.go:117] "RemoveContainer" containerID="1ceed6f5add84033ac8410f88f984ec20df27e6503464fac02cb4408959b2fd5" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.050890 4978 generic.go:334] "Generic (PLEG): container finished" podID="2ab08149-16b3-42ee-be06-5e663fad2df0" containerID="5deb0f5164203d2b369cebbc143c949a54c18541d08040ce69321a3538515667" exitCode=0 Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.050959 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" event={"ID":"2ab08149-16b3-42ee-be06-5e663fad2df0","Type":"ContainerDied","Data":"5deb0f5164203d2b369cebbc143c949a54c18541d08040ce69321a3538515667"} Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.052502 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cb34-account-create-update-k6dk6" event={"ID":"608f63c7-1251-42ea-996d-c7629a992f4a","Type":"ContainerStarted","Data":"9569ef92f4606b61b4ee144b77430b5109d298241ce9e4d4f08627b799dfe74a"} Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.052553 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cb34-account-create-update-k6dk6" event={"ID":"608f63c7-1251-42ea-996d-c7629a992f4a","Type":"ContainerStarted","Data":"6b3cd74a0eb9fd53e0127195d842f998886650fb40418a96a70744db0b7f3603"} Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.063254 4978 generic.go:334] "Generic (PLEG): container finished" podID="c183deb4-f193-4873-9617-5447f5c36e35" containerID="62b83f2cbf47e13ff7892d9042365b10f33ef61ca05bc25efd54f299ca6342e7" exitCode=0 Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.063353 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d4b94" event={"ID":"c183deb4-f193-4873-9617-5447f5c36e35","Type":"ContainerDied","Data":"62b83f2cbf47e13ff7892d9042365b10f33ef61ca05bc25efd54f299ca6342e7"} Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.063435 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d4b94" event={"ID":"c183deb4-f193-4873-9617-5447f5c36e35","Type":"ContainerStarted","Data":"671963c21204a5fdbfd123e6aa150af2ebb491107bee632f7ddcdc88079cfda4"} Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.065252 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-4lv9d" event={"ID":"889b0ea1-aa03-4331-8247-d29684abdcdc","Type":"ContainerDied","Data":"61adfa5198579d9fa832be2066afb9821fa0712d320e73edcccc2bf4110179d5"} Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.065280 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61adfa5198579d9fa832be2066afb9821fa0712d320e73edcccc2bf4110179d5" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.065345 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-4lv9d" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.083586 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-cb34-account-create-update-k6dk6" podStartSLOduration=2.083555463 podStartE2EDuration="2.083555463s" podCreationTimestamp="2026-02-25 07:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:15.073316354 +0000 UTC m=+1268.512572833" watchObservedRunningTime="2026-02-25 07:06:15.083555463 +0000 UTC m=+1268.522811922" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.177445 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.247524 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5s5p\" (UniqueName: \"kubernetes.io/projected/2ab08149-16b3-42ee-be06-5e663fad2df0-kube-api-access-k5s5p\") pod \"2ab08149-16b3-42ee-be06-5e663fad2df0\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.247671 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-dns-svc\") pod \"2ab08149-16b3-42ee-be06-5e663fad2df0\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.247837 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-ovsdbserver-nb\") pod \"2ab08149-16b3-42ee-be06-5e663fad2df0\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.247920 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-config\") pod \"2ab08149-16b3-42ee-be06-5e663fad2df0\" (UID: \"2ab08149-16b3-42ee-be06-5e663fad2df0\") " Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.255799 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ab08149-16b3-42ee-be06-5e663fad2df0-kube-api-access-k5s5p" (OuterVolumeSpecName: "kube-api-access-k5s5p") pod "2ab08149-16b3-42ee-be06-5e663fad2df0" (UID: "2ab08149-16b3-42ee-be06-5e663fad2df0"). InnerVolumeSpecName "kube-api-access-k5s5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.301915 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2ab08149-16b3-42ee-be06-5e663fad2df0" (UID: "2ab08149-16b3-42ee-be06-5e663fad2df0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.305747 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2ab08149-16b3-42ee-be06-5e663fad2df0" (UID: "2ab08149-16b3-42ee-be06-5e663fad2df0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.307041 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-config" (OuterVolumeSpecName: "config") pod "2ab08149-16b3-42ee-be06-5e663fad2df0" (UID: "2ab08149-16b3-42ee-be06-5e663fad2df0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.349557 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.349591 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.349604 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ab08149-16b3-42ee-be06-5e663fad2df0-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.349613 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5s5p\" (UniqueName: \"kubernetes.io/projected/2ab08149-16b3-42ee-be06-5e663fad2df0-kube-api-access-k5s5p\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.453678 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77cf9b784c-524z4"] Feb 25 07:06:15 crc kubenswrapper[4978]: W0225 07:06:15.455695 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod203f2e43_da8a_4be2_a1bb_6a8d5fdc9038.slice/crio-6e793f91d8434d976c51290f95772c5b4b1b7e6970d7c573e210f3569032b141 WatchSource:0}: Error finding container 6e793f91d8434d976c51290f95772c5b4b1b7e6970d7c573e210f3569032b141: Status 404 returned error can't find the container with id 6e793f91d8434d976c51290f95772c5b4b1b7e6970d7c573e210f3569032b141 Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.704784 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 25 07:06:15 crc kubenswrapper[4978]: E0225 07:06:15.705113 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab08149-16b3-42ee-be06-5e663fad2df0" containerName="init" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.705129 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab08149-16b3-42ee-be06-5e663fad2df0" containerName="init" Feb 25 07:06:15 crc kubenswrapper[4978]: E0225 07:06:15.705156 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ab08149-16b3-42ee-be06-5e663fad2df0" containerName="dnsmasq-dns" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.705163 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ab08149-16b3-42ee-be06-5e663fad2df0" containerName="dnsmasq-dns" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.705312 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ab08149-16b3-42ee-be06-5e663fad2df0" containerName="dnsmasq-dns" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.709827 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.711550 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.713051 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-f6svz" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.713141 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.713227 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.727898 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.857851 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-cache\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.858124 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwbxf\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-kube-api-access-lwbxf\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.858223 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-lock\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.858313 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.858463 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.858575 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.908532 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.960403 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.960885 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.960989 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-cache\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.961102 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwbxf\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-kube-api-access-lwbxf\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.961188 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-lock\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.961268 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.961594 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-cache\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: E0225 07:06:15.960595 4978 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 25 07:06:15 crc kubenswrapper[4978]: E0225 07:06:15.961667 4978 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 25 07:06:15 crc kubenswrapper[4978]: E0225 07:06:15.961720 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift podName:9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887 nodeName:}" failed. No retries permitted until 2026-02-25 07:06:16.461700564 +0000 UTC m=+1269.900957043 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift") pod "swift-storage-0" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887") : configmap "swift-ring-files" not found Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.961802 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.961922 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-lock\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.967002 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.983536 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:15 crc kubenswrapper[4978]: I0225 07:06:15.984461 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwbxf\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-kube-api-access-lwbxf\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.074618 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" event={"ID":"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038","Type":"ContainerStarted","Data":"6e793f91d8434d976c51290f95772c5b4b1b7e6970d7c573e210f3569032b141"} Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.084542 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" event={"ID":"2ab08149-16b3-42ee-be06-5e663fad2df0","Type":"ContainerDied","Data":"c2f14279704516a369463eb86c7376a81c146ff90150417ef9e703357bde568d"} Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.084597 4978 scope.go:117] "RemoveContainer" containerID="5deb0f5164203d2b369cebbc143c949a54c18541d08040ce69321a3538515667" Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.084743 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-649c9df595-qpm9m" Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.087667 4978 generic.go:334] "Generic (PLEG): container finished" podID="608f63c7-1251-42ea-996d-c7629a992f4a" containerID="9569ef92f4606b61b4ee144b77430b5109d298241ce9e4d4f08627b799dfe74a" exitCode=0 Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.087840 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cb34-account-create-update-k6dk6" event={"ID":"608f63c7-1251-42ea-996d-c7629a992f4a","Type":"ContainerDied","Data":"9569ef92f4606b61b4ee144b77430b5109d298241ce9e4d4f08627b799dfe74a"} Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.120905 4978 scope.go:117] "RemoveContainer" containerID="be274f99b956d30f11bd4b5e8b166fe685a787c558c954c8d45787aaf386cdcf" Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.128284 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-649c9df595-qpm9m"] Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.136791 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-649c9df595-qpm9m"] Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.477909 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:16 crc kubenswrapper[4978]: E0225 07:06:16.478991 4978 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 25 07:06:16 crc kubenswrapper[4978]: E0225 07:06:16.479031 4978 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 25 07:06:16 crc kubenswrapper[4978]: E0225 07:06:16.479106 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift podName:9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887 nodeName:}" failed. No retries permitted until 2026-02-25 07:06:17.479084058 +0000 UTC m=+1270.918340517 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift") pod "swift-storage-0" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887") : configmap "swift-ring-files" not found Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.484664 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4b94" Feb 25 07:06:16 crc kubenswrapper[4978]: E0225 07:06:16.520516 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod203f2e43_da8a_4be2_a1bb_6a8d5fdc9038.slice/crio-conmon-a5ccdb669ac03bfcffb881734e74afcc02b508fb8978b1671db8c00879702e58.scope\": RecentStats: unable to find data in memory cache]" Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.581894 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c183deb4-f193-4873-9617-5447f5c36e35-operator-scripts\") pod \"c183deb4-f193-4873-9617-5447f5c36e35\" (UID: \"c183deb4-f193-4873-9617-5447f5c36e35\") " Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.581969 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgvkx\" (UniqueName: \"kubernetes.io/projected/c183deb4-f193-4873-9617-5447f5c36e35-kube-api-access-kgvkx\") pod \"c183deb4-f193-4873-9617-5447f5c36e35\" (UID: \"c183deb4-f193-4873-9617-5447f5c36e35\") " Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.583889 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c183deb4-f193-4873-9617-5447f5c36e35-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c183deb4-f193-4873-9617-5447f5c36e35" (UID: "c183deb4-f193-4873-9617-5447f5c36e35"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.590566 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c183deb4-f193-4873-9617-5447f5c36e35-kube-api-access-kgvkx" (OuterVolumeSpecName: "kube-api-access-kgvkx") pod "c183deb4-f193-4873-9617-5447f5c36e35" (UID: "c183deb4-f193-4873-9617-5447f5c36e35"). InnerVolumeSpecName "kube-api-access-kgvkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.684484 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c183deb4-f193-4873-9617-5447f5c36e35-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:16 crc kubenswrapper[4978]: I0225 07:06:16.684524 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgvkx\" (UniqueName: \"kubernetes.io/projected/c183deb4-f193-4873-9617-5447f5c36e35-kube-api-access-kgvkx\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.098970 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-d4b94" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.098963 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-d4b94" event={"ID":"c183deb4-f193-4873-9617-5447f5c36e35","Type":"ContainerDied","Data":"671963c21204a5fdbfd123e6aa150af2ebb491107bee632f7ddcdc88079cfda4"} Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.099582 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="671963c21204a5fdbfd123e6aa150af2ebb491107bee632f7ddcdc88079cfda4" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.101502 4978 generic.go:334] "Generic (PLEG): container finished" podID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerID="a5ccdb669ac03bfcffb881734e74afcc02b508fb8978b1671db8c00879702e58" exitCode=0 Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.101565 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" event={"ID":"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038","Type":"ContainerDied","Data":"a5ccdb669ac03bfcffb881734e74afcc02b508fb8978b1671db8c00879702e58"} Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.339080 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ab08149-16b3-42ee-be06-5e663fad2df0" path="/var/lib/kubelet/pods/2ab08149-16b3-42ee-be06-5e663fad2df0/volumes" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.381016 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-zszqp"] Feb 25 07:06:17 crc kubenswrapper[4978]: E0225 07:06:17.381625 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c183deb4-f193-4873-9617-5447f5c36e35" containerName="mariadb-database-create" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.381651 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c183deb4-f193-4873-9617-5447f5c36e35" containerName="mariadb-database-create" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.381887 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c183deb4-f193-4873-9617-5447f5c36e35" containerName="mariadb-database-create" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.382471 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zszqp" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.394008 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zszqp"] Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.461788 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-1b7b-account-create-update-z4plm"] Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.462681 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1b7b-account-create-update-z4plm" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.467591 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.474828 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-1b7b-account-create-update-z4plm"] Feb 25 07:06:17 crc kubenswrapper[4978]: E0225 07:06:17.496761 4978 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.497015 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:17 crc kubenswrapper[4978]: E0225 07:06:17.497464 4978 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 25 07:06:17 crc kubenswrapper[4978]: E0225 07:06:17.497623 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift podName:9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887 nodeName:}" failed. No retries permitted until 2026-02-25 07:06:19.497600227 +0000 UTC m=+1272.936856686 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift") pod "swift-storage-0" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887") : configmap "swift-ring-files" not found Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.499621 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2a54702-6977-47b4-9e4e-11bc125b4981-operator-scripts\") pod \"glance-db-create-zszqp\" (UID: \"e2a54702-6977-47b4-9e4e-11bc125b4981\") " pod="openstack/glance-db-create-zszqp" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.499723 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2grsv\" (UniqueName: \"kubernetes.io/projected/e2a54702-6977-47b4-9e4e-11bc125b4981-kube-api-access-2grsv\") pod \"glance-db-create-zszqp\" (UID: \"e2a54702-6977-47b4-9e4e-11bc125b4981\") " pod="openstack/glance-db-create-zszqp" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.519952 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cb34-account-create-update-k6dk6" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.601388 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608f63c7-1251-42ea-996d-c7629a992f4a-operator-scripts\") pod \"608f63c7-1251-42ea-996d-c7629a992f4a\" (UID: \"608f63c7-1251-42ea-996d-c7629a992f4a\") " Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.601775 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfnhv\" (UniqueName: \"kubernetes.io/projected/608f63c7-1251-42ea-996d-c7629a992f4a-kube-api-access-lfnhv\") pod \"608f63c7-1251-42ea-996d-c7629a992f4a\" (UID: \"608f63c7-1251-42ea-996d-c7629a992f4a\") " Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.602035 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2a54702-6977-47b4-9e4e-11bc125b4981-operator-scripts\") pod \"glance-db-create-zszqp\" (UID: \"e2a54702-6977-47b4-9e4e-11bc125b4981\") " pod="openstack/glance-db-create-zszqp" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.602113 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2grsv\" (UniqueName: \"kubernetes.io/projected/e2a54702-6977-47b4-9e4e-11bc125b4981-kube-api-access-2grsv\") pod \"glance-db-create-zszqp\" (UID: \"e2a54702-6977-47b4-9e4e-11bc125b4981\") " pod="openstack/glance-db-create-zszqp" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.602240 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b86b8158-9329-44e8-9268-d5ed64155d48-operator-scripts\") pod \"glance-1b7b-account-create-update-z4plm\" (UID: \"b86b8158-9329-44e8-9268-d5ed64155d48\") " pod="openstack/glance-1b7b-account-create-update-z4plm" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.602334 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmhkn\" (UniqueName: \"kubernetes.io/projected/b86b8158-9329-44e8-9268-d5ed64155d48-kube-api-access-hmhkn\") pod \"glance-1b7b-account-create-update-z4plm\" (UID: \"b86b8158-9329-44e8-9268-d5ed64155d48\") " pod="openstack/glance-1b7b-account-create-update-z4plm" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.602805 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/608f63c7-1251-42ea-996d-c7629a992f4a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "608f63c7-1251-42ea-996d-c7629a992f4a" (UID: "608f63c7-1251-42ea-996d-c7629a992f4a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.603193 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2a54702-6977-47b4-9e4e-11bc125b4981-operator-scripts\") pod \"glance-db-create-zszqp\" (UID: \"e2a54702-6977-47b4-9e4e-11bc125b4981\") " pod="openstack/glance-db-create-zszqp" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.607963 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/608f63c7-1251-42ea-996d-c7629a992f4a-kube-api-access-lfnhv" (OuterVolumeSpecName: "kube-api-access-lfnhv") pod "608f63c7-1251-42ea-996d-c7629a992f4a" (UID: "608f63c7-1251-42ea-996d-c7629a992f4a"). InnerVolumeSpecName "kube-api-access-lfnhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.626034 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2grsv\" (UniqueName: \"kubernetes.io/projected/e2a54702-6977-47b4-9e4e-11bc125b4981-kube-api-access-2grsv\") pod \"glance-db-create-zszqp\" (UID: \"e2a54702-6977-47b4-9e4e-11bc125b4981\") " pod="openstack/glance-db-create-zszqp" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.702488 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zszqp" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.704189 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b86b8158-9329-44e8-9268-d5ed64155d48-operator-scripts\") pod \"glance-1b7b-account-create-update-z4plm\" (UID: \"b86b8158-9329-44e8-9268-d5ed64155d48\") " pod="openstack/glance-1b7b-account-create-update-z4plm" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.704432 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmhkn\" (UniqueName: \"kubernetes.io/projected/b86b8158-9329-44e8-9268-d5ed64155d48-kube-api-access-hmhkn\") pod \"glance-1b7b-account-create-update-z4plm\" (UID: \"b86b8158-9329-44e8-9268-d5ed64155d48\") " pod="openstack/glance-1b7b-account-create-update-z4plm" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.704822 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b86b8158-9329-44e8-9268-d5ed64155d48-operator-scripts\") pod \"glance-1b7b-account-create-update-z4plm\" (UID: \"b86b8158-9329-44e8-9268-d5ed64155d48\") " pod="openstack/glance-1b7b-account-create-update-z4plm" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.705410 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfnhv\" (UniqueName: \"kubernetes.io/projected/608f63c7-1251-42ea-996d-c7629a992f4a-kube-api-access-lfnhv\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.705460 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608f63c7-1251-42ea-996d-c7629a992f4a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.720942 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmhkn\" (UniqueName: \"kubernetes.io/projected/b86b8158-9329-44e8-9268-d5ed64155d48-kube-api-access-hmhkn\") pod \"glance-1b7b-account-create-update-z4plm\" (UID: \"b86b8158-9329-44e8-9268-d5ed64155d48\") " pod="openstack/glance-1b7b-account-create-update-z4plm" Feb 25 07:06:17 crc kubenswrapper[4978]: I0225 07:06:17.780637 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1b7b-account-create-update-z4plm" Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.112006 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cb34-account-create-update-k6dk6" event={"ID":"608f63c7-1251-42ea-996d-c7629a992f4a","Type":"ContainerDied","Data":"6b3cd74a0eb9fd53e0127195d842f998886650fb40418a96a70744db0b7f3603"} Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.112264 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b3cd74a0eb9fd53e0127195d842f998886650fb40418a96a70744db0b7f3603" Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.112038 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cb34-account-create-update-k6dk6" Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.114570 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" event={"ID":"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038","Type":"ContainerStarted","Data":"07a057de05f5bc6d8d6e12685b1aaa0d0969c94b182f9211e4ace6841f755b5b"} Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.115817 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.141572 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" podStartSLOduration=4.1415506539999996 podStartE2EDuration="4.141550654s" podCreationTimestamp="2026-02-25 07:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:18.137229654 +0000 UTC m=+1271.576486113" watchObservedRunningTime="2026-02-25 07:06:18.141550654 +0000 UTC m=+1271.580807133" Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.168356 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-zszqp"] Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.268733 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-1b7b-account-create-update-z4plm"] Feb 25 07:06:18 crc kubenswrapper[4978]: W0225 07:06:18.270858 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb86b8158_9329_44e8_9268_d5ed64155d48.slice/crio-43fb51223268f5d3633525ef74a6db3b5c9792d535bbf7cc91189fa0f55b1952 WatchSource:0}: Error finding container 43fb51223268f5d3633525ef74a6db3b5c9792d535bbf7cc91189fa0f55b1952: Status 404 returned error can't find the container with id 43fb51223268f5d3633525ef74a6db3b5c9792d535bbf7cc91189fa0f55b1952 Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.889795 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-4lv9d"] Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.910264 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-4lv9d"] Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.929617 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-f7574"] Feb 25 07:06:18 crc kubenswrapper[4978]: E0225 07:06:18.929999 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="608f63c7-1251-42ea-996d-c7629a992f4a" containerName="mariadb-account-create-update" Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.930015 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="608f63c7-1251-42ea-996d-c7629a992f4a" containerName="mariadb-account-create-update" Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.930212 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="608f63c7-1251-42ea-996d-c7629a992f4a" containerName="mariadb-account-create-update" Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.930773 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-f7574" Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.932715 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 25 07:06:18 crc kubenswrapper[4978]: I0225 07:06:18.936271 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-f7574"] Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.032069 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce72ec5-eb9b-46f3-ae30-b8380615465c-operator-scripts\") pod \"root-account-create-update-f7574\" (UID: \"1ce72ec5-eb9b-46f3-ae30-b8380615465c\") " pod="openstack/root-account-create-update-f7574" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.032154 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb9bg\" (UniqueName: \"kubernetes.io/projected/1ce72ec5-eb9b-46f3-ae30-b8380615465c-kube-api-access-lb9bg\") pod \"root-account-create-update-f7574\" (UID: \"1ce72ec5-eb9b-46f3-ae30-b8380615465c\") " pod="openstack/root-account-create-update-f7574" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.122564 4978 generic.go:334] "Generic (PLEG): container finished" podID="b86b8158-9329-44e8-9268-d5ed64155d48" containerID="5b819e72e4b09f85c33501291e203299fcd870de1d74deb4a6f337a81b3e6c57" exitCode=0 Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.122967 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1b7b-account-create-update-z4plm" event={"ID":"b86b8158-9329-44e8-9268-d5ed64155d48","Type":"ContainerDied","Data":"5b819e72e4b09f85c33501291e203299fcd870de1d74deb4a6f337a81b3e6c57"} Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.122996 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1b7b-account-create-update-z4plm" event={"ID":"b86b8158-9329-44e8-9268-d5ed64155d48","Type":"ContainerStarted","Data":"43fb51223268f5d3633525ef74a6db3b5c9792d535bbf7cc91189fa0f55b1952"} Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.124498 4978 generic.go:334] "Generic (PLEG): container finished" podID="e2a54702-6977-47b4-9e4e-11bc125b4981" containerID="38fe35a66c4eb783bff693706b13667ecdb4496a617dc255fc5aac1f2f5a3bc3" exitCode=0 Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.124557 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zszqp" event={"ID":"e2a54702-6977-47b4-9e4e-11bc125b4981","Type":"ContainerDied","Data":"38fe35a66c4eb783bff693706b13667ecdb4496a617dc255fc5aac1f2f5a3bc3"} Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.124577 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zszqp" event={"ID":"e2a54702-6977-47b4-9e4e-11bc125b4981","Type":"ContainerStarted","Data":"379ed4e2cb0bdd2bb7d3dcc9ec82385fb61dbb2a25044a3530da40de9809e1c1"} Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.133812 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce72ec5-eb9b-46f3-ae30-b8380615465c-operator-scripts\") pod \"root-account-create-update-f7574\" (UID: \"1ce72ec5-eb9b-46f3-ae30-b8380615465c\") " pod="openstack/root-account-create-update-f7574" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.134541 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce72ec5-eb9b-46f3-ae30-b8380615465c-operator-scripts\") pod \"root-account-create-update-f7574\" (UID: \"1ce72ec5-eb9b-46f3-ae30-b8380615465c\") " pod="openstack/root-account-create-update-f7574" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.134675 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb9bg\" (UniqueName: \"kubernetes.io/projected/1ce72ec5-eb9b-46f3-ae30-b8380615465c-kube-api-access-lb9bg\") pod \"root-account-create-update-f7574\" (UID: \"1ce72ec5-eb9b-46f3-ae30-b8380615465c\") " pod="openstack/root-account-create-update-f7574" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.155111 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb9bg\" (UniqueName: \"kubernetes.io/projected/1ce72ec5-eb9b-46f3-ae30-b8380615465c-kube-api-access-lb9bg\") pod \"root-account-create-update-f7574\" (UID: \"1ce72ec5-eb9b-46f3-ae30-b8380615465c\") " pod="openstack/root-account-create-update-f7574" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.247680 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-f7574" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.336208 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="889b0ea1-aa03-4331-8247-d29684abdcdc" path="/var/lib/kubelet/pods/889b0ea1-aa03-4331-8247-d29684abdcdc/volumes" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.549542 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:19 crc kubenswrapper[4978]: E0225 07:06:19.549729 4978 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 25 07:06:19 crc kubenswrapper[4978]: E0225 07:06:19.549756 4978 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 25 07:06:19 crc kubenswrapper[4978]: E0225 07:06:19.549811 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift podName:9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887 nodeName:}" failed. No retries permitted until 2026-02-25 07:06:23.549795162 +0000 UTC m=+1276.989051621 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift") pod "swift-storage-0" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887") : configmap "swift-ring-files" not found Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.632610 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-z5fgm"] Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.634202 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.635983 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.636405 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.637150 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.655085 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-z5fgm"] Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.740218 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-f7574"] Feb 25 07:06:19 crc kubenswrapper[4978]: W0225 07:06:19.742146 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ce72ec5_eb9b_46f3_ae30_b8380615465c.slice/crio-24bd4c593dfa42d7f0ba1fc351339e5779f0a92d6590565c4eeaa9cd2efcb4bd WatchSource:0}: Error finding container 24bd4c593dfa42d7f0ba1fc351339e5779f0a92d6590565c4eeaa9cd2efcb4bd: Status 404 returned error can't find the container with id 24bd4c593dfa42d7f0ba1fc351339e5779f0a92d6590565c4eeaa9cd2efcb4bd Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.751844 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-swiftconf\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.751887 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-dispersionconf\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.751912 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-scripts\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.751935 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7vdm\" (UniqueName: \"kubernetes.io/projected/330d6675-e003-4a9c-862a-88a54e67877e-kube-api-access-m7vdm\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.752514 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-ring-data-devices\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.752598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-combined-ca-bundle\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.752698 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/330d6675-e003-4a9c-862a-88a54e67877e-etc-swift\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.854793 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-ring-data-devices\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.854860 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-combined-ca-bundle\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.854899 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/330d6675-e003-4a9c-862a-88a54e67877e-etc-swift\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.854952 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-swiftconf\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.854973 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-dispersionconf\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.854998 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-scripts\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.855018 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7vdm\" (UniqueName: \"kubernetes.io/projected/330d6675-e003-4a9c-862a-88a54e67877e-kube-api-access-m7vdm\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.856193 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-ring-data-devices\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.857434 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/330d6675-e003-4a9c-862a-88a54e67877e-etc-swift\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.858125 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-scripts\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.861539 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-dispersionconf\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.863624 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-swiftconf\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.864224 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-combined-ca-bundle\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.875631 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7vdm\" (UniqueName: \"kubernetes.io/projected/330d6675-e003-4a9c-862a-88a54e67877e-kube-api-access-m7vdm\") pod \"swift-ring-rebalance-z5fgm\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:19 crc kubenswrapper[4978]: I0225 07:06:19.988583 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.147529 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-f7574" event={"ID":"1ce72ec5-eb9b-46f3-ae30-b8380615465c","Type":"ContainerStarted","Data":"392249f3ec789cc057ca60e633474a4a6604c180310a30bdc6ddcca23a6427f4"} Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.147575 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-f7574" event={"ID":"1ce72ec5-eb9b-46f3-ae30-b8380615465c","Type":"ContainerStarted","Data":"24bd4c593dfa42d7f0ba1fc351339e5779f0a92d6590565c4eeaa9cd2efcb4bd"} Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.168666 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-f7574" podStartSLOduration=2.168645503 podStartE2EDuration="2.168645503s" podCreationTimestamp="2026-02-25 07:06:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:20.161186179 +0000 UTC m=+1273.600442638" watchObservedRunningTime="2026-02-25 07:06:20.168645503 +0000 UTC m=+1273.607901962" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.236272 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-z5fgm"] Feb 25 07:06:20 crc kubenswrapper[4978]: W0225 07:06:20.244551 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod330d6675_e003_4a9c_862a_88a54e67877e.slice/crio-15b7e45b2207a087212683967f513ec289c7f8f0e1faa240f8ed6692d7388a95 WatchSource:0}: Error finding container 15b7e45b2207a087212683967f513ec289c7f8f0e1faa240f8ed6692d7388a95: Status 404 returned error can't find the container with id 15b7e45b2207a087212683967f513ec289c7f8f0e1faa240f8ed6692d7388a95 Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.501592 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zszqp" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.526437 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1b7b-account-create-update-z4plm" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.667099 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmhkn\" (UniqueName: \"kubernetes.io/projected/b86b8158-9329-44e8-9268-d5ed64155d48-kube-api-access-hmhkn\") pod \"b86b8158-9329-44e8-9268-d5ed64155d48\" (UID: \"b86b8158-9329-44e8-9268-d5ed64155d48\") " Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.667165 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b86b8158-9329-44e8-9268-d5ed64155d48-operator-scripts\") pod \"b86b8158-9329-44e8-9268-d5ed64155d48\" (UID: \"b86b8158-9329-44e8-9268-d5ed64155d48\") " Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.667195 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2grsv\" (UniqueName: \"kubernetes.io/projected/e2a54702-6977-47b4-9e4e-11bc125b4981-kube-api-access-2grsv\") pod \"e2a54702-6977-47b4-9e4e-11bc125b4981\" (UID: \"e2a54702-6977-47b4-9e4e-11bc125b4981\") " Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.667327 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2a54702-6977-47b4-9e4e-11bc125b4981-operator-scripts\") pod \"e2a54702-6977-47b4-9e4e-11bc125b4981\" (UID: \"e2a54702-6977-47b4-9e4e-11bc125b4981\") " Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.668061 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b86b8158-9329-44e8-9268-d5ed64155d48-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b86b8158-9329-44e8-9268-d5ed64155d48" (UID: "b86b8158-9329-44e8-9268-d5ed64155d48"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.668088 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2a54702-6977-47b4-9e4e-11bc125b4981-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e2a54702-6977-47b4-9e4e-11bc125b4981" (UID: "e2a54702-6977-47b4-9e4e-11bc125b4981"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.672740 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2a54702-6977-47b4-9e4e-11bc125b4981-kube-api-access-2grsv" (OuterVolumeSpecName: "kube-api-access-2grsv") pod "e2a54702-6977-47b4-9e4e-11bc125b4981" (UID: "e2a54702-6977-47b4-9e4e-11bc125b4981"). InnerVolumeSpecName "kube-api-access-2grsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.673106 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b86b8158-9329-44e8-9268-d5ed64155d48-kube-api-access-hmhkn" (OuterVolumeSpecName: "kube-api-access-hmhkn") pod "b86b8158-9329-44e8-9268-d5ed64155d48" (UID: "b86b8158-9329-44e8-9268-d5ed64155d48"). InnerVolumeSpecName "kube-api-access-hmhkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.769461 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2a54702-6977-47b4-9e4e-11bc125b4981-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.769494 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmhkn\" (UniqueName: \"kubernetes.io/projected/b86b8158-9329-44e8-9268-d5ed64155d48-kube-api-access-hmhkn\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.769505 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b86b8158-9329-44e8-9268-d5ed64155d48-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:20 crc kubenswrapper[4978]: I0225 07:06:20.769513 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2grsv\" (UniqueName: \"kubernetes.io/projected/e2a54702-6977-47b4-9e4e-11bc125b4981-kube-api-access-2grsv\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:21 crc kubenswrapper[4978]: I0225 07:06:21.167032 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z5fgm" event={"ID":"330d6675-e003-4a9c-862a-88a54e67877e","Type":"ContainerStarted","Data":"15b7e45b2207a087212683967f513ec289c7f8f0e1faa240f8ed6692d7388a95"} Feb 25 07:06:21 crc kubenswrapper[4978]: I0225 07:06:21.170901 4978 generic.go:334] "Generic (PLEG): container finished" podID="1ce72ec5-eb9b-46f3-ae30-b8380615465c" containerID="392249f3ec789cc057ca60e633474a4a6604c180310a30bdc6ddcca23a6427f4" exitCode=0 Feb 25 07:06:21 crc kubenswrapper[4978]: I0225 07:06:21.171007 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-f7574" event={"ID":"1ce72ec5-eb9b-46f3-ae30-b8380615465c","Type":"ContainerDied","Data":"392249f3ec789cc057ca60e633474a4a6604c180310a30bdc6ddcca23a6427f4"} Feb 25 07:06:21 crc kubenswrapper[4978]: I0225 07:06:21.175595 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-zszqp" event={"ID":"e2a54702-6977-47b4-9e4e-11bc125b4981","Type":"ContainerDied","Data":"379ed4e2cb0bdd2bb7d3dcc9ec82385fb61dbb2a25044a3530da40de9809e1c1"} Feb 25 07:06:21 crc kubenswrapper[4978]: I0225 07:06:21.175628 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="379ed4e2cb0bdd2bb7d3dcc9ec82385fb61dbb2a25044a3530da40de9809e1c1" Feb 25 07:06:21 crc kubenswrapper[4978]: I0225 07:06:21.175689 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-zszqp" Feb 25 07:06:21 crc kubenswrapper[4978]: I0225 07:06:21.179670 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1b7b-account-create-update-z4plm" event={"ID":"b86b8158-9329-44e8-9268-d5ed64155d48","Type":"ContainerDied","Data":"43fb51223268f5d3633525ef74a6db3b5c9792d535bbf7cc91189fa0f55b1952"} Feb 25 07:06:21 crc kubenswrapper[4978]: I0225 07:06:21.179718 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43fb51223268f5d3633525ef74a6db3b5c9792d535bbf7cc91189fa0f55b1952" Feb 25 07:06:21 crc kubenswrapper[4978]: I0225 07:06:21.179695 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1b7b-account-create-update-z4plm" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.603087 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-7ts6t"] Feb 25 07:06:22 crc kubenswrapper[4978]: E0225 07:06:22.603890 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b86b8158-9329-44e8-9268-d5ed64155d48" containerName="mariadb-account-create-update" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.603902 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b86b8158-9329-44e8-9268-d5ed64155d48" containerName="mariadb-account-create-update" Feb 25 07:06:22 crc kubenswrapper[4978]: E0225 07:06:22.603930 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2a54702-6977-47b4-9e4e-11bc125b4981" containerName="mariadb-database-create" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.603936 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2a54702-6977-47b4-9e4e-11bc125b4981" containerName="mariadb-database-create" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.604082 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b86b8158-9329-44e8-9268-d5ed64155d48" containerName="mariadb-account-create-update" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.604099 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2a54702-6977-47b4-9e4e-11bc125b4981" containerName="mariadb-database-create" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.604800 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.606750 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mrzl7" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.606801 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.613867 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-7ts6t"] Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.700142 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mqrr\" (UniqueName: \"kubernetes.io/projected/4bd97fe5-291b-4130-98c7-4ffbc0cba189-kube-api-access-6mqrr\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.700203 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-db-sync-config-data\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.700399 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-combined-ca-bundle\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.700546 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-config-data\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.802184 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-config-data\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.802280 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mqrr\" (UniqueName: \"kubernetes.io/projected/4bd97fe5-291b-4130-98c7-4ffbc0cba189-kube-api-access-6mqrr\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.802318 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-db-sync-config-data\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.802393 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-combined-ca-bundle\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.808845 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-combined-ca-bundle\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.809472 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-db-sync-config-data\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.814854 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-config-data\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.823106 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mqrr\" (UniqueName: \"kubernetes.io/projected/4bd97fe5-291b-4130-98c7-4ffbc0cba189-kube-api-access-6mqrr\") pod \"glance-db-sync-7ts6t\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:22 crc kubenswrapper[4978]: I0225 07:06:22.922427 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7ts6t" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.089533 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-kvbg5"] Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.090672 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kvbg5" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.100527 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-kvbg5"] Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.189507 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-56f1-account-create-update-5zpqd"] Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.190436 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56f1-account-create-update-5zpqd" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.193219 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.200063 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-56f1-account-create-update-5zpqd"] Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.211359 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-682d5\" (UniqueName: \"kubernetes.io/projected/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-kube-api-access-682d5\") pod \"keystone-db-create-kvbg5\" (UID: \"c953f23d-4f0e-4be8-9b37-b26b7de7cad4\") " pod="openstack/keystone-db-create-kvbg5" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.211441 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-operator-scripts\") pod \"keystone-db-create-kvbg5\" (UID: \"c953f23d-4f0e-4be8-9b37-b26b7de7cad4\") " pod="openstack/keystone-db-create-kvbg5" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.312303 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-operator-scripts\") pod \"keystone-db-create-kvbg5\" (UID: \"c953f23d-4f0e-4be8-9b37-b26b7de7cad4\") " pod="openstack/keystone-db-create-kvbg5" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.312383 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhhkb\" (UniqueName: \"kubernetes.io/projected/d31c2c4a-d61c-496d-aef6-31895fee244f-kube-api-access-hhhkb\") pod \"keystone-56f1-account-create-update-5zpqd\" (UID: \"d31c2c4a-d61c-496d-aef6-31895fee244f\") " pod="openstack/keystone-56f1-account-create-update-5zpqd" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.312421 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d31c2c4a-d61c-496d-aef6-31895fee244f-operator-scripts\") pod \"keystone-56f1-account-create-update-5zpqd\" (UID: \"d31c2c4a-d61c-496d-aef6-31895fee244f\") " pod="openstack/keystone-56f1-account-create-update-5zpqd" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.312508 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-682d5\" (UniqueName: \"kubernetes.io/projected/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-kube-api-access-682d5\") pod \"keystone-db-create-kvbg5\" (UID: \"c953f23d-4f0e-4be8-9b37-b26b7de7cad4\") " pod="openstack/keystone-db-create-kvbg5" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.313574 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-operator-scripts\") pod \"keystone-db-create-kvbg5\" (UID: \"c953f23d-4f0e-4be8-9b37-b26b7de7cad4\") " pod="openstack/keystone-db-create-kvbg5" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.335919 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-682d5\" (UniqueName: \"kubernetes.io/projected/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-kube-api-access-682d5\") pod \"keystone-db-create-kvbg5\" (UID: \"c953f23d-4f0e-4be8-9b37-b26b7de7cad4\") " pod="openstack/keystone-db-create-kvbg5" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.417360 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhhkb\" (UniqueName: \"kubernetes.io/projected/d31c2c4a-d61c-496d-aef6-31895fee244f-kube-api-access-hhhkb\") pod \"keystone-56f1-account-create-update-5zpqd\" (UID: \"d31c2c4a-d61c-496d-aef6-31895fee244f\") " pod="openstack/keystone-56f1-account-create-update-5zpqd" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.417449 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d31c2c4a-d61c-496d-aef6-31895fee244f-operator-scripts\") pod \"keystone-56f1-account-create-update-5zpqd\" (UID: \"d31c2c4a-d61c-496d-aef6-31895fee244f\") " pod="openstack/keystone-56f1-account-create-update-5zpqd" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.418536 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d31c2c4a-d61c-496d-aef6-31895fee244f-operator-scripts\") pod \"keystone-56f1-account-create-update-5zpqd\" (UID: \"d31c2c4a-d61c-496d-aef6-31895fee244f\") " pod="openstack/keystone-56f1-account-create-update-5zpqd" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.425571 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kvbg5" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.437772 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhhkb\" (UniqueName: \"kubernetes.io/projected/d31c2c4a-d61c-496d-aef6-31895fee244f-kube-api-access-hhhkb\") pod \"keystone-56f1-account-create-update-5zpqd\" (UID: \"d31c2c4a-d61c-496d-aef6-31895fee244f\") " pod="openstack/keystone-56f1-account-create-update-5zpqd" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.517206 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56f1-account-create-update-5zpqd" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.621315 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:23 crc kubenswrapper[4978]: E0225 07:06:23.621491 4978 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 25 07:06:23 crc kubenswrapper[4978]: E0225 07:06:23.621507 4978 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 25 07:06:23 crc kubenswrapper[4978]: E0225 07:06:23.621565 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift podName:9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887 nodeName:}" failed. No retries permitted until 2026-02-25 07:06:31.62154757 +0000 UTC m=+1285.060804029 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift") pod "swift-storage-0" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887") : configmap "swift-ring-files" not found Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.798389 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-f7574" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.925663 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb9bg\" (UniqueName: \"kubernetes.io/projected/1ce72ec5-eb9b-46f3-ae30-b8380615465c-kube-api-access-lb9bg\") pod \"1ce72ec5-eb9b-46f3-ae30-b8380615465c\" (UID: \"1ce72ec5-eb9b-46f3-ae30-b8380615465c\") " Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.925803 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce72ec5-eb9b-46f3-ae30-b8380615465c-operator-scripts\") pod \"1ce72ec5-eb9b-46f3-ae30-b8380615465c\" (UID: \"1ce72ec5-eb9b-46f3-ae30-b8380615465c\") " Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.927057 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ce72ec5-eb9b-46f3-ae30-b8380615465c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ce72ec5-eb9b-46f3-ae30-b8380615465c" (UID: "1ce72ec5-eb9b-46f3-ae30-b8380615465c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:23 crc kubenswrapper[4978]: I0225 07:06:23.930320 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ce72ec5-eb9b-46f3-ae30-b8380615465c-kube-api-access-lb9bg" (OuterVolumeSpecName: "kube-api-access-lb9bg") pod "1ce72ec5-eb9b-46f3-ae30-b8380615465c" (UID: "1ce72ec5-eb9b-46f3-ae30-b8380615465c"). InnerVolumeSpecName "kube-api-access-lb9bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:24 crc kubenswrapper[4978]: I0225 07:06:24.027963 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb9bg\" (UniqueName: \"kubernetes.io/projected/1ce72ec5-eb9b-46f3-ae30-b8380615465c-kube-api-access-lb9bg\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:24 crc kubenswrapper[4978]: I0225 07:06:24.028247 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce72ec5-eb9b-46f3-ae30-b8380615465c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:24 crc kubenswrapper[4978]: I0225 07:06:24.222638 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z5fgm" event={"ID":"330d6675-e003-4a9c-862a-88a54e67877e","Type":"ContainerStarted","Data":"ea46e9a5ff7500141828c5e39da95b38f857b39bf42022aa0050cb398f5fd324"} Feb 25 07:06:24 crc kubenswrapper[4978]: I0225 07:06:24.224792 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-f7574" event={"ID":"1ce72ec5-eb9b-46f3-ae30-b8380615465c","Type":"ContainerDied","Data":"24bd4c593dfa42d7f0ba1fc351339e5779f0a92d6590565c4eeaa9cd2efcb4bd"} Feb 25 07:06:24 crc kubenswrapper[4978]: I0225 07:06:24.224833 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24bd4c593dfa42d7f0ba1fc351339e5779f0a92d6590565c4eeaa9cd2efcb4bd" Feb 25 07:06:24 crc kubenswrapper[4978]: I0225 07:06:24.224856 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-f7574" Feb 25 07:06:24 crc kubenswrapper[4978]: I0225 07:06:24.249747 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-z5fgm" podStartSLOduration=1.628073052 podStartE2EDuration="5.249721331s" podCreationTimestamp="2026-02-25 07:06:19 +0000 UTC" firstStartedPulling="2026-02-25 07:06:20.246870139 +0000 UTC m=+1273.686126598" lastFinishedPulling="2026-02-25 07:06:23.868518418 +0000 UTC m=+1277.307774877" observedRunningTime="2026-02-25 07:06:24.241907646 +0000 UTC m=+1277.681164115" watchObservedRunningTime="2026-02-25 07:06:24.249721331 +0000 UTC m=+1277.688977810" Feb 25 07:06:24 crc kubenswrapper[4978]: I0225 07:06:24.318522 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-kvbg5"] Feb 25 07:06:24 crc kubenswrapper[4978]: I0225 07:06:24.326083 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-56f1-account-create-update-5zpqd"] Feb 25 07:06:24 crc kubenswrapper[4978]: W0225 07:06:24.326720 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc953f23d_4f0e_4be8_9b37_b26b7de7cad4.slice/crio-dad917c48cf7501b9ea93e1d5b13c6c6ba7959712f85833f8be279d64ae5e6ca WatchSource:0}: Error finding container dad917c48cf7501b9ea93e1d5b13c6c6ba7959712f85833f8be279d64ae5e6ca: Status 404 returned error can't find the container with id dad917c48cf7501b9ea93e1d5b13c6c6ba7959712f85833f8be279d64ae5e6ca Feb 25 07:06:24 crc kubenswrapper[4978]: W0225 07:06:24.335801 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd31c2c4a_d61c_496d_aef6_31895fee244f.slice/crio-045b7b7fb439cae6ef32a798c585cfbd2c1f7b35d1c7023ed36b45679411cac1 WatchSource:0}: Error finding container 045b7b7fb439cae6ef32a798c585cfbd2c1f7b35d1c7023ed36b45679411cac1: Status 404 returned error can't find the container with id 045b7b7fb439cae6ef32a798c585cfbd2c1f7b35d1c7023ed36b45679411cac1 Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:24.583956 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-7ts6t"] Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:24.961648 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.009876 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9bcc5599-c5twm"] Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.010453 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" podUID="3115eb4c-6af8-41df-9d7e-4db587e37ac5" containerName="dnsmasq-dns" containerID="cri-o://bc7353a13e461bf5e8472c1a0de787f88ed960bbac64db951a039c2f381f9ce5" gracePeriod=10 Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.252231 4978 generic.go:334] "Generic (PLEG): container finished" podID="c953f23d-4f0e-4be8-9b37-b26b7de7cad4" containerID="1d9509bb185bfd8f66b16ea5e86fc0713bd07983268184b5f3f6005b2e7ab1b8" exitCode=0 Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.252305 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kvbg5" event={"ID":"c953f23d-4f0e-4be8-9b37-b26b7de7cad4","Type":"ContainerDied","Data":"1d9509bb185bfd8f66b16ea5e86fc0713bd07983268184b5f3f6005b2e7ab1b8"} Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.252389 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kvbg5" event={"ID":"c953f23d-4f0e-4be8-9b37-b26b7de7cad4","Type":"ContainerStarted","Data":"dad917c48cf7501b9ea93e1d5b13c6c6ba7959712f85833f8be279d64ae5e6ca"} Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.258135 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7ts6t" event={"ID":"4bd97fe5-291b-4130-98c7-4ffbc0cba189","Type":"ContainerStarted","Data":"91a0fb39ecfb77aaf572f0945651a4d418e46efeb2288c713cf21f90ad14899b"} Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.268740 4978 generic.go:334] "Generic (PLEG): container finished" podID="3115eb4c-6af8-41df-9d7e-4db587e37ac5" containerID="bc7353a13e461bf5e8472c1a0de787f88ed960bbac64db951a039c2f381f9ce5" exitCode=0 Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.268826 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" event={"ID":"3115eb4c-6af8-41df-9d7e-4db587e37ac5","Type":"ContainerDied","Data":"bc7353a13e461bf5e8472c1a0de787f88ed960bbac64db951a039c2f381f9ce5"} Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.277569 4978 generic.go:334] "Generic (PLEG): container finished" podID="d31c2c4a-d61c-496d-aef6-31895fee244f" containerID="6921f9a5eef1ac8d074e2d1d00534d70d54e38bc5537860be66cb698876cf0e1" exitCode=0 Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.277650 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-56f1-account-create-update-5zpqd" event={"ID":"d31c2c4a-d61c-496d-aef6-31895fee244f","Type":"ContainerDied","Data":"6921f9a5eef1ac8d074e2d1d00534d70d54e38bc5537860be66cb698876cf0e1"} Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.277712 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-56f1-account-create-update-5zpqd" event={"ID":"d31c2c4a-d61c-496d-aef6-31895fee244f","Type":"ContainerStarted","Data":"045b7b7fb439cae6ef32a798c585cfbd2c1f7b35d1c7023ed36b45679411cac1"} Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.724827 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.875642 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-nb\") pod \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.875974 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmpjw\" (UniqueName: \"kubernetes.io/projected/3115eb4c-6af8-41df-9d7e-4db587e37ac5-kube-api-access-zmpjw\") pod \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.876021 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-config\") pod \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.876053 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-dns-svc\") pod \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.876090 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-sb\") pod \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\" (UID: \"3115eb4c-6af8-41df-9d7e-4db587e37ac5\") " Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.881771 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3115eb4c-6af8-41df-9d7e-4db587e37ac5-kube-api-access-zmpjw" (OuterVolumeSpecName: "kube-api-access-zmpjw") pod "3115eb4c-6af8-41df-9d7e-4db587e37ac5" (UID: "3115eb4c-6af8-41df-9d7e-4db587e37ac5"). InnerVolumeSpecName "kube-api-access-zmpjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.925222 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3115eb4c-6af8-41df-9d7e-4db587e37ac5" (UID: "3115eb4c-6af8-41df-9d7e-4db587e37ac5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.926176 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3115eb4c-6af8-41df-9d7e-4db587e37ac5" (UID: "3115eb4c-6af8-41df-9d7e-4db587e37ac5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.927751 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-config" (OuterVolumeSpecName: "config") pod "3115eb4c-6af8-41df-9d7e-4db587e37ac5" (UID: "3115eb4c-6af8-41df-9d7e-4db587e37ac5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.930755 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3115eb4c-6af8-41df-9d7e-4db587e37ac5" (UID: "3115eb4c-6af8-41df-9d7e-4db587e37ac5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.978115 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.978141 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmpjw\" (UniqueName: \"kubernetes.io/projected/3115eb4c-6af8-41df-9d7e-4db587e37ac5-kube-api-access-zmpjw\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.978154 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.978161 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:25 crc kubenswrapper[4978]: I0225 07:06:25.978170 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3115eb4c-6af8-41df-9d7e-4db587e37ac5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.287335 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.296088 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9bcc5599-c5twm" event={"ID":"3115eb4c-6af8-41df-9d7e-4db587e37ac5","Type":"ContainerDied","Data":"ef4f3364dc9f05a145615387e50a528356278511ff4066da67e7d2bfe3e3f146"} Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.296133 4978 scope.go:117] "RemoveContainer" containerID="bc7353a13e461bf5e8472c1a0de787f88ed960bbac64db951a039c2f381f9ce5" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.322353 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9bcc5599-c5twm"] Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.324106 4978 scope.go:117] "RemoveContainer" containerID="f90d153c5268ec061775ca86fed36beeeb119b85319aded398fd1612371d491c" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.329660 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f9bcc5599-c5twm"] Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.731250 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56f1-account-create-update-5zpqd" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.750737 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kvbg5" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.846434 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.892461 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-operator-scripts\") pod \"c953f23d-4f0e-4be8-9b37-b26b7de7cad4\" (UID: \"c953f23d-4f0e-4be8-9b37-b26b7de7cad4\") " Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.892558 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhhkb\" (UniqueName: \"kubernetes.io/projected/d31c2c4a-d61c-496d-aef6-31895fee244f-kube-api-access-hhhkb\") pod \"d31c2c4a-d61c-496d-aef6-31895fee244f\" (UID: \"d31c2c4a-d61c-496d-aef6-31895fee244f\") " Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.892589 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d31c2c4a-d61c-496d-aef6-31895fee244f-operator-scripts\") pod \"d31c2c4a-d61c-496d-aef6-31895fee244f\" (UID: \"d31c2c4a-d61c-496d-aef6-31895fee244f\") " Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.892666 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-682d5\" (UniqueName: \"kubernetes.io/projected/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-kube-api-access-682d5\") pod \"c953f23d-4f0e-4be8-9b37-b26b7de7cad4\" (UID: \"c953f23d-4f0e-4be8-9b37-b26b7de7cad4\") " Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.893300 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c953f23d-4f0e-4be8-9b37-b26b7de7cad4" (UID: "c953f23d-4f0e-4be8-9b37-b26b7de7cad4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.894048 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d31c2c4a-d61c-496d-aef6-31895fee244f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d31c2c4a-d61c-496d-aef6-31895fee244f" (UID: "d31c2c4a-d61c-496d-aef6-31895fee244f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.909646 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d31c2c4a-d61c-496d-aef6-31895fee244f-kube-api-access-hhhkb" (OuterVolumeSpecName: "kube-api-access-hhhkb") pod "d31c2c4a-d61c-496d-aef6-31895fee244f" (UID: "d31c2c4a-d61c-496d-aef6-31895fee244f"). InnerVolumeSpecName "kube-api-access-hhhkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.911768 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-kube-api-access-682d5" (OuterVolumeSpecName: "kube-api-access-682d5") pod "c953f23d-4f0e-4be8-9b37-b26b7de7cad4" (UID: "c953f23d-4f0e-4be8-9b37-b26b7de7cad4"). InnerVolumeSpecName "kube-api-access-682d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.994496 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.994526 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhhkb\" (UniqueName: \"kubernetes.io/projected/d31c2c4a-d61c-496d-aef6-31895fee244f-kube-api-access-hhhkb\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.994541 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d31c2c4a-d61c-496d-aef6-31895fee244f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:26 crc kubenswrapper[4978]: I0225 07:06:26.994551 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-682d5\" (UniqueName: \"kubernetes.io/projected/c953f23d-4f0e-4be8-9b37-b26b7de7cad4-kube-api-access-682d5\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.298022 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-56f1-account-create-update-5zpqd" event={"ID":"d31c2c4a-d61c-496d-aef6-31895fee244f","Type":"ContainerDied","Data":"045b7b7fb439cae6ef32a798c585cfbd2c1f7b35d1c7023ed36b45679411cac1"} Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.298064 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="045b7b7fb439cae6ef32a798c585cfbd2c1f7b35d1c7023ed36b45679411cac1" Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.298127 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56f1-account-create-update-5zpqd" Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.300676 4978 generic.go:334] "Generic (PLEG): container finished" podID="43bd8325-9408-4a32-9285-d522d497205e" containerID="765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8" exitCode=0 Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.300743 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43bd8325-9408-4a32-9285-d522d497205e","Type":"ContainerDied","Data":"765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8"} Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.304828 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kvbg5" event={"ID":"c953f23d-4f0e-4be8-9b37-b26b7de7cad4","Type":"ContainerDied","Data":"dad917c48cf7501b9ea93e1d5b13c6c6ba7959712f85833f8be279d64ae5e6ca"} Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.305026 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dad917c48cf7501b9ea93e1d5b13c6c6ba7959712f85833f8be279d64ae5e6ca" Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.305131 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kvbg5" Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.312306 4978 generic.go:334] "Generic (PLEG): container finished" podID="09112202-a3ef-4742-bdbe-b46c1c9a55ba" containerID="62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732" exitCode=0 Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.312430 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"09112202-a3ef-4742-bdbe-b46c1c9a55ba","Type":"ContainerDied","Data":"62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732"} Feb 25 07:06:27 crc kubenswrapper[4978]: I0225 07:06:27.337742 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3115eb4c-6af8-41df-9d7e-4db587e37ac5" path="/var/lib/kubelet/pods/3115eb4c-6af8-41df-9d7e-4db587e37ac5/volumes" Feb 25 07:06:28 crc kubenswrapper[4978]: I0225 07:06:28.328693 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43bd8325-9408-4a32-9285-d522d497205e","Type":"ContainerStarted","Data":"2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5"} Feb 25 07:06:28 crc kubenswrapper[4978]: I0225 07:06:28.329128 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:06:28 crc kubenswrapper[4978]: I0225 07:06:28.331818 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"09112202-a3ef-4742-bdbe-b46c1c9a55ba","Type":"ContainerStarted","Data":"0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33"} Feb 25 07:06:28 crc kubenswrapper[4978]: I0225 07:06:28.331964 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 25 07:06:28 crc kubenswrapper[4978]: I0225 07:06:28.361203 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.937614169 podStartE2EDuration="1m1.361177284s" podCreationTimestamp="2026-02-25 07:05:27 +0000 UTC" firstStartedPulling="2026-02-25 07:05:29.979617778 +0000 UTC m=+1223.418874237" lastFinishedPulling="2026-02-25 07:05:53.403180853 +0000 UTC m=+1246.842437352" observedRunningTime="2026-02-25 07:06:28.360999198 +0000 UTC m=+1281.800255667" watchObservedRunningTime="2026-02-25 07:06:28.361177284 +0000 UTC m=+1281.800433743" Feb 25 07:06:28 crc kubenswrapper[4978]: I0225 07:06:28.392130 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371975.46267 podStartE2EDuration="1m1.392105685s" podCreationTimestamp="2026-02-25 07:05:27 +0000 UTC" firstStartedPulling="2026-02-25 07:05:28.951517542 +0000 UTC m=+1222.390774001" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:28.385870447 +0000 UTC m=+1281.825126936" watchObservedRunningTime="2026-02-25 07:06:28.392105685 +0000 UTC m=+1281.831362144" Feb 25 07:06:31 crc kubenswrapper[4978]: I0225 07:06:31.364814 4978 generic.go:334] "Generic (PLEG): container finished" podID="330d6675-e003-4a9c-862a-88a54e67877e" containerID="ea46e9a5ff7500141828c5e39da95b38f857b39bf42022aa0050cb398f5fd324" exitCode=0 Feb 25 07:06:31 crc kubenswrapper[4978]: I0225 07:06:31.364946 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z5fgm" event={"ID":"330d6675-e003-4a9c-862a-88a54e67877e","Type":"ContainerDied","Data":"ea46e9a5ff7500141828c5e39da95b38f857b39bf42022aa0050cb398f5fd324"} Feb 25 07:06:31 crc kubenswrapper[4978]: I0225 07:06:31.691978 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:31 crc kubenswrapper[4978]: I0225 07:06:31.704652 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift\") pod \"swift-storage-0\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " pod="openstack/swift-storage-0" Feb 25 07:06:31 crc kubenswrapper[4978]: I0225 07:06:31.924010 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 25 07:06:32 crc kubenswrapper[4978]: I0225 07:06:32.964914 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-r45vg" podUID="9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" containerName="ovn-controller" probeResult="failure" output=< Feb 25 07:06:32 crc kubenswrapper[4978]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 25 07:06:32 crc kubenswrapper[4978]: > Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.010295 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.019259 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.253662 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-r45vg-config-2z6jv"] Feb 25 07:06:33 crc kubenswrapper[4978]: E0225 07:06:33.254265 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3115eb4c-6af8-41df-9d7e-4db587e37ac5" containerName="dnsmasq-dns" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.254286 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3115eb4c-6af8-41df-9d7e-4db587e37ac5" containerName="dnsmasq-dns" Feb 25 07:06:33 crc kubenswrapper[4978]: E0225 07:06:33.254307 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d31c2c4a-d61c-496d-aef6-31895fee244f" containerName="mariadb-account-create-update" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.254314 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d31c2c4a-d61c-496d-aef6-31895fee244f" containerName="mariadb-account-create-update" Feb 25 07:06:33 crc kubenswrapper[4978]: E0225 07:06:33.254330 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3115eb4c-6af8-41df-9d7e-4db587e37ac5" containerName="init" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.254336 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3115eb4c-6af8-41df-9d7e-4db587e37ac5" containerName="init" Feb 25 07:06:33 crc kubenswrapper[4978]: E0225 07:06:33.254351 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c953f23d-4f0e-4be8-9b37-b26b7de7cad4" containerName="mariadb-database-create" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.254358 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c953f23d-4f0e-4be8-9b37-b26b7de7cad4" containerName="mariadb-database-create" Feb 25 07:06:33 crc kubenswrapper[4978]: E0225 07:06:33.254395 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce72ec5-eb9b-46f3-ae30-b8380615465c" containerName="mariadb-account-create-update" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.254403 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce72ec5-eb9b-46f3-ae30-b8380615465c" containerName="mariadb-account-create-update" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.254577 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c953f23d-4f0e-4be8-9b37-b26b7de7cad4" containerName="mariadb-database-create" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.254597 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d31c2c4a-d61c-496d-aef6-31895fee244f" containerName="mariadb-account-create-update" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.254615 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3115eb4c-6af8-41df-9d7e-4db587e37ac5" containerName="dnsmasq-dns" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.254635 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ce72ec5-eb9b-46f3-ae30-b8380615465c" containerName="mariadb-account-create-update" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.256147 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.258811 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.265945 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-r45vg-config-2z6jv"] Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.325000 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-log-ovn\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.325098 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lfwg\" (UniqueName: \"kubernetes.io/projected/df970d60-1d8e-497c-b86e-f1787a3f3bd8-kube-api-access-4lfwg\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.325181 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-additional-scripts\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.325269 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.325468 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-scripts\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.325561 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run-ovn\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.427630 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.427761 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-scripts\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.428029 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.428713 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run-ovn\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.428820 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run-ovn\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.428892 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-log-ovn\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.428961 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-log-ovn\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.429017 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lfwg\" (UniqueName: \"kubernetes.io/projected/df970d60-1d8e-497c-b86e-f1787a3f3bd8-kube-api-access-4lfwg\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.429962 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-additional-scripts\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.430230 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-scripts\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.430659 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-additional-scripts\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.447291 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lfwg\" (UniqueName: \"kubernetes.io/projected/df970d60-1d8e-497c-b86e-f1787a3f3bd8-kube-api-access-4lfwg\") pod \"ovn-controller-r45vg-config-2z6jv\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:33 crc kubenswrapper[4978]: I0225 07:06:33.586784 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:37 crc kubenswrapper[4978]: I0225 07:06:37.967592 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-r45vg" podUID="9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" containerName="ovn-controller" probeResult="failure" output=< Feb 25 07:06:37 crc kubenswrapper[4978]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 25 07:06:37 crc kubenswrapper[4978]: > Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.411698 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.794983 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-99k22"] Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.796158 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-99k22" Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.815197 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-99k22"] Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.835354 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51a73e3f-747f-4467-95c5-aebf41f4cc9b-operator-scripts\") pod \"cinder-db-create-99k22\" (UID: \"51a73e3f-747f-4467-95c5-aebf41f4cc9b\") " pod="openstack/cinder-db-create-99k22" Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.835432 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfldb\" (UniqueName: \"kubernetes.io/projected/51a73e3f-747f-4467-95c5-aebf41f4cc9b-kube-api-access-qfldb\") pod \"cinder-db-create-99k22\" (UID: \"51a73e3f-747f-4467-95c5-aebf41f4cc9b\") " pod="openstack/cinder-db-create-99k22" Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.936868 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51a73e3f-747f-4467-95c5-aebf41f4cc9b-operator-scripts\") pod \"cinder-db-create-99k22\" (UID: \"51a73e3f-747f-4467-95c5-aebf41f4cc9b\") " pod="openstack/cinder-db-create-99k22" Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.936915 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfldb\" (UniqueName: \"kubernetes.io/projected/51a73e3f-747f-4467-95c5-aebf41f4cc9b-kube-api-access-qfldb\") pod \"cinder-db-create-99k22\" (UID: \"51a73e3f-747f-4467-95c5-aebf41f4cc9b\") " pod="openstack/cinder-db-create-99k22" Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.937681 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51a73e3f-747f-4467-95c5-aebf41f4cc9b-operator-scripts\") pod \"cinder-db-create-99k22\" (UID: \"51a73e3f-747f-4467-95c5-aebf41f4cc9b\") " pod="openstack/cinder-db-create-99k22" Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.945133 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-9eb7-account-create-update-zszr6"] Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.946031 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9eb7-account-create-update-zszr6" Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.948177 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.955403 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9eb7-account-create-update-zszr6"] Feb 25 07:06:38 crc kubenswrapper[4978]: I0225 07:06:38.966098 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfldb\" (UniqueName: \"kubernetes.io/projected/51a73e3f-747f-4467-95c5-aebf41f4cc9b-kube-api-access-qfldb\") pod \"cinder-db-create-99k22\" (UID: \"51a73e3f-747f-4467-95c5-aebf41f4cc9b\") " pod="openstack/cinder-db-create-99k22" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.028564 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8a8c-account-create-update-sffwp"] Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.029694 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-sffwp" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.034700 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.038261 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bndtj\" (UniqueName: \"kubernetes.io/projected/333c5a0b-874e-45cc-a44e-4b836b4bd850-kube-api-access-bndtj\") pod \"cinder-9eb7-account-create-update-zszr6\" (UID: \"333c5a0b-874e-45cc-a44e-4b836b4bd850\") " pod="openstack/cinder-9eb7-account-create-update-zszr6" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.038325 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333c5a0b-874e-45cc-a44e-4b836b4bd850-operator-scripts\") pod \"cinder-9eb7-account-create-update-zszr6\" (UID: \"333c5a0b-874e-45cc-a44e-4b836b4bd850\") " pod="openstack/cinder-9eb7-account-create-update-zszr6" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.043513 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-w7ztt"] Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.044530 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-w7ztt" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.058782 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8a8c-account-create-update-sffwp"] Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.076420 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-w7ztt"] Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.126417 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-c75lr"] Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.127391 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-c75lr" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.128673 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-99k22" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.140196 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljhbc\" (UniqueName: \"kubernetes.io/projected/f3143797-ced6-46e4-a06a-b3790384dc8c-kube-api-access-ljhbc\") pod \"barbican-8a8c-account-create-update-sffwp\" (UID: \"f3143797-ced6-46e4-a06a-b3790384dc8c\") " pod="openstack/barbican-8a8c-account-create-update-sffwp" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.140252 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfd8z\" (UniqueName: \"kubernetes.io/projected/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-kube-api-access-xfd8z\") pod \"barbican-db-create-w7ztt\" (UID: \"2fcf1162-2ff4-4faa-89f0-272fd5b2669b\") " pod="openstack/barbican-db-create-w7ztt" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.140460 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bndtj\" (UniqueName: \"kubernetes.io/projected/333c5a0b-874e-45cc-a44e-4b836b4bd850-kube-api-access-bndtj\") pod \"cinder-9eb7-account-create-update-zszr6\" (UID: \"333c5a0b-874e-45cc-a44e-4b836b4bd850\") " pod="openstack/cinder-9eb7-account-create-update-zszr6" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.140557 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-operator-scripts\") pod \"barbican-db-create-w7ztt\" (UID: \"2fcf1162-2ff4-4faa-89f0-272fd5b2669b\") " pod="openstack/barbican-db-create-w7ztt" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.140598 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333c5a0b-874e-45cc-a44e-4b836b4bd850-operator-scripts\") pod \"cinder-9eb7-account-create-update-zszr6\" (UID: \"333c5a0b-874e-45cc-a44e-4b836b4bd850\") " pod="openstack/cinder-9eb7-account-create-update-zszr6" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.140623 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3143797-ced6-46e4-a06a-b3790384dc8c-operator-scripts\") pod \"barbican-8a8c-account-create-update-sffwp\" (UID: \"f3143797-ced6-46e4-a06a-b3790384dc8c\") " pod="openstack/barbican-8a8c-account-create-update-sffwp" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.141010 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-c75lr"] Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.141806 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333c5a0b-874e-45cc-a44e-4b836b4bd850-operator-scripts\") pod \"cinder-9eb7-account-create-update-zszr6\" (UID: \"333c5a0b-874e-45cc-a44e-4b836b4bd850\") " pod="openstack/cinder-9eb7-account-create-update-zszr6" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.204345 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bndtj\" (UniqueName: \"kubernetes.io/projected/333c5a0b-874e-45cc-a44e-4b836b4bd850-kube-api-access-bndtj\") pod \"cinder-9eb7-account-create-update-zszr6\" (UID: \"333c5a0b-874e-45cc-a44e-4b836b4bd850\") " pod="openstack/cinder-9eb7-account-create-update-zszr6" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.241916 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-operator-scripts\") pod \"barbican-db-create-w7ztt\" (UID: \"2fcf1162-2ff4-4faa-89f0-272fd5b2669b\") " pod="openstack/barbican-db-create-w7ztt" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.241967 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3143797-ced6-46e4-a06a-b3790384dc8c-operator-scripts\") pod \"barbican-8a8c-account-create-update-sffwp\" (UID: \"f3143797-ced6-46e4-a06a-b3790384dc8c\") " pod="openstack/barbican-8a8c-account-create-update-sffwp" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.241994 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvbxg\" (UniqueName: \"kubernetes.io/projected/1878ccac-42b1-4199-9008-eca6dcad41f6-kube-api-access-vvbxg\") pod \"neutron-db-create-c75lr\" (UID: \"1878ccac-42b1-4199-9008-eca6dcad41f6\") " pod="openstack/neutron-db-create-c75lr" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.242015 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1878ccac-42b1-4199-9008-eca6dcad41f6-operator-scripts\") pod \"neutron-db-create-c75lr\" (UID: \"1878ccac-42b1-4199-9008-eca6dcad41f6\") " pod="openstack/neutron-db-create-c75lr" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.242068 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljhbc\" (UniqueName: \"kubernetes.io/projected/f3143797-ced6-46e4-a06a-b3790384dc8c-kube-api-access-ljhbc\") pod \"barbican-8a8c-account-create-update-sffwp\" (UID: \"f3143797-ced6-46e4-a06a-b3790384dc8c\") " pod="openstack/barbican-8a8c-account-create-update-sffwp" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.242151 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfd8z\" (UniqueName: \"kubernetes.io/projected/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-kube-api-access-xfd8z\") pod \"barbican-db-create-w7ztt\" (UID: \"2fcf1162-2ff4-4faa-89f0-272fd5b2669b\") " pod="openstack/barbican-db-create-w7ztt" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.242761 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3143797-ced6-46e4-a06a-b3790384dc8c-operator-scripts\") pod \"barbican-8a8c-account-create-update-sffwp\" (UID: \"f3143797-ced6-46e4-a06a-b3790384dc8c\") " pod="openstack/barbican-8a8c-account-create-update-sffwp" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.243054 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-operator-scripts\") pod \"barbican-db-create-w7ztt\" (UID: \"2fcf1162-2ff4-4faa-89f0-272fd5b2669b\") " pod="openstack/barbican-db-create-w7ztt" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.246546 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-phpsb"] Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.247426 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.250429 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.250464 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.250507 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.250854 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dvmg8" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.261337 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9eb7-account-create-update-zszr6" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.261426 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfd8z\" (UniqueName: \"kubernetes.io/projected/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-kube-api-access-xfd8z\") pod \"barbican-db-create-w7ztt\" (UID: \"2fcf1162-2ff4-4faa-89f0-272fd5b2669b\") " pod="openstack/barbican-db-create-w7ztt" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.265509 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-phpsb"] Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.267197 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljhbc\" (UniqueName: \"kubernetes.io/projected/f3143797-ced6-46e4-a06a-b3790384dc8c-kube-api-access-ljhbc\") pod \"barbican-8a8c-account-create-update-sffwp\" (UID: \"f3143797-ced6-46e4-a06a-b3790384dc8c\") " pod="openstack/barbican-8a8c-account-create-update-sffwp" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.325750 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6ec4-account-create-update-8mwd5"] Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.326818 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ec4-account-create-update-8mwd5" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.328343 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.337688 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6ec4-account-create-update-8mwd5"] Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.343947 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqfpk\" (UniqueName: \"kubernetes.io/projected/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-kube-api-access-gqfpk\") pod \"keystone-db-sync-phpsb\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.343987 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-config-data\") pod \"keystone-db-sync-phpsb\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.344019 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-combined-ca-bundle\") pod \"keystone-db-sync-phpsb\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.344053 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvbxg\" (UniqueName: \"kubernetes.io/projected/1878ccac-42b1-4199-9008-eca6dcad41f6-kube-api-access-vvbxg\") pod \"neutron-db-create-c75lr\" (UID: \"1878ccac-42b1-4199-9008-eca6dcad41f6\") " pod="openstack/neutron-db-create-c75lr" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.344072 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1878ccac-42b1-4199-9008-eca6dcad41f6-operator-scripts\") pod \"neutron-db-create-c75lr\" (UID: \"1878ccac-42b1-4199-9008-eca6dcad41f6\") " pod="openstack/neutron-db-create-c75lr" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.344727 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1878ccac-42b1-4199-9008-eca6dcad41f6-operator-scripts\") pod \"neutron-db-create-c75lr\" (UID: \"1878ccac-42b1-4199-9008-eca6dcad41f6\") " pod="openstack/neutron-db-create-c75lr" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.352810 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-sffwp" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.362713 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-w7ztt" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.364509 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvbxg\" (UniqueName: \"kubernetes.io/projected/1878ccac-42b1-4199-9008-eca6dcad41f6-kube-api-access-vvbxg\") pod \"neutron-db-create-c75lr\" (UID: \"1878ccac-42b1-4199-9008-eca6dcad41f6\") " pod="openstack/neutron-db-create-c75lr" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.381149 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.445728 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccea553d-1b65-436c-aa35-3b0217c92689-operator-scripts\") pod \"neutron-6ec4-account-create-update-8mwd5\" (UID: \"ccea553d-1b65-436c-aa35-3b0217c92689\") " pod="openstack/neutron-6ec4-account-create-update-8mwd5" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.445782 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6rj8\" (UniqueName: \"kubernetes.io/projected/ccea553d-1b65-436c-aa35-3b0217c92689-kube-api-access-c6rj8\") pod \"neutron-6ec4-account-create-update-8mwd5\" (UID: \"ccea553d-1b65-436c-aa35-3b0217c92689\") " pod="openstack/neutron-6ec4-account-create-update-8mwd5" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.445817 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqfpk\" (UniqueName: \"kubernetes.io/projected/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-kube-api-access-gqfpk\") pod \"keystone-db-sync-phpsb\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.445850 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-config-data\") pod \"keystone-db-sync-phpsb\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.445998 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-combined-ca-bundle\") pod \"keystone-db-sync-phpsb\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.450407 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-combined-ca-bundle\") pod \"keystone-db-sync-phpsb\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.450623 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-config-data\") pod \"keystone-db-sync-phpsb\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.456290 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-c75lr" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.467232 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqfpk\" (UniqueName: \"kubernetes.io/projected/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-kube-api-access-gqfpk\") pod \"keystone-db-sync-phpsb\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.547581 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccea553d-1b65-436c-aa35-3b0217c92689-operator-scripts\") pod \"neutron-6ec4-account-create-update-8mwd5\" (UID: \"ccea553d-1b65-436c-aa35-3b0217c92689\") " pod="openstack/neutron-6ec4-account-create-update-8mwd5" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.547673 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6rj8\" (UniqueName: \"kubernetes.io/projected/ccea553d-1b65-436c-aa35-3b0217c92689-kube-api-access-c6rj8\") pod \"neutron-6ec4-account-create-update-8mwd5\" (UID: \"ccea553d-1b65-436c-aa35-3b0217c92689\") " pod="openstack/neutron-6ec4-account-create-update-8mwd5" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.554104 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccea553d-1b65-436c-aa35-3b0217c92689-operator-scripts\") pod \"neutron-6ec4-account-create-update-8mwd5\" (UID: \"ccea553d-1b65-436c-aa35-3b0217c92689\") " pod="openstack/neutron-6ec4-account-create-update-8mwd5" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.562068 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.572318 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6rj8\" (UniqueName: \"kubernetes.io/projected/ccea553d-1b65-436c-aa35-3b0217c92689-kube-api-access-c6rj8\") pod \"neutron-6ec4-account-create-update-8mwd5\" (UID: \"ccea553d-1b65-436c-aa35-3b0217c92689\") " pod="openstack/neutron-6ec4-account-create-update-8mwd5" Feb 25 07:06:39 crc kubenswrapper[4978]: I0225 07:06:39.643490 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ec4-account-create-update-8mwd5" Feb 25 07:06:41 crc kubenswrapper[4978]: E0225 07:06:41.581199 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:69927bf0036b7c213c1fa7c3879c3d6a7690de68bbbe38f078f45f21708e3416" Feb 25 07:06:41 crc kubenswrapper[4978]: E0225 07:06:41.582043 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:69927bf0036b7c213c1fa7c3879c3d6a7690de68bbbe38f078f45f21708e3416,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6mqrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-7ts6t_openstack(4bd97fe5-291b-4130-98c7-4ffbc0cba189): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 07:06:41 crc kubenswrapper[4978]: E0225 07:06:41.583449 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-7ts6t" podUID="4bd97fe5-291b-4130-98c7-4ffbc0cba189" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.750872 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.785966 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-combined-ca-bundle\") pod \"330d6675-e003-4a9c-862a-88a54e67877e\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.786055 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-ring-data-devices\") pod \"330d6675-e003-4a9c-862a-88a54e67877e\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.786096 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/330d6675-e003-4a9c-862a-88a54e67877e-etc-swift\") pod \"330d6675-e003-4a9c-862a-88a54e67877e\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.786127 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-dispersionconf\") pod \"330d6675-e003-4a9c-862a-88a54e67877e\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.786169 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-scripts\") pod \"330d6675-e003-4a9c-862a-88a54e67877e\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.786283 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-swiftconf\") pod \"330d6675-e003-4a9c-862a-88a54e67877e\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.786350 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7vdm\" (UniqueName: \"kubernetes.io/projected/330d6675-e003-4a9c-862a-88a54e67877e-kube-api-access-m7vdm\") pod \"330d6675-e003-4a9c-862a-88a54e67877e\" (UID: \"330d6675-e003-4a9c-862a-88a54e67877e\") " Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.787710 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/330d6675-e003-4a9c-862a-88a54e67877e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "330d6675-e003-4a9c-862a-88a54e67877e" (UID: "330d6675-e003-4a9c-862a-88a54e67877e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.788324 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "330d6675-e003-4a9c-862a-88a54e67877e" (UID: "330d6675-e003-4a9c-862a-88a54e67877e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.794825 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/330d6675-e003-4a9c-862a-88a54e67877e-kube-api-access-m7vdm" (OuterVolumeSpecName: "kube-api-access-m7vdm") pod "330d6675-e003-4a9c-862a-88a54e67877e" (UID: "330d6675-e003-4a9c-862a-88a54e67877e"). InnerVolumeSpecName "kube-api-access-m7vdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.813632 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "330d6675-e003-4a9c-862a-88a54e67877e" (UID: "330d6675-e003-4a9c-862a-88a54e67877e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.825514 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-scripts" (OuterVolumeSpecName: "scripts") pod "330d6675-e003-4a9c-862a-88a54e67877e" (UID: "330d6675-e003-4a9c-862a-88a54e67877e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.832748 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "330d6675-e003-4a9c-862a-88a54e67877e" (UID: "330d6675-e003-4a9c-862a-88a54e67877e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.838617 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "330d6675-e003-4a9c-862a-88a54e67877e" (UID: "330d6675-e003-4a9c-862a-88a54e67877e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.888802 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7vdm\" (UniqueName: \"kubernetes.io/projected/330d6675-e003-4a9c-862a-88a54e67877e-kube-api-access-m7vdm\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.889011 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.889021 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.889031 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/330d6675-e003-4a9c-862a-88a54e67877e-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.889040 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.889049 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/330d6675-e003-4a9c-862a-88a54e67877e-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:41 crc kubenswrapper[4978]: I0225 07:06:41.889060 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/330d6675-e003-4a9c-862a-88a54e67877e-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.453319 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-z5fgm" event={"ID":"330d6675-e003-4a9c-862a-88a54e67877e","Type":"ContainerDied","Data":"15b7e45b2207a087212683967f513ec289c7f8f0e1faa240f8ed6692d7388a95"} Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.453359 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15b7e45b2207a087212683967f513ec289c7f8f0e1faa240f8ed6692d7388a95" Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.453391 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-z5fgm" Feb 25 07:06:42 crc kubenswrapper[4978]: E0225 07:06:42.454355 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:69927bf0036b7c213c1fa7c3879c3d6a7690de68bbbe38f078f45f21708e3416\\\"\"" pod="openstack/glance-db-sync-7ts6t" podUID="4bd97fe5-291b-4130-98c7-4ffbc0cba189" Feb 25 07:06:42 crc kubenswrapper[4978]: W0225 07:06:42.474650 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a73e3f_747f_4467_95c5_aebf41f4cc9b.slice/crio-cb70b171d73bb3fcd7e16d41ad40c8c9142d55146ec275ca3d1924675c7bc170 WatchSource:0}: Error finding container cb70b171d73bb3fcd7e16d41ad40c8c9142d55146ec275ca3d1924675c7bc170: Status 404 returned error can't find the container with id cb70b171d73bb3fcd7e16d41ad40c8c9142d55146ec275ca3d1924675c7bc170 Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.490245 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-99k22"] Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.660981 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-w7ztt"] Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.680261 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-phpsb"] Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.692948 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9eb7-account-create-update-zszr6"] Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.746790 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-r45vg-config-2z6jv"] Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.761461 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8a8c-account-create-update-sffwp"] Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.770735 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-c75lr"] Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.779602 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6ec4-account-create-update-8mwd5"] Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.823708 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 25 07:06:42 crc kubenswrapper[4978]: I0225 07:06:42.955780 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-r45vg" podUID="9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" containerName="ovn-controller" probeResult="failure" output=< Feb 25 07:06:42 crc kubenswrapper[4978]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 25 07:06:42 crc kubenswrapper[4978]: > Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.461590 4978 generic.go:334] "Generic (PLEG): container finished" podID="df970d60-1d8e-497c-b86e-f1787a3f3bd8" containerID="3c6dcd5048df9f2163eaef44d3728e6bf73e055ec2ab84cad23993f70dbd7b2e" exitCode=0 Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.461683 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r45vg-config-2z6jv" event={"ID":"df970d60-1d8e-497c-b86e-f1787a3f3bd8","Type":"ContainerDied","Data":"3c6dcd5048df9f2163eaef44d3728e6bf73e055ec2ab84cad23993f70dbd7b2e"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.461715 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r45vg-config-2z6jv" event={"ID":"df970d60-1d8e-497c-b86e-f1787a3f3bd8","Type":"ContainerStarted","Data":"ea2ce4ad5e7091e432a755a9f5fbae94ae391373da2c109b5c7b07b31f9fb370"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.462724 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-phpsb" event={"ID":"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9","Type":"ContainerStarted","Data":"ca598b92e0b6bbac0da5a14b63a9ec19f3344ab6d995daf3e1b7377de1aef97b"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.464333 4978 generic.go:334] "Generic (PLEG): container finished" podID="51a73e3f-747f-4467-95c5-aebf41f4cc9b" containerID="5520aff92f97a896a5a806d71f4fb6585ad2a8f952fe7c239751ef1d30bae74e" exitCode=0 Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.464450 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-99k22" event={"ID":"51a73e3f-747f-4467-95c5-aebf41f4cc9b","Type":"ContainerDied","Data":"5520aff92f97a896a5a806d71f4fb6585ad2a8f952fe7c239751ef1d30bae74e"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.464493 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-99k22" event={"ID":"51a73e3f-747f-4467-95c5-aebf41f4cc9b","Type":"ContainerStarted","Data":"cb70b171d73bb3fcd7e16d41ad40c8c9142d55146ec275ca3d1924675c7bc170"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.466228 4978 generic.go:334] "Generic (PLEG): container finished" podID="ccea553d-1b65-436c-aa35-3b0217c92689" containerID="8e2f22cb510c2be5bede5f9167ccd0a70c7f5db1dec604bd4fea4484d24efb52" exitCode=0 Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.466289 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ec4-account-create-update-8mwd5" event={"ID":"ccea553d-1b65-436c-aa35-3b0217c92689","Type":"ContainerDied","Data":"8e2f22cb510c2be5bede5f9167ccd0a70c7f5db1dec604bd4fea4484d24efb52"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.466326 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ec4-account-create-update-8mwd5" event={"ID":"ccea553d-1b65-436c-aa35-3b0217c92689","Type":"ContainerStarted","Data":"c22cef9af9e589248c155536d7beb4a38ee1d0981cacdab872507c6c2c80ae37"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.467480 4978 generic.go:334] "Generic (PLEG): container finished" podID="333c5a0b-874e-45cc-a44e-4b836b4bd850" containerID="3027c6d98887a040746bbb52842379fe73693d836e727a5d3de95c63fc2b5e30" exitCode=0 Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.467532 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9eb7-account-create-update-zszr6" event={"ID":"333c5a0b-874e-45cc-a44e-4b836b4bd850","Type":"ContainerDied","Data":"3027c6d98887a040746bbb52842379fe73693d836e727a5d3de95c63fc2b5e30"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.467548 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9eb7-account-create-update-zszr6" event={"ID":"333c5a0b-874e-45cc-a44e-4b836b4bd850","Type":"ContainerStarted","Data":"35e65c20446f72892e4584951f13a3ce8bdc4542f2f5bb52fcf00cdcea61db68"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.468706 4978 generic.go:334] "Generic (PLEG): container finished" podID="2fcf1162-2ff4-4faa-89f0-272fd5b2669b" containerID="5b1a33fcd51e2cec3e100c3d54c2ea14e7ea10f1fc1625a323da5788dd492729" exitCode=0 Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.468740 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-w7ztt" event={"ID":"2fcf1162-2ff4-4faa-89f0-272fd5b2669b","Type":"ContainerDied","Data":"5b1a33fcd51e2cec3e100c3d54c2ea14e7ea10f1fc1625a323da5788dd492729"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.468794 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-w7ztt" event={"ID":"2fcf1162-2ff4-4faa-89f0-272fd5b2669b","Type":"ContainerStarted","Data":"54509fe5471c8f023c85c21002d66f35d7fbf35af8aeff8f4add7fc459c596b6"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.470235 4978 generic.go:334] "Generic (PLEG): container finished" podID="1878ccac-42b1-4199-9008-eca6dcad41f6" containerID="2b9d4c187e335c3723ec5d19e25fdb6cedda0a9852c880709cea81dfaf1e5c3b" exitCode=0 Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.470275 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-c75lr" event={"ID":"1878ccac-42b1-4199-9008-eca6dcad41f6","Type":"ContainerDied","Data":"2b9d4c187e335c3723ec5d19e25fdb6cedda0a9852c880709cea81dfaf1e5c3b"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.470289 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-c75lr" event={"ID":"1878ccac-42b1-4199-9008-eca6dcad41f6","Type":"ContainerStarted","Data":"e09936aa9c3bf638cc1bec220ebf305d44d0f31e506773fabc2ea7b27212023d"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.471750 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"d1f7e3a38ee0eaf8e5860da6b520860688f76a4907f43b92be68142e113845f5"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.473211 4978 generic.go:334] "Generic (PLEG): container finished" podID="f3143797-ced6-46e4-a06a-b3790384dc8c" containerID="51facbbef619d03278afe02c3c8fbe41a2c4e98c1f3161602d2cb500952a63a4" exitCode=0 Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.473259 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a8c-account-create-update-sffwp" event={"ID":"f3143797-ced6-46e4-a06a-b3790384dc8c","Type":"ContainerDied","Data":"51facbbef619d03278afe02c3c8fbe41a2c4e98c1f3161602d2cb500952a63a4"} Feb 25 07:06:43 crc kubenswrapper[4978]: I0225 07:06:43.473285 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a8c-account-create-update-sffwp" event={"ID":"f3143797-ced6-46e4-a06a-b3790384dc8c","Type":"ContainerStarted","Data":"9fb59ea19fe2dfdc9c49da57d2a9850828304b9eab95086e6ccf497b6866ff87"} Feb 25 07:06:44 crc kubenswrapper[4978]: I0225 07:06:44.504960 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"bb9960184024cbe585b528d58a3dd1a8b9fdeda2c6566bd7d76615473e70311e"} Feb 25 07:06:44 crc kubenswrapper[4978]: I0225 07:06:44.505578 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"fe36f33cfa027ca4f84b465a94aa8c1bbd4436f1982627dab087ee9bdbe66214"} Feb 25 07:06:45 crc kubenswrapper[4978]: I0225 07:06:45.519647 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"25a3891e332bc2d905bcc5238aea824519156ba640e57709ba7a03f1986ab36b"} Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.446510 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9eb7-account-create-update-zszr6" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.460500 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ec4-account-create-update-8mwd5" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.465816 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-sffwp" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.475966 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.514581 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-c75lr" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.547191 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a8c-account-create-update-sffwp" event={"ID":"f3143797-ced6-46e4-a06a-b3790384dc8c","Type":"ContainerDied","Data":"9fb59ea19fe2dfdc9c49da57d2a9850828304b9eab95086e6ccf497b6866ff87"} Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.547250 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fb59ea19fe2dfdc9c49da57d2a9850828304b9eab95086e6ccf497b6866ff87" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.547207 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-sffwp" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.552568 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-99k22" event={"ID":"51a73e3f-747f-4467-95c5-aebf41f4cc9b","Type":"ContainerDied","Data":"cb70b171d73bb3fcd7e16d41ad40c8c9142d55146ec275ca3d1924675c7bc170"} Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.552607 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb70b171d73bb3fcd7e16d41ad40c8c9142d55146ec275ca3d1924675c7bc170" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.554114 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ec4-account-create-update-8mwd5" event={"ID":"ccea553d-1b65-436c-aa35-3b0217c92689","Type":"ContainerDied","Data":"c22cef9af9e589248c155536d7beb4a38ee1d0981cacdab872507c6c2c80ae37"} Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.554131 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c22cef9af9e589248c155536d7beb4a38ee1d0981cacdab872507c6c2c80ae37" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.554181 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ec4-account-create-update-8mwd5" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.568205 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9eb7-account-create-update-zszr6" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.568228 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9eb7-account-create-update-zszr6" event={"ID":"333c5a0b-874e-45cc-a44e-4b836b4bd850","Type":"ContainerDied","Data":"35e65c20446f72892e4584951f13a3ce8bdc4542f2f5bb52fcf00cdcea61db68"} Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.568276 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35e65c20446f72892e4584951f13a3ce8bdc4542f2f5bb52fcf00cdcea61db68" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.569433 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-w7ztt" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.569982 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-w7ztt" event={"ID":"2fcf1162-2ff4-4faa-89f0-272fd5b2669b","Type":"ContainerDied","Data":"54509fe5471c8f023c85c21002d66f35d7fbf35af8aeff8f4add7fc459c596b6"} Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.570017 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54509fe5471c8f023c85c21002d66f35d7fbf35af8aeff8f4add7fc459c596b6" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.571539 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r45vg-config-2z6jv" event={"ID":"df970d60-1d8e-497c-b86e-f1787a3f3bd8","Type":"ContainerDied","Data":"ea2ce4ad5e7091e432a755a9f5fbae94ae391373da2c109b5c7b07b31f9fb370"} Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.571567 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea2ce4ad5e7091e432a755a9f5fbae94ae391373da2c109b5c7b07b31f9fb370" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.571611 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r45vg-config-2z6jv" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.577404 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-c75lr" event={"ID":"1878ccac-42b1-4199-9008-eca6dcad41f6","Type":"ContainerDied","Data":"e09936aa9c3bf638cc1bec220ebf305d44d0f31e506773fabc2ea7b27212023d"} Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.577447 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e09936aa9c3bf638cc1bec220ebf305d44d0f31e506773fabc2ea7b27212023d" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.577516 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-c75lr" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.587441 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-99k22" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.612719 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccea553d-1b65-436c-aa35-3b0217c92689-operator-scripts\") pod \"ccea553d-1b65-436c-aa35-3b0217c92689\" (UID: \"ccea553d-1b65-436c-aa35-3b0217c92689\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.612836 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lfwg\" (UniqueName: \"kubernetes.io/projected/df970d60-1d8e-497c-b86e-f1787a3f3bd8-kube-api-access-4lfwg\") pod \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.612870 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333c5a0b-874e-45cc-a44e-4b836b4bd850-operator-scripts\") pod \"333c5a0b-874e-45cc-a44e-4b836b4bd850\" (UID: \"333c5a0b-874e-45cc-a44e-4b836b4bd850\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.612890 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1878ccac-42b1-4199-9008-eca6dcad41f6-operator-scripts\") pod \"1878ccac-42b1-4199-9008-eca6dcad41f6\" (UID: \"1878ccac-42b1-4199-9008-eca6dcad41f6\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.612911 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljhbc\" (UniqueName: \"kubernetes.io/projected/f3143797-ced6-46e4-a06a-b3790384dc8c-kube-api-access-ljhbc\") pod \"f3143797-ced6-46e4-a06a-b3790384dc8c\" (UID: \"f3143797-ced6-46e4-a06a-b3790384dc8c\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.612960 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run\") pod \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.612992 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-log-ovn\") pod \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613014 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvbxg\" (UniqueName: \"kubernetes.io/projected/1878ccac-42b1-4199-9008-eca6dcad41f6-kube-api-access-vvbxg\") pod \"1878ccac-42b1-4199-9008-eca6dcad41f6\" (UID: \"1878ccac-42b1-4199-9008-eca6dcad41f6\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613076 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-scripts\") pod \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613123 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bndtj\" (UniqueName: \"kubernetes.io/projected/333c5a0b-874e-45cc-a44e-4b836b4bd850-kube-api-access-bndtj\") pod \"333c5a0b-874e-45cc-a44e-4b836b4bd850\" (UID: \"333c5a0b-874e-45cc-a44e-4b836b4bd850\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613149 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6rj8\" (UniqueName: \"kubernetes.io/projected/ccea553d-1b65-436c-aa35-3b0217c92689-kube-api-access-c6rj8\") pod \"ccea553d-1b65-436c-aa35-3b0217c92689\" (UID: \"ccea553d-1b65-436c-aa35-3b0217c92689\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613186 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-additional-scripts\") pod \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613225 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3143797-ced6-46e4-a06a-b3790384dc8c-operator-scripts\") pod \"f3143797-ced6-46e4-a06a-b3790384dc8c\" (UID: \"f3143797-ced6-46e4-a06a-b3790384dc8c\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613253 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run-ovn\") pod \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\" (UID: \"df970d60-1d8e-497c-b86e-f1787a3f3bd8\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613671 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333c5a0b-874e-45cc-a44e-4b836b4bd850-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "333c5a0b-874e-45cc-a44e-4b836b4bd850" (UID: "333c5a0b-874e-45cc-a44e-4b836b4bd850"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613716 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "df970d60-1d8e-497c-b86e-f1787a3f3bd8" (UID: "df970d60-1d8e-497c-b86e-f1787a3f3bd8"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613733 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run" (OuterVolumeSpecName: "var-run") pod "df970d60-1d8e-497c-b86e-f1787a3f3bd8" (UID: "df970d60-1d8e-497c-b86e-f1787a3f3bd8"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613827 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/333c5a0b-874e-45cc-a44e-4b836b4bd850-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613846 4978 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.613861 4978 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.614175 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1878ccac-42b1-4199-9008-eca6dcad41f6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1878ccac-42b1-4199-9008-eca6dcad41f6" (UID: "1878ccac-42b1-4199-9008-eca6dcad41f6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.614189 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccea553d-1b65-436c-aa35-3b0217c92689-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ccea553d-1b65-436c-aa35-3b0217c92689" (UID: "ccea553d-1b65-436c-aa35-3b0217c92689"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.614255 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "df970d60-1d8e-497c-b86e-f1787a3f3bd8" (UID: "df970d60-1d8e-497c-b86e-f1787a3f3bd8"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.614304 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3143797-ced6-46e4-a06a-b3790384dc8c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f3143797-ced6-46e4-a06a-b3790384dc8c" (UID: "f3143797-ced6-46e4-a06a-b3790384dc8c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.614337 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "df970d60-1d8e-497c-b86e-f1787a3f3bd8" (UID: "df970d60-1d8e-497c-b86e-f1787a3f3bd8"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.614656 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-scripts" (OuterVolumeSpecName: "scripts") pod "df970d60-1d8e-497c-b86e-f1787a3f3bd8" (UID: "df970d60-1d8e-497c-b86e-f1787a3f3bd8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.617962 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccea553d-1b65-436c-aa35-3b0217c92689-kube-api-access-c6rj8" (OuterVolumeSpecName: "kube-api-access-c6rj8") pod "ccea553d-1b65-436c-aa35-3b0217c92689" (UID: "ccea553d-1b65-436c-aa35-3b0217c92689"). InnerVolumeSpecName "kube-api-access-c6rj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.618076 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3143797-ced6-46e4-a06a-b3790384dc8c-kube-api-access-ljhbc" (OuterVolumeSpecName: "kube-api-access-ljhbc") pod "f3143797-ced6-46e4-a06a-b3790384dc8c" (UID: "f3143797-ced6-46e4-a06a-b3790384dc8c"). InnerVolumeSpecName "kube-api-access-ljhbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.618179 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/333c5a0b-874e-45cc-a44e-4b836b4bd850-kube-api-access-bndtj" (OuterVolumeSpecName: "kube-api-access-bndtj") pod "333c5a0b-874e-45cc-a44e-4b836b4bd850" (UID: "333c5a0b-874e-45cc-a44e-4b836b4bd850"). InnerVolumeSpecName "kube-api-access-bndtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.618470 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df970d60-1d8e-497c-b86e-f1787a3f3bd8-kube-api-access-4lfwg" (OuterVolumeSpecName: "kube-api-access-4lfwg") pod "df970d60-1d8e-497c-b86e-f1787a3f3bd8" (UID: "df970d60-1d8e-497c-b86e-f1787a3f3bd8"). InnerVolumeSpecName "kube-api-access-4lfwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.619997 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1878ccac-42b1-4199-9008-eca6dcad41f6-kube-api-access-vvbxg" (OuterVolumeSpecName: "kube-api-access-vvbxg") pod "1878ccac-42b1-4199-9008-eca6dcad41f6" (UID: "1878ccac-42b1-4199-9008-eca6dcad41f6"). InnerVolumeSpecName "kube-api-access-vvbxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.715313 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfldb\" (UniqueName: \"kubernetes.io/projected/51a73e3f-747f-4467-95c5-aebf41f4cc9b-kube-api-access-qfldb\") pod \"51a73e3f-747f-4467-95c5-aebf41f4cc9b\" (UID: \"51a73e3f-747f-4467-95c5-aebf41f4cc9b\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.715657 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51a73e3f-747f-4467-95c5-aebf41f4cc9b-operator-scripts\") pod \"51a73e3f-747f-4467-95c5-aebf41f4cc9b\" (UID: \"51a73e3f-747f-4467-95c5-aebf41f4cc9b\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.715810 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfd8z\" (UniqueName: \"kubernetes.io/projected/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-kube-api-access-xfd8z\") pod \"2fcf1162-2ff4-4faa-89f0-272fd5b2669b\" (UID: \"2fcf1162-2ff4-4faa-89f0-272fd5b2669b\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.715903 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-operator-scripts\") pod \"2fcf1162-2ff4-4faa-89f0-272fd5b2669b\" (UID: \"2fcf1162-2ff4-4faa-89f0-272fd5b2669b\") " Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716290 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a73e3f-747f-4467-95c5-aebf41f4cc9b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "51a73e3f-747f-4467-95c5-aebf41f4cc9b" (UID: "51a73e3f-747f-4467-95c5-aebf41f4cc9b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716323 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccea553d-1b65-436c-aa35-3b0217c92689-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716444 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lfwg\" (UniqueName: \"kubernetes.io/projected/df970d60-1d8e-497c-b86e-f1787a3f3bd8-kube-api-access-4lfwg\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716475 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1878ccac-42b1-4199-9008-eca6dcad41f6-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716496 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljhbc\" (UniqueName: \"kubernetes.io/projected/f3143797-ced6-46e4-a06a-b3790384dc8c-kube-api-access-ljhbc\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716500 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2fcf1162-2ff4-4faa-89f0-272fd5b2669b" (UID: "2fcf1162-2ff4-4faa-89f0-272fd5b2669b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716516 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvbxg\" (UniqueName: \"kubernetes.io/projected/1878ccac-42b1-4199-9008-eca6dcad41f6-kube-api-access-vvbxg\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716573 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716585 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bndtj\" (UniqueName: \"kubernetes.io/projected/333c5a0b-874e-45cc-a44e-4b836b4bd850-kube-api-access-bndtj\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716596 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6rj8\" (UniqueName: \"kubernetes.io/projected/ccea553d-1b65-436c-aa35-3b0217c92689-kube-api-access-c6rj8\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716606 4978 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/df970d60-1d8e-497c-b86e-f1787a3f3bd8-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716615 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3143797-ced6-46e4-a06a-b3790384dc8c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.716625 4978 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/df970d60-1d8e-497c-b86e-f1787a3f3bd8-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.718579 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a73e3f-747f-4467-95c5-aebf41f4cc9b-kube-api-access-qfldb" (OuterVolumeSpecName: "kube-api-access-qfldb") pod "51a73e3f-747f-4467-95c5-aebf41f4cc9b" (UID: "51a73e3f-747f-4467-95c5-aebf41f4cc9b"). InnerVolumeSpecName "kube-api-access-qfldb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.720246 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-kube-api-access-xfd8z" (OuterVolumeSpecName: "kube-api-access-xfd8z") pod "2fcf1162-2ff4-4faa-89f0-272fd5b2669b" (UID: "2fcf1162-2ff4-4faa-89f0-272fd5b2669b"). InnerVolumeSpecName "kube-api-access-xfd8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.818916 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.819394 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfldb\" (UniqueName: \"kubernetes.io/projected/51a73e3f-747f-4467-95c5-aebf41f4cc9b-kube-api-access-qfldb\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.820020 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51a73e3f-747f-4467-95c5-aebf41f4cc9b-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.820069 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfd8z\" (UniqueName: \"kubernetes.io/projected/2fcf1162-2ff4-4faa-89f0-272fd5b2669b-kube-api-access-xfd8z\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:47 crc kubenswrapper[4978]: I0225 07:06:47.943410 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-r45vg" Feb 25 07:06:48 crc kubenswrapper[4978]: I0225 07:06:48.625354 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-phpsb" event={"ID":"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9","Type":"ContainerStarted","Data":"15a2b4fd13fadf8eed227080adbec60612493687b1b860efa155a9d0a8f50be7"} Feb 25 07:06:48 crc kubenswrapper[4978]: I0225 07:06:48.643696 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-r45vg-config-2z6jv"] Feb 25 07:06:48 crc kubenswrapper[4978]: I0225 07:06:48.644335 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-99k22" Feb 25 07:06:48 crc kubenswrapper[4978]: I0225 07:06:48.648430 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"c196bca824e71c59dc1d9939514fc9f5d009c35f1cb8b40a2d2b99b8d722f60b"} Feb 25 07:06:48 crc kubenswrapper[4978]: I0225 07:06:48.648534 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-w7ztt" Feb 25 07:06:48 crc kubenswrapper[4978]: I0225 07:06:48.656176 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-r45vg-config-2z6jv"] Feb 25 07:06:48 crc kubenswrapper[4978]: I0225 07:06:48.657562 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-phpsb" podStartSLOduration=4.968647622 podStartE2EDuration="9.657550448s" podCreationTimestamp="2026-02-25 07:06:39 +0000 UTC" firstStartedPulling="2026-02-25 07:06:42.689190282 +0000 UTC m=+1296.128446741" lastFinishedPulling="2026-02-25 07:06:47.378093078 +0000 UTC m=+1300.817349567" observedRunningTime="2026-02-25 07:06:48.643669999 +0000 UTC m=+1302.082926448" watchObservedRunningTime="2026-02-25 07:06:48.657550448 +0000 UTC m=+1302.096806907" Feb 25 07:06:49 crc kubenswrapper[4978]: I0225 07:06:49.342698 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df970d60-1d8e-497c-b86e-f1787a3f3bd8" path="/var/lib/kubelet/pods/df970d60-1d8e-497c-b86e-f1787a3f3bd8/volumes" Feb 25 07:06:49 crc kubenswrapper[4978]: I0225 07:06:49.663627 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"c247f762435a5bf5a26de2224e27277501e932169aad0cc00bd5b31c4c134647"} Feb 25 07:06:49 crc kubenswrapper[4978]: I0225 07:06:49.663881 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"36f71fa276b25e88398462e3bbc6d6b87e2bee8c2aade1e2f5f0ffdc61440473"} Feb 25 07:06:49 crc kubenswrapper[4978]: I0225 07:06:49.663891 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"2c1a63624536475bb718a04a658404433f192ece6f6682b1e5ca98142ac9eb79"} Feb 25 07:06:50 crc kubenswrapper[4978]: I0225 07:06:50.682638 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"f892ef9c825b59066a0c023cd3b507cce63da08b00077af4f81ced0830c89f1c"} Feb 25 07:06:51 crc kubenswrapper[4978]: I0225 07:06:51.701042 4978 generic.go:334] "Generic (PLEG): container finished" podID="fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9" containerID="15a2b4fd13fadf8eed227080adbec60612493687b1b860efa155a9d0a8f50be7" exitCode=0 Feb 25 07:06:51 crc kubenswrapper[4978]: I0225 07:06:51.701140 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-phpsb" event={"ID":"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9","Type":"ContainerDied","Data":"15a2b4fd13fadf8eed227080adbec60612493687b1b860efa155a9d0a8f50be7"} Feb 25 07:06:52 crc kubenswrapper[4978]: I0225 07:06:52.738162 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"4dfed42feb8ac7a42412c46dcae347795ab03fb0af477b3ffbb5372f2f1e92c3"} Feb 25 07:06:52 crc kubenswrapper[4978]: I0225 07:06:52.738733 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"c5c66777c2f36490e61ea071a8cde68af4d8aa4113f9db174c104aae89ea3baf"} Feb 25 07:06:52 crc kubenswrapper[4978]: I0225 07:06:52.738747 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"1ec7901b94aa973e76af694ac0b6f07b5c9788edfe7b888e4ae6f0ec6806b20b"} Feb 25 07:06:52 crc kubenswrapper[4978]: I0225 07:06:52.738756 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"6fefdf96f9ea31d2462085c35bca5e00e4f192d43b6b9dbe2f57601d1afde09c"} Feb 25 07:06:52 crc kubenswrapper[4978]: I0225 07:06:52.738766 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"cfdc985f7b5bb332d961a90b34a12545a7880322a9b3eaa992362b444ccba073"} Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.009073 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.015821 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-combined-ca-bundle\") pod \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.015991 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-config-data\") pod \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.016064 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqfpk\" (UniqueName: \"kubernetes.io/projected/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-kube-api-access-gqfpk\") pod \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\" (UID: \"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9\") " Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.020772 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-kube-api-access-gqfpk" (OuterVolumeSpecName: "kube-api-access-gqfpk") pod "fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9" (UID: "fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9"). InnerVolumeSpecName "kube-api-access-gqfpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.042273 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9" (UID: "fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.070985 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-config-data" (OuterVolumeSpecName: "config-data") pod "fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9" (UID: "fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.118357 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqfpk\" (UniqueName: \"kubernetes.io/projected/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-kube-api-access-gqfpk\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.118467 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.118486 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.763048 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-phpsb" Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.763295 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-phpsb" event={"ID":"fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9","Type":"ContainerDied","Data":"ca598b92e0b6bbac0da5a14b63a9ec19f3344ab6d995daf3e1b7377de1aef97b"} Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.763336 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca598b92e0b6bbac0da5a14b63a9ec19f3344ab6d995daf3e1b7377de1aef97b" Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.772914 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"ce64ee6e29ee6ea1ff2d376f27c1117c98cf4060372883033763a60e602660f9"} Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.773141 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerStarted","Data":"d908490f57d2f2a43ad75a1a8576681b6638c3b6cc2217e47bfbe190b3f31f3b"} Feb 25 07:06:53 crc kubenswrapper[4978]: I0225 07:06:53.825431 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=31.078093929 podStartE2EDuration="39.82540524s" podCreationTimestamp="2026-02-25 07:06:14 +0000 UTC" firstStartedPulling="2026-02-25 07:06:42.825742425 +0000 UTC m=+1296.264998884" lastFinishedPulling="2026-02-25 07:06:51.573053726 +0000 UTC m=+1305.012310195" observedRunningTime="2026-02-25 07:06:53.817830352 +0000 UTC m=+1307.257086911" watchObservedRunningTime="2026-02-25 07:06:53.82540524 +0000 UTC m=+1307.264661709" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:53.999531 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66cfcc97b7-7h75q"] Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:53.999956 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fcf1162-2ff4-4faa-89f0-272fd5b2669b" containerName="mariadb-database-create" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:53.999972 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fcf1162-2ff4-4faa-89f0-272fd5b2669b" containerName="mariadb-database-create" Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:53.999989 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a73e3f-747f-4467-95c5-aebf41f4cc9b" containerName="mariadb-database-create" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:53.999996 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a73e3f-747f-4467-95c5-aebf41f4cc9b" containerName="mariadb-database-create" Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:54.000008 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccea553d-1b65-436c-aa35-3b0217c92689" containerName="mariadb-account-create-update" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000017 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccea553d-1b65-436c-aa35-3b0217c92689" containerName="mariadb-account-create-update" Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:54.000031 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1878ccac-42b1-4199-9008-eca6dcad41f6" containerName="mariadb-database-create" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000039 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1878ccac-42b1-4199-9008-eca6dcad41f6" containerName="mariadb-database-create" Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:54.000052 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3143797-ced6-46e4-a06a-b3790384dc8c" containerName="mariadb-account-create-update" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000059 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3143797-ced6-46e4-a06a-b3790384dc8c" containerName="mariadb-account-create-update" Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:54.000078 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333c5a0b-874e-45cc-a44e-4b836b4bd850" containerName="mariadb-account-create-update" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000086 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="333c5a0b-874e-45cc-a44e-4b836b4bd850" containerName="mariadb-account-create-update" Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:54.000113 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330d6675-e003-4a9c-862a-88a54e67877e" containerName="swift-ring-rebalance" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000122 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="330d6675-e003-4a9c-862a-88a54e67877e" containerName="swift-ring-rebalance" Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:54.000137 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9" containerName="keystone-db-sync" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000146 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9" containerName="keystone-db-sync" Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:54.000159 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df970d60-1d8e-497c-b86e-f1787a3f3bd8" containerName="ovn-config" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000166 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="df970d60-1d8e-497c-b86e-f1787a3f3bd8" containerName="ovn-config" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000383 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fcf1162-2ff4-4faa-89f0-272fd5b2669b" containerName="mariadb-database-create" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000406 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3143797-ced6-46e4-a06a-b3790384dc8c" containerName="mariadb-account-create-update" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000421 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9" containerName="keystone-db-sync" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000433 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a73e3f-747f-4467-95c5-aebf41f4cc9b" containerName="mariadb-database-create" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000444 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="df970d60-1d8e-497c-b86e-f1787a3f3bd8" containerName="ovn-config" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000456 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1878ccac-42b1-4199-9008-eca6dcad41f6" containerName="mariadb-database-create" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000463 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="333c5a0b-874e-45cc-a44e-4b836b4bd850" containerName="mariadb-account-create-update" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000475 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccea553d-1b65-436c-aa35-3b0217c92689" containerName="mariadb-account-create-update" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.000495 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="330d6675-e003-4a9c-862a-88a54e67877e" containerName="swift-ring-rebalance" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.001495 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.013603 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sfj52"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.023624 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.027435 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66cfcc97b7-7h75q"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.031247 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.031441 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dvmg8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.031546 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.031661 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.031795 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.038074 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sfj52"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136455 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-fernet-keys\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136503 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-config\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136536 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-sb\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136554 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-dns-svc\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136575 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-credential-keys\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136594 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hw7z\" (UniqueName: \"kubernetes.io/projected/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-kube-api-access-5hw7z\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136613 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f4mk\" (UniqueName: \"kubernetes.io/projected/8742604d-31f7-4533-be72-ed92a2e290c8-kube-api-access-2f4mk\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136631 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-scripts\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136646 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-config-data\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136688 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-nb\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.136711 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-combined-ca-bundle\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.151290 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.153327 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.155048 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.155290 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.165392 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.188466 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-tc5ks"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.189521 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.193770 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zzfvc" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.198489 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.199106 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.226267 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tc5ks"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.236191 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66cfcc97b7-7h75q"] Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:54.236822 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-5hw7z ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" podUID="7ecda08f-b8ff-4c8b-ab54-c55791e8868a" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237235 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-nb\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237269 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-combined-ca-bundle\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237324 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-fernet-keys\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237346 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-config\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237391 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-sb\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237408 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-dns-svc\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237427 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-credential-keys\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237445 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hw7z\" (UniqueName: \"kubernetes.io/projected/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-kube-api-access-5hw7z\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237460 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f4mk\" (UniqueName: \"kubernetes.io/projected/8742604d-31f7-4533-be72-ed92a2e290c8-kube-api-access-2f4mk\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237475 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-scripts\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.237494 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-config-data\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.242353 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-sb\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.242789 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-dns-svc\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.245306 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-config\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.245845 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-nb\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.246490 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-combined-ca-bundle\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.251173 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-fernet-keys\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.260936 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-config-data\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.273590 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-scripts\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.281515 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-credential-keys\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.282105 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hw7z\" (UniqueName: \"kubernetes.io/projected/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-kube-api-access-5hw7z\") pod \"dnsmasq-dns-66cfcc97b7-7h75q\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.290614 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b68b66885-6nm8s"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.291903 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.294635 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.315717 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f4mk\" (UniqueName: \"kubernetes.io/projected/8742604d-31f7-4533-be72-ed92a2e290c8-kube-api-access-2f4mk\") pod \"keystone-bootstrap-sfj52\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.324192 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2sqb8"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.325223 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.327745 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.327965 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zhc9h" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.329654 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.331534 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b68b66885-6nm8s"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342205 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342284 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342335 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-combined-ca-bundle\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342389 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-config-data\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342407 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kq7t\" (UniqueName: \"kubernetes.io/projected/f812baa4-8d36-4a69-b0c1-49fb714beced-kube-api-access-5kq7t\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342432 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-scripts\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342453 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-log-httpd\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342470 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342493 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9ttn\" (UniqueName: \"kubernetes.io/projected/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-kube-api-access-s9ttn\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342533 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-config-data\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342563 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-scripts\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342580 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-etc-machine-id\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-run-httpd\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342620 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-db-sync-config-data\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.342735 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2sqb8"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.373434 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-xpwds"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.395722 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.401227 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.401579 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-zx5c6" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.448725 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzwzj\" (UniqueName: \"kubernetes.io/projected/83a291e6-466e-456d-8674-690f1d8d4a1b-kube-api-access-kzwzj\") pod \"neutron-db-sync-2sqb8\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.448789 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-config-data\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.448815 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kq7t\" (UniqueName: \"kubernetes.io/projected/f812baa4-8d36-4a69-b0c1-49fb714beced-kube-api-access-5kq7t\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.448838 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvc85\" (UniqueName: \"kubernetes.io/projected/6da3632a-96b9-4f97-bddd-7afe0739b449-kube-api-access-jvc85\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.448864 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-scripts\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.448886 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-log-httpd\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.448903 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.448925 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-config\") pod \"neutron-db-sync-2sqb8\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.448940 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9ttn\" (UniqueName: \"kubernetes.io/projected/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-kube-api-access-s9ttn\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.448969 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-svc\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.450104 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xpwds"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.451556 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-sb\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.451666 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-config-data\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.451758 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-combined-ca-bundle\") pod \"neutron-db-sync-2sqb8\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.451871 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-scripts\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.451953 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-etc-machine-id\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.452017 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-run-httpd\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.452110 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-db-sync-config-data\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.452198 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-swift-storage-0\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.452274 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.452421 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-nb\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.452501 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-combined-ca-bundle\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.452598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-config\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.453812 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-config-data\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.454888 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-run-httpd\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.455111 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-log-httpd\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.456189 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-config-data\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.456404 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.456453 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-etc-machine-id\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.457087 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-scripts\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.459447 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.460476 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-db-sync-config-data\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.463839 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-scripts\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.465674 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-combined-ca-bundle\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.475733 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kq7t\" (UniqueName: \"kubernetes.io/projected/f812baa4-8d36-4a69-b0c1-49fb714beced-kube-api-access-5kq7t\") pod \"ceilometer-0\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.483060 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9ttn\" (UniqueName: \"kubernetes.io/projected/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-kube-api-access-s9ttn\") pod \"cinder-db-sync-tc5ks\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.499168 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b68b66885-6nm8s"] Feb 25 07:06:54 crc kubenswrapper[4978]: E0225 07:06:54.503962 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-jvc85 ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" podUID="6da3632a-96b9-4f97-bddd-7afe0739b449" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.509859 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-flx4x"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.510999 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.512842 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.513333 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-nzrxz" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.513494 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.513648 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.519966 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-flx4x"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.535683 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bd5bd7985-l4b4x"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.537029 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.553860 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-combined-ca-bundle\") pod \"neutron-db-sync-2sqb8\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.553901 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-svc\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.553937 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-combined-ca-bundle\") pod \"barbican-db-sync-xpwds\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.553958 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.553984 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ee32dc0-fe37-4be7-b87c-f605febf9692-logs\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554007 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-swift-storage-0\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554036 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-config-data\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554058 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554081 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-config\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554097 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-nb\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554113 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kbpn\" (UniqueName: \"kubernetes.io/projected/2ee32dc0-fe37-4be7-b87c-f605febf9692-kube-api-access-4kbpn\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554146 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-config\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554163 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzwzj\" (UniqueName: \"kubernetes.io/projected/83a291e6-466e-456d-8674-690f1d8d4a1b-kube-api-access-kzwzj\") pod \"neutron-db-sync-2sqb8\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554183 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-swift-storage-0\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554204 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvc85\" (UniqueName: \"kubernetes.io/projected/6da3632a-96b9-4f97-bddd-7afe0739b449-kube-api-access-jvc85\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554224 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-combined-ca-bundle\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554247 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-db-sync-config-data\") pod \"barbican-db-sync-xpwds\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554271 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhl77\" (UniqueName: \"kubernetes.io/projected/4d7807bf-3cea-45ed-b534-29b26ebe2c12-kube-api-access-bhl77\") pod \"barbican-db-sync-xpwds\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554287 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs7j2\" (UniqueName: \"kubernetes.io/projected/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-kube-api-access-hs7j2\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554308 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-config\") pod \"neutron-db-sync-2sqb8\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554335 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-svc\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554356 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-scripts\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.554436 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-sb\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.555215 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-sb\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.556892 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-swift-storage-0\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.558036 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-config\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.558579 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-svc\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.558783 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-nb\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.559119 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-combined-ca-bundle\") pod \"neutron-db-sync-2sqb8\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.560561 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-config\") pod \"neutron-db-sync-2sqb8\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.573114 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd5bd7985-l4b4x"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.575573 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzwzj\" (UniqueName: \"kubernetes.io/projected/83a291e6-466e-456d-8674-690f1d8d4a1b-kube-api-access-kzwzj\") pod \"neutron-db-sync-2sqb8\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.577186 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvc85\" (UniqueName: \"kubernetes.io/projected/6da3632a-96b9-4f97-bddd-7afe0739b449-kube-api-access-jvc85\") pod \"dnsmasq-dns-5b68b66885-6nm8s\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657316 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-swift-storage-0\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657663 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-combined-ca-bundle\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657700 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-db-sync-config-data\") pod \"barbican-db-sync-xpwds\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657732 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhl77\" (UniqueName: \"kubernetes.io/projected/4d7807bf-3cea-45ed-b534-29b26ebe2c12-kube-api-access-bhl77\") pod \"barbican-db-sync-xpwds\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657759 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs7j2\" (UniqueName: \"kubernetes.io/projected/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-kube-api-access-hs7j2\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657802 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-scripts\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657847 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-svc\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657878 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-combined-ca-bundle\") pod \"barbican-db-sync-xpwds\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657903 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657927 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ee32dc0-fe37-4be7-b87c-f605febf9692-logs\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657957 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-config-data\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.657975 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.658002 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-config\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.658026 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kbpn\" (UniqueName: \"kubernetes.io/projected/2ee32dc0-fe37-4be7-b87c-f605febf9692-kube-api-access-4kbpn\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.659291 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-swift-storage-0\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.665805 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-combined-ca-bundle\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.670243 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-db-sync-config-data\") pod \"barbican-db-sync-xpwds\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.674081 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ee32dc0-fe37-4be7-b87c-f605febf9692-logs\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.674737 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-svc\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.676022 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-scripts\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.683700 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-nb\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.684319 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-sb\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.689436 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs7j2\" (UniqueName: \"kubernetes.io/projected/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-kube-api-access-hs7j2\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.692918 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-combined-ca-bundle\") pod \"barbican-db-sync-xpwds\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.693159 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-config-data\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.697135 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-config\") pod \"dnsmasq-dns-7bd5bd7985-l4b4x\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.703850 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhl77\" (UniqueName: \"kubernetes.io/projected/4d7807bf-3cea-45ed-b534-29b26ebe2c12-kube-api-access-bhl77\") pod \"barbican-db-sync-xpwds\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.704041 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kbpn\" (UniqueName: \"kubernetes.io/projected/2ee32dc0-fe37-4be7-b87c-f605febf9692-kube-api-access-4kbpn\") pod \"placement-db-sync-flx4x\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.738811 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sfj52"] Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.770411 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.771884 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.788987 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sfj52" event={"ID":"8742604d-31f7-4533-be72-ed92a2e290c8","Type":"ContainerStarted","Data":"ab6be39b1e348e8136f9cc58571bc0ca00746634a137c8dbe0423ad9722dec00"} Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.789008 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.789019 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.794341 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xpwds" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.831553 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-flx4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.835043 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.844792 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.865851 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.969868 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-sb\") pod \"6da3632a-96b9-4f97-bddd-7afe0739b449\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.969933 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvc85\" (UniqueName: \"kubernetes.io/projected/6da3632a-96b9-4f97-bddd-7afe0739b449-kube-api-access-jvc85\") pod \"6da3632a-96b9-4f97-bddd-7afe0739b449\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.969982 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-nb\") pod \"6da3632a-96b9-4f97-bddd-7afe0739b449\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.970019 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-dns-svc\") pod \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.970066 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-config\") pod \"6da3632a-96b9-4f97-bddd-7afe0739b449\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.970092 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-config\") pod \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.970161 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hw7z\" (UniqueName: \"kubernetes.io/projected/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-kube-api-access-5hw7z\") pod \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.970195 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-svc\") pod \"6da3632a-96b9-4f97-bddd-7afe0739b449\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.970221 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-nb\") pod \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.970253 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-sb\") pod \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\" (UID: \"7ecda08f-b8ff-4c8b-ab54-c55791e8868a\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.970274 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-swift-storage-0\") pod \"6da3632a-96b9-4f97-bddd-7afe0739b449\" (UID: \"6da3632a-96b9-4f97-bddd-7afe0739b449\") " Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.971976 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7ecda08f-b8ff-4c8b-ab54-c55791e8868a" (UID: "7ecda08f-b8ff-4c8b-ab54-c55791e8868a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.972297 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6da3632a-96b9-4f97-bddd-7afe0739b449" (UID: "6da3632a-96b9-4f97-bddd-7afe0739b449"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.972592 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7ecda08f-b8ff-4c8b-ab54-c55791e8868a" (UID: "7ecda08f-b8ff-4c8b-ab54-c55791e8868a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.973775 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6da3632a-96b9-4f97-bddd-7afe0739b449" (UID: "6da3632a-96b9-4f97-bddd-7afe0739b449"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.974278 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6da3632a-96b9-4f97-bddd-7afe0739b449" (UID: "6da3632a-96b9-4f97-bddd-7afe0739b449"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.974580 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-config" (OuterVolumeSpecName: "config") pod "6da3632a-96b9-4f97-bddd-7afe0739b449" (UID: "6da3632a-96b9-4f97-bddd-7afe0739b449"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.974862 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-config" (OuterVolumeSpecName: "config") pod "7ecda08f-b8ff-4c8b-ab54-c55791e8868a" (UID: "7ecda08f-b8ff-4c8b-ab54-c55791e8868a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.977015 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-kube-api-access-5hw7z" (OuterVolumeSpecName: "kube-api-access-5hw7z") pod "7ecda08f-b8ff-4c8b-ab54-c55791e8868a" (UID: "7ecda08f-b8ff-4c8b-ab54-c55791e8868a"). InnerVolumeSpecName "kube-api-access-5hw7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.981455 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da3632a-96b9-4f97-bddd-7afe0739b449-kube-api-access-jvc85" (OuterVolumeSpecName: "kube-api-access-jvc85") pod "6da3632a-96b9-4f97-bddd-7afe0739b449" (UID: "6da3632a-96b9-4f97-bddd-7afe0739b449"). InnerVolumeSpecName "kube-api-access-jvc85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.982641 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6da3632a-96b9-4f97-bddd-7afe0739b449" (UID: "6da3632a-96b9-4f97-bddd-7afe0739b449"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:54 crc kubenswrapper[4978]: I0225 07:06:54.982822 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7ecda08f-b8ff-4c8b-ab54-c55791e8868a" (UID: "7ecda08f-b8ff-4c8b-ab54-c55791e8868a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071762 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvc85\" (UniqueName: \"kubernetes.io/projected/6da3632a-96b9-4f97-bddd-7afe0739b449-kube-api-access-jvc85\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071823 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071835 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071844 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071852 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071860 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hw7z\" (UniqueName: \"kubernetes.io/projected/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-kube-api-access-5hw7z\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071868 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071875 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071885 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ecda08f-b8ff-4c8b-ab54-c55791e8868a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071892 4978 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.071900 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6da3632a-96b9-4f97-bddd-7afe0739b449-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.109695 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-tc5ks"] Feb 25 07:06:55 crc kubenswrapper[4978]: W0225 07:06:55.140017 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf18ee3ff_994c_46c4_8f0c_76b1ee45da50.slice/crio-a1c9a063536805161964b6f751b3329edfe8bd7212f701ca56b03986b200184a WatchSource:0}: Error finding container a1c9a063536805161964b6f751b3329edfe8bd7212f701ca56b03986b200184a: Status 404 returned error can't find the container with id a1c9a063536805161964b6f751b3329edfe8bd7212f701ca56b03986b200184a Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.396592 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.526908 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2sqb8"] Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.533904 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xpwds"] Feb 25 07:06:55 crc kubenswrapper[4978]: W0225 07:06:55.537287 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a291e6_466e_456d_8674_690f1d8d4a1b.slice/crio-4ccd4bd3cb023bfc4bf1cd00518eefa858fd7ee5885638f0ab150e5bfd3beb84 WatchSource:0}: Error finding container 4ccd4bd3cb023bfc4bf1cd00518eefa858fd7ee5885638f0ab150e5bfd3beb84: Status 404 returned error can't find the container with id 4ccd4bd3cb023bfc4bf1cd00518eefa858fd7ee5885638f0ab150e5bfd3beb84 Feb 25 07:06:55 crc kubenswrapper[4978]: W0225 07:06:55.546906 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d7807bf_3cea_45ed_b534_29b26ebe2c12.slice/crio-e767316480b28229f447bef48ca6576c4bc55b93b5f0975506f046f01c71b538 WatchSource:0}: Error finding container e767316480b28229f447bef48ca6576c4bc55b93b5f0975506f046f01c71b538: Status 404 returned error can't find the container with id e767316480b28229f447bef48ca6576c4bc55b93b5f0975506f046f01c71b538 Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.631995 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-flx4x"] Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.727120 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd5bd7985-l4b4x"] Feb 25 07:06:55 crc kubenswrapper[4978]: W0225 07:06:55.731342 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eb078ad_fc4f_479f_b330_69f8b2fe62e8.slice/crio-9b33afadec79918c6788b4c2c7e34bd710efb3369f49347121c909ddf7562a7f WatchSource:0}: Error finding container 9b33afadec79918c6788b4c2c7e34bd710efb3369f49347121c909ddf7562a7f: Status 404 returned error can't find the container with id 9b33afadec79918c6788b4c2c7e34bd710efb3369f49347121c909ddf7562a7f Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.812484 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tc5ks" event={"ID":"f18ee3ff-994c-46c4-8f0c-76b1ee45da50","Type":"ContainerStarted","Data":"a1c9a063536805161964b6f751b3329edfe8bd7212f701ca56b03986b200184a"} Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.816785 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sfj52" event={"ID":"8742604d-31f7-4533-be72-ed92a2e290c8","Type":"ContainerStarted","Data":"8032219bff21948de2c2feac0cab9661d2687ea4f33eb4dc5b640069888bce22"} Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.818729 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" event={"ID":"1eb078ad-fc4f-479f-b330-69f8b2fe62e8","Type":"ContainerStarted","Data":"9b33afadec79918c6788b4c2c7e34bd710efb3369f49347121c909ddf7562a7f"} Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.820181 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2sqb8" event={"ID":"83a291e6-466e-456d-8674-690f1d8d4a1b","Type":"ContainerStarted","Data":"cd94dd53f9e56529db66e99175ac2ac797be2dd86cf8eb9e8918194438ca01bf"} Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.820203 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2sqb8" event={"ID":"83a291e6-466e-456d-8674-690f1d8d4a1b","Type":"ContainerStarted","Data":"4ccd4bd3cb023bfc4bf1cd00518eefa858fd7ee5885638f0ab150e5bfd3beb84"} Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.822525 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-flx4x" event={"ID":"2ee32dc0-fe37-4be7-b87c-f605febf9692","Type":"ContainerStarted","Data":"cf67e3bb996efbb0b427d99418654a8ed5138586530681652bfc426d0022e8a3"} Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.823680 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f812baa4-8d36-4a69-b0c1-49fb714beced","Type":"ContainerStarted","Data":"684932be1fff80d666c4c7817523272c34daa5f728400430e15d1e71eb499300"} Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.824513 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xpwds" event={"ID":"4d7807bf-3cea-45ed-b534-29b26ebe2c12","Type":"ContainerStarted","Data":"e767316480b28229f447bef48ca6576c4bc55b93b5f0975506f046f01c71b538"} Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.825712 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b68b66885-6nm8s" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.826234 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7ts6t" event={"ID":"4bd97fe5-291b-4130-98c7-4ffbc0cba189","Type":"ContainerStarted","Data":"6d5a6575f5371601ecab8ff2f8053d7907cb7b78d92f5ec99f37ea255f9dd065"} Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.826625 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cfcc97b7-7h75q" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.843255 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sfj52" podStartSLOduration=2.84323724 podStartE2EDuration="2.84323724s" podCreationTimestamp="2026-02-25 07:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:55.836143257 +0000 UTC m=+1309.275399726" watchObservedRunningTime="2026-02-25 07:06:55.84323724 +0000 UTC m=+1309.282493699" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.855195 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-7ts6t" podStartSLOduration=3.695873292 podStartE2EDuration="33.85517614s" podCreationTimestamp="2026-02-25 07:06:22 +0000 UTC" firstStartedPulling="2026-02-25 07:06:24.609679754 +0000 UTC m=+1278.048936213" lastFinishedPulling="2026-02-25 07:06:54.768982602 +0000 UTC m=+1308.208239061" observedRunningTime="2026-02-25 07:06:55.854605883 +0000 UTC m=+1309.293862342" watchObservedRunningTime="2026-02-25 07:06:55.85517614 +0000 UTC m=+1309.294432599" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.875282 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2sqb8" podStartSLOduration=1.875263865 podStartE2EDuration="1.875263865s" podCreationTimestamp="2026-02-25 07:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:55.869209613 +0000 UTC m=+1309.308466072" watchObservedRunningTime="2026-02-25 07:06:55.875263865 +0000 UTC m=+1309.314520324" Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.951789 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b68b66885-6nm8s"] Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.954536 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b68b66885-6nm8s"] Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.974967 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66cfcc97b7-7h75q"] Feb 25 07:06:55 crc kubenswrapper[4978]: I0225 07:06:55.981237 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66cfcc97b7-7h75q"] Feb 25 07:06:56 crc kubenswrapper[4978]: I0225 07:06:56.442475 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:06:56 crc kubenswrapper[4978]: I0225 07:06:56.838609 4978 generic.go:334] "Generic (PLEG): container finished" podID="1eb078ad-fc4f-479f-b330-69f8b2fe62e8" containerID="9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd" exitCode=0 Feb 25 07:06:56 crc kubenswrapper[4978]: I0225 07:06:56.839448 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" event={"ID":"1eb078ad-fc4f-479f-b330-69f8b2fe62e8","Type":"ContainerDied","Data":"9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd"} Feb 25 07:06:57 crc kubenswrapper[4978]: I0225 07:06:57.348063 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da3632a-96b9-4f97-bddd-7afe0739b449" path="/var/lib/kubelet/pods/6da3632a-96b9-4f97-bddd-7afe0739b449/volumes" Feb 25 07:06:57 crc kubenswrapper[4978]: I0225 07:06:57.348737 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ecda08f-b8ff-4c8b-ab54-c55791e8868a" path="/var/lib/kubelet/pods/7ecda08f-b8ff-4c8b-ab54-c55791e8868a/volumes" Feb 25 07:06:57 crc kubenswrapper[4978]: I0225 07:06:57.849188 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" event={"ID":"1eb078ad-fc4f-479f-b330-69f8b2fe62e8","Type":"ContainerStarted","Data":"33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970"} Feb 25 07:06:57 crc kubenswrapper[4978]: I0225 07:06:57.850299 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:06:58 crc kubenswrapper[4978]: I0225 07:06:58.858683 4978 generic.go:334] "Generic (PLEG): container finished" podID="8742604d-31f7-4533-be72-ed92a2e290c8" containerID="8032219bff21948de2c2feac0cab9661d2687ea4f33eb4dc5b640069888bce22" exitCode=0 Feb 25 07:06:58 crc kubenswrapper[4978]: I0225 07:06:58.858916 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sfj52" event={"ID":"8742604d-31f7-4533-be72-ed92a2e290c8","Type":"ContainerDied","Data":"8032219bff21948de2c2feac0cab9661d2687ea4f33eb4dc5b640069888bce22"} Feb 25 07:06:58 crc kubenswrapper[4978]: I0225 07:06:58.883254 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" podStartSLOduration=4.8832302389999995 podStartE2EDuration="4.883230239s" podCreationTimestamp="2026-02-25 07:06:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:06:57.873857236 +0000 UTC m=+1311.313113705" watchObservedRunningTime="2026-02-25 07:06:58.883230239 +0000 UTC m=+1312.322486698" Feb 25 07:07:04 crc kubenswrapper[4978]: I0225 07:07:04.868136 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:07:04 crc kubenswrapper[4978]: I0225 07:07:04.931111 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77cf9b784c-524z4"] Feb 25 07:07:04 crc kubenswrapper[4978]: I0225 07:07:04.931429 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" podUID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerName="dnsmasq-dns" containerID="cri-o://07a057de05f5bc6d8d6e12685b1aaa0d0969c94b182f9211e4ace6841f755b5b" gracePeriod=10 Feb 25 07:07:04 crc kubenswrapper[4978]: I0225 07:07:04.959517 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" podUID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.690703 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.798490 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-config-data\") pod \"8742604d-31f7-4533-be72-ed92a2e290c8\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.798543 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f4mk\" (UniqueName: \"kubernetes.io/projected/8742604d-31f7-4533-be72-ed92a2e290c8-kube-api-access-2f4mk\") pod \"8742604d-31f7-4533-be72-ed92a2e290c8\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.798587 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-credential-keys\") pod \"8742604d-31f7-4533-be72-ed92a2e290c8\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.798660 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-fernet-keys\") pod \"8742604d-31f7-4533-be72-ed92a2e290c8\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.798755 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-scripts\") pod \"8742604d-31f7-4533-be72-ed92a2e290c8\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.798808 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-combined-ca-bundle\") pod \"8742604d-31f7-4533-be72-ed92a2e290c8\" (UID: \"8742604d-31f7-4533-be72-ed92a2e290c8\") " Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.805487 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-scripts" (OuterVolumeSpecName: "scripts") pod "8742604d-31f7-4533-be72-ed92a2e290c8" (UID: "8742604d-31f7-4533-be72-ed92a2e290c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.806082 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8742604d-31f7-4533-be72-ed92a2e290c8" (UID: "8742604d-31f7-4533-be72-ed92a2e290c8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.810527 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8742604d-31f7-4533-be72-ed92a2e290c8" (UID: "8742604d-31f7-4533-be72-ed92a2e290c8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.814853 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8742604d-31f7-4533-be72-ed92a2e290c8-kube-api-access-2f4mk" (OuterVolumeSpecName: "kube-api-access-2f4mk") pod "8742604d-31f7-4533-be72-ed92a2e290c8" (UID: "8742604d-31f7-4533-be72-ed92a2e290c8"). InnerVolumeSpecName "kube-api-access-2f4mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.838965 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8742604d-31f7-4533-be72-ed92a2e290c8" (UID: "8742604d-31f7-4533-be72-ed92a2e290c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.856465 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-config-data" (OuterVolumeSpecName: "config-data") pod "8742604d-31f7-4533-be72-ed92a2e290c8" (UID: "8742604d-31f7-4533-be72-ed92a2e290c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.900685 4978 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.900714 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.900723 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.900732 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.900740 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f4mk\" (UniqueName: \"kubernetes.io/projected/8742604d-31f7-4533-be72-ed92a2e290c8-kube-api-access-2f4mk\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.900748 4978 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8742604d-31f7-4533-be72-ed92a2e290c8-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.926541 4978 generic.go:334] "Generic (PLEG): container finished" podID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerID="07a057de05f5bc6d8d6e12685b1aaa0d0969c94b182f9211e4ace6841f755b5b" exitCode=0 Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.926596 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" event={"ID":"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038","Type":"ContainerDied","Data":"07a057de05f5bc6d8d6e12685b1aaa0d0969c94b182f9211e4ace6841f755b5b"} Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.928130 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sfj52" event={"ID":"8742604d-31f7-4533-be72-ed92a2e290c8","Type":"ContainerDied","Data":"ab6be39b1e348e8136f9cc58571bc0ca00746634a137c8dbe0423ad9722dec00"} Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.928154 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab6be39b1e348e8136f9cc58571bc0ca00746634a137c8dbe0423ad9722dec00" Feb 25 07:07:05 crc kubenswrapper[4978]: I0225 07:07:05.928215 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sfj52" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.787879 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sfj52"] Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.796162 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sfj52"] Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.872325 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-q4lmf"] Feb 25 07:07:06 crc kubenswrapper[4978]: E0225 07:07:06.872687 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8742604d-31f7-4533-be72-ed92a2e290c8" containerName="keystone-bootstrap" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.872703 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8742604d-31f7-4533-be72-ed92a2e290c8" containerName="keystone-bootstrap" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.872854 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8742604d-31f7-4533-be72-ed92a2e290c8" containerName="keystone-bootstrap" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.873331 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.876407 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.876725 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.876932 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.877578 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.877993 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-dvmg8" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.894180 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-q4lmf"] Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.915786 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-credential-keys\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.916728 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsmtl\" (UniqueName: \"kubernetes.io/projected/df4bb775-9429-45bf-af46-93dcf3b4a82d-kube-api-access-bsmtl\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.916860 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-fernet-keys\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.917037 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-combined-ca-bundle\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.917200 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-scripts\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:06 crc kubenswrapper[4978]: I0225 07:07:06.917312 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-config-data\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.018909 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-combined-ca-bundle\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.018975 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-scripts\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.018998 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-config-data\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.019066 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-credential-keys\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.019109 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsmtl\" (UniqueName: \"kubernetes.io/projected/df4bb775-9429-45bf-af46-93dcf3b4a82d-kube-api-access-bsmtl\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.019138 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-fernet-keys\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.025012 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-combined-ca-bundle\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.025402 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-scripts\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.025702 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-fernet-keys\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.026983 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-config-data\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.027287 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-credential-keys\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.040682 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsmtl\" (UniqueName: \"kubernetes.io/projected/df4bb775-9429-45bf-af46-93dcf3b4a82d-kube-api-access-bsmtl\") pod \"keystone-bootstrap-q4lmf\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.194190 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:07 crc kubenswrapper[4978]: I0225 07:07:07.348322 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8742604d-31f7-4533-be72-ed92a2e290c8" path="/var/lib/kubelet/pods/8742604d-31f7-4533-be72-ed92a2e290c8/volumes" Feb 25 07:07:09 crc kubenswrapper[4978]: I0225 07:07:09.959540 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" podUID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Feb 25 07:07:10 crc kubenswrapper[4978]: I0225 07:07:10.979200 4978 generic.go:334] "Generic (PLEG): container finished" podID="4bd97fe5-291b-4130-98c7-4ffbc0cba189" containerID="6d5a6575f5371601ecab8ff2f8053d7907cb7b78d92f5ec99f37ea255f9dd065" exitCode=0 Feb 25 07:07:10 crc kubenswrapper[4978]: I0225 07:07:10.979290 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7ts6t" event={"ID":"4bd97fe5-291b-4130-98c7-4ffbc0cba189","Type":"ContainerDied","Data":"6d5a6575f5371601ecab8ff2f8053d7907cb7b78d92f5ec99f37ea255f9dd065"} Feb 25 07:07:15 crc kubenswrapper[4978]: E0225 07:07:15.727789 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4b52f033a38f9c82ba4b5a24a50b343738c826babad6993e6dc744a6981e8cab" Feb 25 07:07:15 crc kubenswrapper[4978]: E0225 07:07:15.728696 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4b52f033a38f9c82ba4b5a24a50b343738c826babad6993e6dc744a6981e8cab,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bhl77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-xpwds_openstack(4d7807bf-3cea-45ed-b534-29b26ebe2c12): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 07:07:15 crc kubenswrapper[4978]: E0225 07:07:15.729960 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-xpwds" podUID="4d7807bf-3cea-45ed-b534-29b26ebe2c12" Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.838198 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7ts6t" Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.892773 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-config-data\") pod \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.892848 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mqrr\" (UniqueName: \"kubernetes.io/projected/4bd97fe5-291b-4130-98c7-4ffbc0cba189-kube-api-access-6mqrr\") pod \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.892925 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-combined-ca-bundle\") pod \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.892979 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-db-sync-config-data\") pod \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\" (UID: \"4bd97fe5-291b-4130-98c7-4ffbc0cba189\") " Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.912765 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bd97fe5-291b-4130-98c7-4ffbc0cba189-kube-api-access-6mqrr" (OuterVolumeSpecName: "kube-api-access-6mqrr") pod "4bd97fe5-291b-4130-98c7-4ffbc0cba189" (UID: "4bd97fe5-291b-4130-98c7-4ffbc0cba189"). InnerVolumeSpecName "kube-api-access-6mqrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.915342 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4bd97fe5-291b-4130-98c7-4ffbc0cba189" (UID: "4bd97fe5-291b-4130-98c7-4ffbc0cba189"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.921598 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bd97fe5-291b-4130-98c7-4ffbc0cba189" (UID: "4bd97fe5-291b-4130-98c7-4ffbc0cba189"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.947428 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-config-data" (OuterVolumeSpecName: "config-data") pod "4bd97fe5-291b-4130-98c7-4ffbc0cba189" (UID: "4bd97fe5-291b-4130-98c7-4ffbc0cba189"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.994573 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.994637 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mqrr\" (UniqueName: \"kubernetes.io/projected/4bd97fe5-291b-4130-98c7-4ffbc0cba189-kube-api-access-6mqrr\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.994657 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:15 crc kubenswrapper[4978]: I0225 07:07:15.994671 4978 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4bd97fe5-291b-4130-98c7-4ffbc0cba189-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:16 crc kubenswrapper[4978]: I0225 07:07:16.042327 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-7ts6t" Feb 25 07:07:16 crc kubenswrapper[4978]: I0225 07:07:16.042470 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-7ts6t" event={"ID":"4bd97fe5-291b-4130-98c7-4ffbc0cba189","Type":"ContainerDied","Data":"91a0fb39ecfb77aaf572f0945651a4d418e46efeb2288c713cf21f90ad14899b"} Feb 25 07:07:16 crc kubenswrapper[4978]: I0225 07:07:16.042509 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91a0fb39ecfb77aaf572f0945651a4d418e46efeb2288c713cf21f90ad14899b" Feb 25 07:07:16 crc kubenswrapper[4978]: E0225 07:07:16.044642 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4b52f033a38f9c82ba4b5a24a50b343738c826babad6993e6dc744a6981e8cab\\\"\"" pod="openstack/barbican-db-sync-xpwds" podUID="4d7807bf-3cea-45ed-b534-29b26ebe2c12" Feb 25 07:07:16 crc kubenswrapper[4978]: I0225 07:07:16.540732 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:07:16 crc kubenswrapper[4978]: I0225 07:07:16.540804 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:07:16 crc kubenswrapper[4978]: E0225 07:07:16.928741 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:76b63ab76a2865b38e0fc61c5800b33683b6bd2f6b77eb1c791aad230bbebead" Feb 25 07:07:16 crc kubenswrapper[4978]: E0225 07:07:16.929110 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:76b63ab76a2865b38e0fc61c5800b33683b6bd2f6b77eb1c791aad230bbebead,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s9ttn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-tc5ks_openstack(f18ee3ff-994c-46c4-8f0c-76b1ee45da50): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 07:07:16 crc kubenswrapper[4978]: E0225 07:07:16.932472 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-tc5ks" podUID="f18ee3ff-994c-46c4-8f0c-76b1ee45da50" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.089806 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" event={"ID":"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038","Type":"ContainerDied","Data":"6e793f91d8434d976c51290f95772c5b4b1b7e6970d7c573e210f3569032b141"} Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.090180 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e793f91d8434d976c51290f95772c5b4b1b7e6970d7c573e210f3569032b141" Feb 25 07:07:17 crc kubenswrapper[4978]: E0225 07:07:17.093634 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:76b63ab76a2865b38e0fc61c5800b33683b6bd2f6b77eb1c791aad230bbebead\\\"\"" pod="openstack/cinder-db-sync-tc5ks" podUID="f18ee3ff-994c-46c4-8f0c-76b1ee45da50" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.222000 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86df48c85f-nmjcf"] Feb 25 07:07:17 crc kubenswrapper[4978]: E0225 07:07:17.222401 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bd97fe5-291b-4130-98c7-4ffbc0cba189" containerName="glance-db-sync" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.222419 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bd97fe5-291b-4130-98c7-4ffbc0cba189" containerName="glance-db-sync" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.222598 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bd97fe5-291b-4130-98c7-4ffbc0cba189" containerName="glance-db-sync" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.223434 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.224655 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.235466 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86df48c85f-nmjcf"] Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.315115 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-sb\") pod \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.315197 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-dns-svc\") pod \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.315217 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-nb\") pod \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.315577 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-config\") pod \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.315620 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2d9b\" (UniqueName: \"kubernetes.io/projected/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-kube-api-access-z2d9b\") pod \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\" (UID: \"203f2e43-da8a-4be2-a1bb-6a8d5fdc9038\") " Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.315850 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-sb\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.315949 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-swift-storage-0\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.316717 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvwc6\" (UniqueName: \"kubernetes.io/projected/ee76c60c-2ef4-4ba7-b864-156f132d03d3-kube-api-access-fvwc6\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.316754 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-config\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.316786 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-svc\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.316816 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-nb\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.331586 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-kube-api-access-z2d9b" (OuterVolumeSpecName: "kube-api-access-z2d9b") pod "203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" (UID: "203f2e43-da8a-4be2-a1bb-6a8d5fdc9038"). InnerVolumeSpecName "kube-api-access-z2d9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.419458 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvwc6\" (UniqueName: \"kubernetes.io/projected/ee76c60c-2ef4-4ba7-b864-156f132d03d3-kube-api-access-fvwc6\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.419469 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-q4lmf"] Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.419749 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-config\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.419858 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-svc\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.419944 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-nb\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.420043 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-sb\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.420238 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-swift-storage-0\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.420310 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2d9b\" (UniqueName: \"kubernetes.io/projected/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-kube-api-access-z2d9b\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.420551 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-config\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.420919 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-swift-storage-0\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.421473 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-sb\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.421654 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-svc\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.422165 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-nb\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.441892 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvwc6\" (UniqueName: \"kubernetes.io/projected/ee76c60c-2ef4-4ba7-b864-156f132d03d3-kube-api-access-fvwc6\") pod \"dnsmasq-dns-86df48c85f-nmjcf\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.444064 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.500661 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" (UID: "203f2e43-da8a-4be2-a1bb-6a8d5fdc9038"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.520715 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" (UID: "203f2e43-da8a-4be2-a1bb-6a8d5fdc9038"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.522734 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.522768 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.540831 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-config" (OuterVolumeSpecName: "config") pod "203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" (UID: "203f2e43-da8a-4be2-a1bb-6a8d5fdc9038"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.547131 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" (UID: "203f2e43-da8a-4be2-a1bb-6a8d5fdc9038"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.580529 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.623904 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:17 crc kubenswrapper[4978]: I0225 07:07:17.623928 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.117266 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q4lmf" event={"ID":"df4bb775-9429-45bf-af46-93dcf3b4a82d","Type":"ContainerStarted","Data":"caed680d248b73867bcd7477d8eb3c3cdbb4b1d1ceff2297baed0b3c1eed6964"} Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.117711 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q4lmf" event={"ID":"df4bb775-9429-45bf-af46-93dcf3b4a82d","Type":"ContainerStarted","Data":"191b687dd267b5da9db654307f5907fc2fd7ff78005e8946476615d1b3d994a2"} Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.125231 4978 generic.go:334] "Generic (PLEG): container finished" podID="83a291e6-466e-456d-8674-690f1d8d4a1b" containerID="cd94dd53f9e56529db66e99175ac2ac797be2dd86cf8eb9e8918194438ca01bf" exitCode=0 Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.125283 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2sqb8" event={"ID":"83a291e6-466e-456d-8674-690f1d8d4a1b","Type":"ContainerDied","Data":"cd94dd53f9e56529db66e99175ac2ac797be2dd86cf8eb9e8918194438ca01bf"} Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.127761 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-flx4x" event={"ID":"2ee32dc0-fe37-4be7-b87c-f605febf9692","Type":"ContainerStarted","Data":"6341648a449ad9f4c7c225487199328ca24ce228204b741a39e695b53437b28d"} Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.131113 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f812baa4-8d36-4a69-b0c1-49fb714beced","Type":"ContainerStarted","Data":"8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b"} Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.131395 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.143971 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-q4lmf" podStartSLOduration=12.143954158 podStartE2EDuration="12.143954158s" podCreationTimestamp="2026-02-25 07:07:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:18.130804152 +0000 UTC m=+1331.570060611" watchObservedRunningTime="2026-02-25 07:07:18.143954158 +0000 UTC m=+1331.583210617" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.171445 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-flx4x" podStartSLOduration=2.95537019 podStartE2EDuration="24.171432075s" podCreationTimestamp="2026-02-25 07:06:54 +0000 UTC" firstStartedPulling="2026-02-25 07:06:55.634087111 +0000 UTC m=+1309.073343570" lastFinishedPulling="2026-02-25 07:07:16.850148976 +0000 UTC m=+1330.289405455" observedRunningTime="2026-02-25 07:07:18.166679462 +0000 UTC m=+1331.605935921" watchObservedRunningTime="2026-02-25 07:07:18.171432075 +0000 UTC m=+1331.610688534" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.185639 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77cf9b784c-524z4"] Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.193149 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77cf9b784c-524z4"] Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.255848 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:07:18 crc kubenswrapper[4978]: E0225 07:07:18.256188 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerName="dnsmasq-dns" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.256208 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerName="dnsmasq-dns" Feb 25 07:07:18 crc kubenswrapper[4978]: E0225 07:07:18.256224 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerName="init" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.256231 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerName="init" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.256384 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerName="dnsmasq-dns" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.257180 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.260182 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.260428 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mrzl7" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.260755 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.283352 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.329806 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86df48c85f-nmjcf"] Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.348478 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-config-data\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.348532 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9z4b\" (UniqueName: \"kubernetes.io/projected/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-kube-api-access-h9z4b\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.348554 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.348606 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.348621 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-scripts\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.348675 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.348698 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-logs\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.455265 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.455573 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-scripts\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.455630 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.455658 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-logs\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.455695 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-config-data\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.455748 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9z4b\" (UniqueName: \"kubernetes.io/projected/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-kube-api-access-h9z4b\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.455768 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.458593 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.464444 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.467135 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.467422 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-logs\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.470937 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-scripts\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.480331 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-config-data\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.491250 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.492660 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.496344 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9z4b\" (UniqueName: \"kubernetes.io/projected/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-kube-api-access-h9z4b\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.498206 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.505662 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.546437 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.557124 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-logs\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.557175 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttjn8\" (UniqueName: \"kubernetes.io/projected/bfe41d4c-9ca3-4772-b036-84b5573eb44e-kube-api-access-ttjn8\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.557210 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.557239 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.557294 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.557317 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.557354 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.572043 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.658326 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttjn8\" (UniqueName: \"kubernetes.io/projected/bfe41d4c-9ca3-4772-b036-84b5573eb44e-kube-api-access-ttjn8\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.658404 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.658439 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.658473 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.658500 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.658535 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.658595 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-logs\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.659064 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-logs\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.659434 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.661422 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.664920 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.669352 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.677124 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.683689 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttjn8\" (UniqueName: \"kubernetes.io/projected/bfe41d4c-9ca3-4772-b036-84b5573eb44e-kube-api-access-ttjn8\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.719463 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:18 crc kubenswrapper[4978]: I0225 07:07:18.836825 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.182322 4978 generic.go:334] "Generic (PLEG): container finished" podID="ee76c60c-2ef4-4ba7-b864-156f132d03d3" containerID="429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae" exitCode=0 Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.182567 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" event={"ID":"ee76c60c-2ef4-4ba7-b864-156f132d03d3","Type":"ContainerDied","Data":"429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae"} Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.182668 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" event={"ID":"ee76c60c-2ef4-4ba7-b864-156f132d03d3","Type":"ContainerStarted","Data":"383c66d93eefee72ea72dccec191153daff0abd6495defbc18c476dd0b971e8f"} Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.340981 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" path="/var/lib/kubelet/pods/203f2e43-da8a-4be2-a1bb-6a8d5fdc9038/volumes" Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.615004 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.704535 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.777345 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-config\") pod \"83a291e6-466e-456d-8674-690f1d8d4a1b\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.777451 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-combined-ca-bundle\") pod \"83a291e6-466e-456d-8674-690f1d8d4a1b\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.777553 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzwzj\" (UniqueName: \"kubernetes.io/projected/83a291e6-466e-456d-8674-690f1d8d4a1b-kube-api-access-kzwzj\") pod \"83a291e6-466e-456d-8674-690f1d8d4a1b\" (UID: \"83a291e6-466e-456d-8674-690f1d8d4a1b\") " Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.782153 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a291e6-466e-456d-8674-690f1d8d4a1b-kube-api-access-kzwzj" (OuterVolumeSpecName: "kube-api-access-kzwzj") pod "83a291e6-466e-456d-8674-690f1d8d4a1b" (UID: "83a291e6-466e-456d-8674-690f1d8d4a1b"). InnerVolumeSpecName "kube-api-access-kzwzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.816885 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-config" (OuterVolumeSpecName: "config") pod "83a291e6-466e-456d-8674-690f1d8d4a1b" (UID: "83a291e6-466e-456d-8674-690f1d8d4a1b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.828262 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83a291e6-466e-456d-8674-690f1d8d4a1b" (UID: "83a291e6-466e-456d-8674-690f1d8d4a1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.879588 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.879623 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzwzj\" (UniqueName: \"kubernetes.io/projected/83a291e6-466e-456d-8674-690f1d8d4a1b-kube-api-access-kzwzj\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.879636 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/83a291e6-466e-456d-8674-690f1d8d4a1b-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.953894 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:07:19 crc kubenswrapper[4978]: I0225 07:07:19.959585 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77cf9b784c-524z4" podUID="203f2e43-da8a-4be2-a1bb-6a8d5fdc9038" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.012468 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.192263 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e","Type":"ContainerStarted","Data":"8c3804f11c6eee51db79a152498fadec5b9f4e6f8497e2d9a272cb0edc7e7195"} Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.204861 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2sqb8" event={"ID":"83a291e6-466e-456d-8674-690f1d8d4a1b","Type":"ContainerDied","Data":"4ccd4bd3cb023bfc4bf1cd00518eefa858fd7ee5885638f0ab150e5bfd3beb84"} Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.204906 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ccd4bd3cb023bfc4bf1cd00518eefa858fd7ee5885638f0ab150e5bfd3beb84" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.204982 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2sqb8" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.216060 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" event={"ID":"ee76c60c-2ef4-4ba7-b864-156f132d03d3","Type":"ContainerStarted","Data":"dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e"} Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.216508 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.251212 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" podStartSLOduration=3.251190031 podStartE2EDuration="3.251190031s" podCreationTimestamp="2026-02-25 07:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:20.242439858 +0000 UTC m=+1333.681696317" watchObservedRunningTime="2026-02-25 07:07:20.251190031 +0000 UTC m=+1333.690446490" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.451177 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86df48c85f-nmjcf"] Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.533443 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78b5478d49-qg7sz"] Feb 25 07:07:20 crc kubenswrapper[4978]: E0225 07:07:20.533857 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a291e6-466e-456d-8674-690f1d8d4a1b" containerName="neutron-db-sync" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.533868 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a291e6-466e-456d-8674-690f1d8d4a1b" containerName="neutron-db-sync" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.534050 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a291e6-466e-456d-8674-690f1d8d4a1b" containerName="neutron-db-sync" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.534977 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.553182 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5964547b68-crjwr"] Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.555121 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.563773 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.563856 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-zhc9h" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.563780 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.564524 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.566136 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78b5478d49-qg7sz"] Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.587427 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5964547b68-crjwr"] Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.699766 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-nb\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.700094 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-ovndb-tls-certs\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.700159 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-sb\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.700187 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v8cw\" (UniqueName: \"kubernetes.io/projected/9ee02b80-8075-47b3-8fe8-52b8a2e17371-kube-api-access-9v8cw\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.700252 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-combined-ca-bundle\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.700277 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-httpd-config\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.700311 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-config\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.700432 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-svc\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.700573 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-config\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.700598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-swift-storage-0\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.700644 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9hp9\" (UniqueName: \"kubernetes.io/projected/bd220760-4e92-4cac-b4f0-8eb2c38f67be-kube-api-access-d9hp9\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.721492 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.802929 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-config\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.802992 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-swift-storage-0\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.803038 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9hp9\" (UniqueName: \"kubernetes.io/projected/bd220760-4e92-4cac-b4f0-8eb2c38f67be-kube-api-access-d9hp9\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.803087 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-nb\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.803125 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-ovndb-tls-certs\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.803187 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-sb\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.803212 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v8cw\" (UniqueName: \"kubernetes.io/projected/9ee02b80-8075-47b3-8fe8-52b8a2e17371-kube-api-access-9v8cw\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.803261 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-combined-ca-bundle\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.803293 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-httpd-config\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.803328 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-config\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.803352 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-svc\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.813039 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-svc\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.814159 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-combined-ca-bundle\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.814552 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-sb\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.814702 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-nb\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.814822 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-config\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.814904 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-swift-storage-0\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.816145 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-config\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.817262 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-ovndb-tls-certs\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.817283 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-httpd-config\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.823572 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9hp9\" (UniqueName: \"kubernetes.io/projected/bd220760-4e92-4cac-b4f0-8eb2c38f67be-kube-api-access-d9hp9\") pod \"dnsmasq-dns-78b5478d49-qg7sz\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.826854 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v8cw\" (UniqueName: \"kubernetes.io/projected/9ee02b80-8075-47b3-8fe8-52b8a2e17371-kube-api-access-9v8cw\") pod \"neutron-5964547b68-crjwr\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.866730 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:20 crc kubenswrapper[4978]: I0225 07:07:20.908776 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:21 crc kubenswrapper[4978]: I0225 07:07:21.240422 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e","Type":"ContainerStarted","Data":"607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257"} Feb 25 07:07:21 crc kubenswrapper[4978]: I0225 07:07:21.252310 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bfe41d4c-9ca3-4772-b036-84b5573eb44e","Type":"ContainerStarted","Data":"cdf5b5a9a9e7c2989b301194e51f534b7afb711852c8a5b3ec187c63760b13c7"} Feb 25 07:07:21 crc kubenswrapper[4978]: I0225 07:07:21.269953 4978 generic.go:334] "Generic (PLEG): container finished" podID="2ee32dc0-fe37-4be7-b87c-f605febf9692" containerID="6341648a449ad9f4c7c225487199328ca24ce228204b741a39e695b53437b28d" exitCode=0 Feb 25 07:07:21 crc kubenswrapper[4978]: I0225 07:07:21.270401 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-flx4x" event={"ID":"2ee32dc0-fe37-4be7-b87c-f605febf9692","Type":"ContainerDied","Data":"6341648a449ad9f4c7c225487199328ca24ce228204b741a39e695b53437b28d"} Feb 25 07:07:21 crc kubenswrapper[4978]: I0225 07:07:21.433128 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78b5478d49-qg7sz"] Feb 25 07:07:21 crc kubenswrapper[4978]: I0225 07:07:21.668733 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5964547b68-crjwr"] Feb 25 07:07:21 crc kubenswrapper[4978]: W0225 07:07:21.676032 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ee02b80_8075_47b3_8fe8_52b8a2e17371.slice/crio-33bea0fd49f502a85f8efa8d5c6ffccd81b68b31c773f26ffa1d0c16dbec49ac WatchSource:0}: Error finding container 33bea0fd49f502a85f8efa8d5c6ffccd81b68b31c773f26ffa1d0c16dbec49ac: Status 404 returned error can't find the container with id 33bea0fd49f502a85f8efa8d5c6ffccd81b68b31c773f26ffa1d0c16dbec49ac Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.283309 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bfe41d4c-9ca3-4772-b036-84b5573eb44e","Type":"ContainerStarted","Data":"81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7"} Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.285263 4978 generic.go:334] "Generic (PLEG): container finished" podID="bd220760-4e92-4cac-b4f0-8eb2c38f67be" containerID="8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f" exitCode=0 Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.285333 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" event={"ID":"bd220760-4e92-4cac-b4f0-8eb2c38f67be","Type":"ContainerDied","Data":"8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f"} Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.285386 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" event={"ID":"bd220760-4e92-4cac-b4f0-8eb2c38f67be","Type":"ContainerStarted","Data":"5aa5223f9fc92e20c81d58a8c59d9fbb0939e1e284234687230fff236d873bb6"} Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.287718 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f812baa4-8d36-4a69-b0c1-49fb714beced","Type":"ContainerStarted","Data":"84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779"} Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.301917 4978 generic.go:334] "Generic (PLEG): container finished" podID="df4bb775-9429-45bf-af46-93dcf3b4a82d" containerID="caed680d248b73867bcd7477d8eb3c3cdbb4b1d1ceff2297baed0b3c1eed6964" exitCode=0 Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.301991 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q4lmf" event={"ID":"df4bb775-9429-45bf-af46-93dcf3b4a82d","Type":"ContainerDied","Data":"caed680d248b73867bcd7477d8eb3c3cdbb4b1d1ceff2297baed0b3c1eed6964"} Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.308047 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e","Type":"ContainerStarted","Data":"0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0"} Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.308238 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" containerName="glance-log" containerID="cri-o://607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257" gracePeriod=30 Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.308488 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" containerName="glance-httpd" containerID="cri-o://0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0" gracePeriod=30 Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.340763 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" podUID="ee76c60c-2ef4-4ba7-b864-156f132d03d3" containerName="dnsmasq-dns" containerID="cri-o://dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e" gracePeriod=10 Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.341126 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5964547b68-crjwr" event={"ID":"9ee02b80-8075-47b3-8fe8-52b8a2e17371","Type":"ContainerStarted","Data":"0d1ec2a95bdf608b1bae7e76043322c9d54dc4357493235e1965da313e3493ee"} Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.341179 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5964547b68-crjwr" event={"ID":"9ee02b80-8075-47b3-8fe8-52b8a2e17371","Type":"ContainerStarted","Data":"af6d5e300eab1e02eed4638bd374ff091a24a5f26d9debc40f54880c4cadf6b8"} Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.341192 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5964547b68-crjwr" event={"ID":"9ee02b80-8075-47b3-8fe8-52b8a2e17371","Type":"ContainerStarted","Data":"33bea0fd49f502a85f8efa8d5c6ffccd81b68b31c773f26ffa1d0c16dbec49ac"} Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.341595 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.418524 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.418503003 podStartE2EDuration="5.418503003s" podCreationTimestamp="2026-02-25 07:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:22.381628283 +0000 UTC m=+1335.820884742" watchObservedRunningTime="2026-02-25 07:07:22.418503003 +0000 UTC m=+1335.857759462" Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.450630 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5964547b68-crjwr" podStartSLOduration=2.450611271 podStartE2EDuration="2.450611271s" podCreationTimestamp="2026-02-25 07:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:22.419458292 +0000 UTC m=+1335.858714761" watchObservedRunningTime="2026-02-25 07:07:22.450611271 +0000 UTC m=+1335.889867730" Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.839854 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-flx4x" Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.880303 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-scripts\") pod \"2ee32dc0-fe37-4be7-b87c-f605febf9692\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.880781 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-combined-ca-bundle\") pod \"2ee32dc0-fe37-4be7-b87c-f605febf9692\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.880872 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ee32dc0-fe37-4be7-b87c-f605febf9692-logs\") pod \"2ee32dc0-fe37-4be7-b87c-f605febf9692\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.880921 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-config-data\") pod \"2ee32dc0-fe37-4be7-b87c-f605febf9692\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.880955 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kbpn\" (UniqueName: \"kubernetes.io/projected/2ee32dc0-fe37-4be7-b87c-f605febf9692-kube-api-access-4kbpn\") pod \"2ee32dc0-fe37-4be7-b87c-f605febf9692\" (UID: \"2ee32dc0-fe37-4be7-b87c-f605febf9692\") " Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.882357 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ee32dc0-fe37-4be7-b87c-f605febf9692-logs" (OuterVolumeSpecName: "logs") pod "2ee32dc0-fe37-4be7-b87c-f605febf9692" (UID: "2ee32dc0-fe37-4be7-b87c-f605febf9692"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.882593 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2ee32dc0-fe37-4be7-b87c-f605febf9692-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.900704 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-scripts" (OuterVolumeSpecName: "scripts") pod "2ee32dc0-fe37-4be7-b87c-f605febf9692" (UID: "2ee32dc0-fe37-4be7-b87c-f605febf9692"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.901567 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ee32dc0-fe37-4be7-b87c-f605febf9692-kube-api-access-4kbpn" (OuterVolumeSpecName: "kube-api-access-4kbpn") pod "2ee32dc0-fe37-4be7-b87c-f605febf9692" (UID: "2ee32dc0-fe37-4be7-b87c-f605febf9692"). InnerVolumeSpecName "kube-api-access-4kbpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.958802 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-config-data" (OuterVolumeSpecName: "config-data") pod "2ee32dc0-fe37-4be7-b87c-f605febf9692" (UID: "2ee32dc0-fe37-4be7-b87c-f605febf9692"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:22 crc kubenswrapper[4978]: I0225 07:07:22.969125 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ee32dc0-fe37-4be7-b87c-f605febf9692" (UID: "2ee32dc0-fe37-4be7-b87c-f605febf9692"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.006242 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.006277 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.006291 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ee32dc0-fe37-4be7-b87c-f605febf9692-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.006300 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kbpn\" (UniqueName: \"kubernetes.io/projected/2ee32dc0-fe37-4be7-b87c-f605febf9692-kube-api-access-4kbpn\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.113957 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.216334 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-swift-storage-0\") pod \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.216792 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvwc6\" (UniqueName: \"kubernetes.io/projected/ee76c60c-2ef4-4ba7-b864-156f132d03d3-kube-api-access-fvwc6\") pod \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.216922 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-nb\") pod \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.216946 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-sb\") pod \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.217001 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-svc\") pod \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.217050 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-config\") pod \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\" (UID: \"ee76c60c-2ef4-4ba7-b864-156f132d03d3\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.223920 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee76c60c-2ef4-4ba7-b864-156f132d03d3-kube-api-access-fvwc6" (OuterVolumeSpecName: "kube-api-access-fvwc6") pod "ee76c60c-2ef4-4ba7-b864-156f132d03d3" (UID: "ee76c60c-2ef4-4ba7-b864-156f132d03d3"). InnerVolumeSpecName "kube-api-access-fvwc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.278002 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-config" (OuterVolumeSpecName: "config") pod "ee76c60c-2ef4-4ba7-b864-156f132d03d3" (UID: "ee76c60c-2ef4-4ba7-b864-156f132d03d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.320498 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvwc6\" (UniqueName: \"kubernetes.io/projected/ee76c60c-2ef4-4ba7-b864-156f132d03d3-kube-api-access-fvwc6\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.320523 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.343144 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ee76c60c-2ef4-4ba7-b864-156f132d03d3" (UID: "ee76c60c-2ef4-4ba7-b864-156f132d03d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.351636 4978 generic.go:334] "Generic (PLEG): container finished" podID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" containerID="0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0" exitCode=0 Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.351667 4978 generic.go:334] "Generic (PLEG): container finished" podID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" containerID="607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257" exitCode=143 Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.356621 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.364089 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee76c60c-2ef4-4ba7-b864-156f132d03d3" (UID: "ee76c60c-2ef4-4ba7-b864-156f132d03d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.366654 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e","Type":"ContainerDied","Data":"0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0"} Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.366681 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e","Type":"ContainerDied","Data":"607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257"} Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.366691 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e","Type":"ContainerDied","Data":"8c3804f11c6eee51db79a152498fadec5b9f4e6f8497e2d9a272cb0edc7e7195"} Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.366708 4978 scope.go:117] "RemoveContainer" containerID="0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.367895 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bfe41d4c-9ca3-4772-b036-84b5573eb44e","Type":"ContainerStarted","Data":"db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069"} Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.368007 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" containerName="glance-log" containerID="cri-o://81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7" gracePeriod=30 Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.368232 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" containerName="glance-httpd" containerID="cri-o://db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069" gracePeriod=30 Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.370820 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ee76c60c-2ef4-4ba7-b864-156f132d03d3" (UID: "ee76c60c-2ef4-4ba7-b864-156f132d03d3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.374158 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ee76c60c-2ef4-4ba7-b864-156f132d03d3" (UID: "ee76c60c-2ef4-4ba7-b864-156f132d03d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.390544 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" event={"ID":"bd220760-4e92-4cac-b4f0-8eb2c38f67be","Type":"ContainerStarted","Data":"5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b"} Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.390613 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.401960 4978 generic.go:334] "Generic (PLEG): container finished" podID="ee76c60c-2ef4-4ba7-b864-156f132d03d3" containerID="dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e" exitCode=0 Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.402023 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" event={"ID":"ee76c60c-2ef4-4ba7-b864-156f132d03d3","Type":"ContainerDied","Data":"dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e"} Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.402049 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" event={"ID":"ee76c60c-2ef4-4ba7-b864-156f132d03d3","Type":"ContainerDied","Data":"383c66d93eefee72ea72dccec191153daff0abd6495defbc18c476dd0b971e8f"} Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.402103 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86df48c85f-nmjcf" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.422168 4978 scope.go:117] "RemoveContainer" containerID="607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.430734 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-scripts\") pod \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.430890 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9z4b\" (UniqueName: \"kubernetes.io/projected/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-kube-api-access-h9z4b\") pod \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.431096 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-logs\") pod \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.432019 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-combined-ca-bundle\") pod \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.432125 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.432280 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-config-data\") pod \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.432380 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-httpd-run\") pod \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\" (UID: \"3bf7fe24-c0df-4823-bb46-bfd3c7ef942e\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.433842 4978 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.435468 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.435546 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.435616 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee76c60c-2ef4-4ba7-b864-156f132d03d3-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.436334 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-logs" (OuterVolumeSpecName: "logs") pod "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" (UID: "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.438005 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-flx4x" event={"ID":"2ee32dc0-fe37-4be7-b87c-f605febf9692","Type":"ContainerDied","Data":"cf67e3bb996efbb0b427d99418654a8ed5138586530681652bfc426d0022e8a3"} Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.438068 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf67e3bb996efbb0b427d99418654a8ed5138586530681652bfc426d0022e8a3" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.438286 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-flx4x" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.439574 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" (UID: "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.444061 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-kube-api-access-h9z4b" (OuterVolumeSpecName: "kube-api-access-h9z4b") pod "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" (UID: "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e"). InnerVolumeSpecName "kube-api-access-h9z4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.444156 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-scripts" (OuterVolumeSpecName: "scripts") pod "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" (UID: "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.452648 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" (UID: "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.465794 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" podStartSLOduration=3.465773398 podStartE2EDuration="3.465773398s" podCreationTimestamp="2026-02-25 07:07:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:23.430195937 +0000 UTC m=+1336.869452406" watchObservedRunningTime="2026-02-25 07:07:23.465773398 +0000 UTC m=+1336.905029857" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.466232 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.466227372 podStartE2EDuration="6.466227372s" podCreationTimestamp="2026-02-25 07:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:23.453747147 +0000 UTC m=+1336.893003606" watchObservedRunningTime="2026-02-25 07:07:23.466227372 +0000 UTC m=+1336.905483831" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.488438 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-776b767d76-hl99l"] Feb 25 07:07:23 crc kubenswrapper[4978]: E0225 07:07:23.488854 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" containerName="glance-httpd" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.488866 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" containerName="glance-httpd" Feb 25 07:07:23 crc kubenswrapper[4978]: E0225 07:07:23.488880 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee76c60c-2ef4-4ba7-b864-156f132d03d3" containerName="dnsmasq-dns" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.488887 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee76c60c-2ef4-4ba7-b864-156f132d03d3" containerName="dnsmasq-dns" Feb 25 07:07:23 crc kubenswrapper[4978]: E0225 07:07:23.488902 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ee32dc0-fe37-4be7-b87c-f605febf9692" containerName="placement-db-sync" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.488910 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ee32dc0-fe37-4be7-b87c-f605febf9692" containerName="placement-db-sync" Feb 25 07:07:23 crc kubenswrapper[4978]: E0225 07:07:23.488917 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee76c60c-2ef4-4ba7-b864-156f132d03d3" containerName="init" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.488923 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee76c60c-2ef4-4ba7-b864-156f132d03d3" containerName="init" Feb 25 07:07:23 crc kubenswrapper[4978]: E0225 07:07:23.488952 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" containerName="glance-log" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.488958 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" containerName="glance-log" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.489111 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" containerName="glance-log" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.489122 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" containerName="glance-httpd" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.489134 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ee32dc0-fe37-4be7-b87c-f605febf9692" containerName="placement-db-sync" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.489147 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee76c60c-2ef4-4ba7-b864-156f132d03d3" containerName="dnsmasq-dns" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.491390 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" (UID: "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.491887 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.496762 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.497101 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.497219 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-nzrxz" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.497333 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.496824 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.514847 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86df48c85f-nmjcf"] Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.528179 4978 scope.go:117] "RemoveContainer" containerID="0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.528393 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86df48c85f-nmjcf"] Feb 25 07:07:23 crc kubenswrapper[4978]: E0225 07:07:23.528699 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0\": container with ID starting with 0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0 not found: ID does not exist" containerID="0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.528748 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0"} err="failed to get container status \"0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0\": rpc error: code = NotFound desc = could not find container \"0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0\": container with ID starting with 0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0 not found: ID does not exist" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.528774 4978 scope.go:117] "RemoveContainer" containerID="607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257" Feb 25 07:07:23 crc kubenswrapper[4978]: E0225 07:07:23.531175 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257\": container with ID starting with 607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257 not found: ID does not exist" containerID="607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.531203 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257"} err="failed to get container status \"607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257\": rpc error: code = NotFound desc = could not find container \"607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257\": container with ID starting with 607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257 not found: ID does not exist" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.531219 4978 scope.go:117] "RemoveContainer" containerID="0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.533091 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0"} err="failed to get container status \"0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0\": rpc error: code = NotFound desc = could not find container \"0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0\": container with ID starting with 0744126843dadaa66c379fcb1ed8f36efb1bef32fd0242a0be037650118eeff0 not found: ID does not exist" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.533115 4978 scope.go:117] "RemoveContainer" containerID="607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.533325 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257"} err="failed to get container status \"607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257\": rpc error: code = NotFound desc = could not find container \"607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257\": container with ID starting with 607c320d80845b6d6f9269d5dde22f68a8d3294dc4b04e61765f5f75a296f257 not found: ID does not exist" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.533346 4978 scope.go:117] "RemoveContainer" containerID="dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.537607 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-public-tls-certs\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.537637 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-logs\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.537671 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-combined-ca-bundle\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.537702 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qrgh\" (UniqueName: \"kubernetes.io/projected/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-kube-api-access-4qrgh\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.537838 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-scripts\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.537980 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-config-data\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.538186 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-internal-tls-certs\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.538281 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.538299 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.538320 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.538332 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.538341 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.538378 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9z4b\" (UniqueName: \"kubernetes.io/projected/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-kube-api-access-h9z4b\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.540924 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-776b767d76-hl99l"] Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.556595 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.562303 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-config-data" (OuterVolumeSpecName: "config-data") pod "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" (UID: "3bf7fe24-c0df-4823-bb46-bfd3c7ef942e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.567648 4978 scope.go:117] "RemoveContainer" containerID="429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.619010 4978 scope.go:117] "RemoveContainer" containerID="dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e" Feb 25 07:07:23 crc kubenswrapper[4978]: E0225 07:07:23.619659 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e\": container with ID starting with dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e not found: ID does not exist" containerID="dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.619694 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e"} err="failed to get container status \"dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e\": rpc error: code = NotFound desc = could not find container \"dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e\": container with ID starting with dd7b235e8eea86a210d7bf6f14c9906e4579d3b53a62f588fe4665a3ce9bd23e not found: ID does not exist" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.619715 4978 scope.go:117] "RemoveContainer" containerID="429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae" Feb 25 07:07:23 crc kubenswrapper[4978]: E0225 07:07:23.620198 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae\": container with ID starting with 429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae not found: ID does not exist" containerID="429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.620222 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae"} err="failed to get container status \"429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae\": rpc error: code = NotFound desc = could not find container \"429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae\": container with ID starting with 429ac3416087585d2aeb7987027d2fe0c2d9be1ea8d8c3dd2e4d9dbd9c0903ae not found: ID does not exist" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.640723 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-internal-tls-certs\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.641835 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-public-tls-certs\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.641866 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-logs\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.641893 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-combined-ca-bundle\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.641935 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qrgh\" (UniqueName: \"kubernetes.io/projected/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-kube-api-access-4qrgh\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.642006 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-scripts\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.642047 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-config-data\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.642233 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.642248 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.644203 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-logs\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.644764 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-internal-tls-certs\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.645407 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-combined-ca-bundle\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.645856 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-config-data\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.647326 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-scripts\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.648472 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-public-tls-certs\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.664998 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qrgh\" (UniqueName: \"kubernetes.io/projected/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-kube-api-access-4qrgh\") pod \"placement-776b767d76-hl99l\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.830715 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.831910 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.955805 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-combined-ca-bundle\") pod \"df4bb775-9429-45bf-af46-93dcf3b4a82d\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.955869 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-credential-keys\") pod \"df4bb775-9429-45bf-af46-93dcf3b4a82d\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.955896 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-scripts\") pod \"df4bb775-9429-45bf-af46-93dcf3b4a82d\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.955925 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsmtl\" (UniqueName: \"kubernetes.io/projected/df4bb775-9429-45bf-af46-93dcf3b4a82d-kube-api-access-bsmtl\") pod \"df4bb775-9429-45bf-af46-93dcf3b4a82d\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.956110 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-fernet-keys\") pod \"df4bb775-9429-45bf-af46-93dcf3b4a82d\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.956137 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-config-data\") pod \"df4bb775-9429-45bf-af46-93dcf3b4a82d\" (UID: \"df4bb775-9429-45bf-af46-93dcf3b4a82d\") " Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.962056 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-scripts" (OuterVolumeSpecName: "scripts") pod "df4bb775-9429-45bf-af46-93dcf3b4a82d" (UID: "df4bb775-9429-45bf-af46-93dcf3b4a82d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.964319 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df4bb775-9429-45bf-af46-93dcf3b4a82d-kube-api-access-bsmtl" (OuterVolumeSpecName: "kube-api-access-bsmtl") pod "df4bb775-9429-45bf-af46-93dcf3b4a82d" (UID: "df4bb775-9429-45bf-af46-93dcf3b4a82d"). InnerVolumeSpecName "kube-api-access-bsmtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.967577 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "df4bb775-9429-45bf-af46-93dcf3b4a82d" (UID: "df4bb775-9429-45bf-af46-93dcf3b4a82d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.973159 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "df4bb775-9429-45bf-af46-93dcf3b4a82d" (UID: "df4bb775-9429-45bf-af46-93dcf3b4a82d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:23 crc kubenswrapper[4978]: I0225 07:07:23.993096 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-config-data" (OuterVolumeSpecName: "config-data") pod "df4bb775-9429-45bf-af46-93dcf3b4a82d" (UID: "df4bb775-9429-45bf-af46-93dcf3b4a82d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.011716 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df4bb775-9429-45bf-af46-93dcf3b4a82d" (UID: "df4bb775-9429-45bf-af46-93dcf3b4a82d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.057666 4978 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.057690 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.057701 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.057710 4978 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.057720 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df4bb775-9429-45bf-af46-93dcf3b4a82d-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.057729 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsmtl\" (UniqueName: \"kubernetes.io/projected/df4bb775-9429-45bf-af46-93dcf3b4a82d-kube-api-access-bsmtl\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.095655 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5987f7d87c-fxhz5"] Feb 25 07:07:24 crc kubenswrapper[4978]: E0225 07:07:24.096064 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df4bb775-9429-45bf-af46-93dcf3b4a82d" containerName="keystone-bootstrap" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.096076 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="df4bb775-9429-45bf-af46-93dcf3b4a82d" containerName="keystone-bootstrap" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.096253 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="df4bb775-9429-45bf-af46-93dcf3b4a82d" containerName="keystone-bootstrap" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.097126 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.102787 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.103006 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.111428 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5987f7d87c-fxhz5"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.159613 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-internal-tls-certs\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.159666 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-httpd-config\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.159742 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-ovndb-tls-certs\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.159763 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-combined-ca-bundle\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.159801 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-public-tls-certs\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.159837 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-config\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.159854 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gsf8\" (UniqueName: \"kubernetes.io/projected/5ad94c00-0a28-4933-ad5c-27966cc8b437-kube-api-access-4gsf8\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.259930 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.261035 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-ovndb-tls-certs\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.261065 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-combined-ca-bundle\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.261085 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-public-tls-certs\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.261119 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-config\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.261137 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gsf8\" (UniqueName: \"kubernetes.io/projected/5ad94c00-0a28-4933-ad5c-27966cc8b437-kube-api-access-4gsf8\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.261168 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-internal-tls-certs\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.261199 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-httpd-config\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.269275 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-config\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.269700 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-internal-tls-certs\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.270724 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-public-tls-certs\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.271586 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-ovndb-tls-certs\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.272929 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-httpd-config\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.274104 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-combined-ca-bundle\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.294921 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gsf8\" (UniqueName: \"kubernetes.io/projected/5ad94c00-0a28-4933-ad5c-27966cc8b437-kube-api-access-4gsf8\") pod \"neutron-5987f7d87c-fxhz5\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.361763 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-scripts\") pod \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.361814 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttjn8\" (UniqueName: \"kubernetes.io/projected/bfe41d4c-9ca3-4772-b036-84b5573eb44e-kube-api-access-ttjn8\") pod \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.361839 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-combined-ca-bundle\") pod \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.361868 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-httpd-run\") pod \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.361938 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.362077 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-logs\") pod \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.362120 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-config-data\") pod \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\" (UID: \"bfe41d4c-9ca3-4772-b036-84b5573eb44e\") " Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.365982 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-logs" (OuterVolumeSpecName: "logs") pod "bfe41d4c-9ca3-4772-b036-84b5573eb44e" (UID: "bfe41d4c-9ca3-4772-b036-84b5573eb44e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.367624 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bfe41d4c-9ca3-4772-b036-84b5573eb44e" (UID: "bfe41d4c-9ca3-4772-b036-84b5573eb44e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.369604 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-scripts" (OuterVolumeSpecName: "scripts") pod "bfe41d4c-9ca3-4772-b036-84b5573eb44e" (UID: "bfe41d4c-9ca3-4772-b036-84b5573eb44e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.371512 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe41d4c-9ca3-4772-b036-84b5573eb44e-kube-api-access-ttjn8" (OuterVolumeSpecName: "kube-api-access-ttjn8") pod "bfe41d4c-9ca3-4772-b036-84b5573eb44e" (UID: "bfe41d4c-9ca3-4772-b036-84b5573eb44e"). InnerVolumeSpecName "kube-api-access-ttjn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.385720 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "bfe41d4c-9ca3-4772-b036-84b5573eb44e" (UID: "bfe41d4c-9ca3-4772-b036-84b5573eb44e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.415905 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfe41d4c-9ca3-4772-b036-84b5573eb44e" (UID: "bfe41d4c-9ca3-4772-b036-84b5573eb44e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.427911 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7dd4fddc6c-nv49c"] Feb 25 07:07:24 crc kubenswrapper[4978]: E0225 07:07:24.428790 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" containerName="glance-log" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.428810 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" containerName="glance-log" Feb 25 07:07:24 crc kubenswrapper[4978]: E0225 07:07:24.428840 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" containerName="glance-httpd" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.428849 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" containerName="glance-httpd" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.429290 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" containerName="glance-httpd" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.429353 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" containerName="glance-log" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.430315 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.432533 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.436096 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.437676 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.441188 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7dd4fddc6c-nv49c"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.469325 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-config-data\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.469632 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-fernet-keys\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.478815 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-credential-keys\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.478972 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-combined-ca-bundle\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479041 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-scripts\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.476301 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-config-data" (OuterVolumeSpecName: "config-data") pod "bfe41d4c-9ca3-4772-b036-84b5573eb44e" (UID: "bfe41d4c-9ca3-4772-b036-84b5573eb44e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479127 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzhn8\" (UniqueName: \"kubernetes.io/projected/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-kube-api-access-tzhn8\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479345 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-internal-tls-certs\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479409 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-public-tls-certs\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479876 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479898 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479909 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479918 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttjn8\" (UniqueName: \"kubernetes.io/projected/bfe41d4c-9ca3-4772-b036-84b5573eb44e-kube-api-access-ttjn8\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479928 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe41d4c-9ca3-4772-b036-84b5573eb44e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479936 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfe41d4c-9ca3-4772-b036-84b5573eb44e-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.479954 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.480553 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q4lmf" event={"ID":"df4bb775-9429-45bf-af46-93dcf3b4a82d","Type":"ContainerDied","Data":"191b687dd267b5da9db654307f5907fc2fd7ff78005e8946476615d1b3d994a2"} Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.480671 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="191b687dd267b5da9db654307f5907fc2fd7ff78005e8946476615d1b3d994a2" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.480809 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q4lmf" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.487818 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.505201 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.507220 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-776b767d76-hl99l"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.514801 4978 generic.go:334] "Generic (PLEG): container finished" podID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" containerID="db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069" exitCode=0 Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.514834 4978 generic.go:334] "Generic (PLEG): container finished" podID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" containerID="81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7" exitCode=143 Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.514910 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bfe41d4c-9ca3-4772-b036-84b5573eb44e","Type":"ContainerDied","Data":"db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069"} Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.514952 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bfe41d4c-9ca3-4772-b036-84b5573eb44e","Type":"ContainerDied","Data":"81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7"} Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.514968 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bfe41d4c-9ca3-4772-b036-84b5573eb44e","Type":"ContainerDied","Data":"cdf5b5a9a9e7c2989b301194e51f534b7afb711852c8a5b3ec187c63760b13c7"} Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.514998 4978 scope.go:117] "RemoveContainer" containerID="db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.515216 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.540253 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.556431 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.572089 4978 scope.go:117] "RemoveContainer" containerID="81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.581218 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-credential-keys\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.581296 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-combined-ca-bundle\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.581316 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-scripts\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.581341 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzhn8\" (UniqueName: \"kubernetes.io/projected/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-kube-api-access-tzhn8\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.581485 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-internal-tls-certs\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.581517 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-public-tls-certs\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.581545 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-config-data\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.581594 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-fernet-keys\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.581642 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.586845 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-fernet-keys\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.588706 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.590244 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.593238 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.595350 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.595502 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.595650 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mrzl7" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.600841 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-credential-keys\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.602399 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-config-data\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.604890 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.607149 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-internal-tls-certs\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.616825 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-combined-ca-bundle\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.617494 4978 scope.go:117] "RemoveContainer" containerID="db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069" Feb 25 07:07:24 crc kubenswrapper[4978]: E0225 07:07:24.617810 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069\": container with ID starting with db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069 not found: ID does not exist" containerID="db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.617851 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069"} err="failed to get container status \"db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069\": rpc error: code = NotFound desc = could not find container \"db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069\": container with ID starting with db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069 not found: ID does not exist" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.617869 4978 scope.go:117] "RemoveContainer" containerID="81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7" Feb 25 07:07:24 crc kubenswrapper[4978]: E0225 07:07:24.618060 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7\": container with ID starting with 81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7 not found: ID does not exist" containerID="81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.618107 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7"} err="failed to get container status \"81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7\": rpc error: code = NotFound desc = could not find container \"81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7\": container with ID starting with 81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7 not found: ID does not exist" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.618120 4978 scope.go:117] "RemoveContainer" containerID="db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.618320 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069"} err="failed to get container status \"db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069\": rpc error: code = NotFound desc = could not find container \"db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069\": container with ID starting with db4d12095f1197fd2e959574895398e37df607d1d785df0a7891232421aac069 not found: ID does not exist" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.618333 4978 scope.go:117] "RemoveContainer" containerID="81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.618744 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7"} err="failed to get container status \"81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7\": rpc error: code = NotFound desc = could not find container \"81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7\": container with ID starting with 81c70c86e4d0eda9c66256a110cb78624d5cad4f222179a0c6f196168c20ddd7 not found: ID does not exist" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.619911 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.620328 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzhn8\" (UniqueName: \"kubernetes.io/projected/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-kube-api-access-tzhn8\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.641464 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.641525 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.642944 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.643024 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.644096 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-public-tls-certs\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.645578 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.650634 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-scripts\") pod \"keystone-7dd4fddc6c-nv49c\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.650921 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.761323 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784419 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784458 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djv55\" (UniqueName: \"kubernetes.io/projected/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-kube-api-access-djv55\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784494 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-scripts\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784515 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784541 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784557 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784597 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784611 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-logs\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784632 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-logs\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784657 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784678 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784694 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784708 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-config-data\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784727 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784766 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.784783 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q45fb\" (UniqueName: \"kubernetes.io/projected/667d3386-e724-42f3-940b-82af7780255b-kube-api-access-q45fb\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.886631 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-scripts\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.888135 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.888175 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.888194 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.888449 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889670 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889696 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-logs\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889730 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-logs\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889764 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889791 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889809 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889828 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-config-data\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889848 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889905 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889926 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q45fb\" (UniqueName: \"kubernetes.io/projected/667d3386-e724-42f3-940b-82af7780255b-kube-api-access-q45fb\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889974 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.889995 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djv55\" (UniqueName: \"kubernetes.io/projected/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-kube-api-access-djv55\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.890588 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.890614 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.890854 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.891324 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-logs\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.891839 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.893509 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-logs\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.908098 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.908334 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.909253 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.923018 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.923866 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-config-data\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.924341 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q45fb\" (UniqueName: \"kubernetes.io/projected/667d3386-e724-42f3-940b-82af7780255b-kube-api-access-q45fb\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.933742 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-scripts\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.945685 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djv55\" (UniqueName: \"kubernetes.io/projected/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-kube-api-access-djv55\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.954600 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:24 crc kubenswrapper[4978]: I0225 07:07:24.960732 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.001868 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.020542 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " pod="openstack/glance-default-external-api-0" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.188962 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5987f7d87c-fxhz5"] Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.275975 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.323292 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7dd4fddc6c-nv49c"] Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.341620 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bf7fe24-c0df-4823-bb46-bfd3c7ef942e" path="/var/lib/kubelet/pods/3bf7fe24-c0df-4823-bb46-bfd3c7ef942e/volumes" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.342398 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfe41d4c-9ca3-4772-b036-84b5573eb44e" path="/var/lib/kubelet/pods/bfe41d4c-9ca3-4772-b036-84b5573eb44e/volumes" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.342950 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee76c60c-2ef4-4ba7-b864-156f132d03d3" path="/var/lib/kubelet/pods/ee76c60c-2ef4-4ba7-b864-156f132d03d3/volumes" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.605633 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-776b767d76-hl99l" event={"ID":"be86cdbd-c5d3-465a-96fb-426b76a8b6bb","Type":"ContainerStarted","Data":"3738d8be721c087bd8f44c86484a7539eb563b4afdd76d2543dcc52eb0e30860"} Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.605925 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-776b767d76-hl99l" event={"ID":"be86cdbd-c5d3-465a-96fb-426b76a8b6bb","Type":"ContainerStarted","Data":"1fb7bf2f0b0317ab071ae00aecb34bc39e311241db22967f13d25e3a933d30e5"} Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.605938 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-776b767d76-hl99l" event={"ID":"be86cdbd-c5d3-465a-96fb-426b76a8b6bb","Type":"ContainerStarted","Data":"566e36ddc4b6b30b8144c94b686a183e417c2e1130150217e9c40be2cd8047d8"} Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.605988 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.606006 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.610908 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5987f7d87c-fxhz5" event={"ID":"5ad94c00-0a28-4933-ad5c-27966cc8b437","Type":"ContainerStarted","Data":"6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a"} Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.610949 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5987f7d87c-fxhz5" event={"ID":"5ad94c00-0a28-4933-ad5c-27966cc8b437","Type":"ContainerStarted","Data":"b86d30a69e41ee0a7893929787a3b26e8f0ad417dc776d5a7c7325639399e4c0"} Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.612326 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7dd4fddc6c-nv49c" event={"ID":"3ee398cf-5bd3-4905-80f2-b66d67c57f3d","Type":"ContainerStarted","Data":"b6bce4a4fde76c203ae95aa89120619243abd412f96cdabdc6c8cd947deefeee"} Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.613218 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.640106 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-776b767d76-hl99l" podStartSLOduration=2.640085932 podStartE2EDuration="2.640085932s" podCreationTimestamp="2026-02-25 07:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:25.628686309 +0000 UTC m=+1339.067942788" watchObservedRunningTime="2026-02-25 07:07:25.640085932 +0000 UTC m=+1339.079342391" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.692806 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7dd4fddc6c-nv49c" podStartSLOduration=1.692761779 podStartE2EDuration="1.692761779s" podCreationTimestamp="2026-02-25 07:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:25.665132407 +0000 UTC m=+1339.104388876" watchObservedRunningTime="2026-02-25 07:07:25.692761779 +0000 UTC m=+1339.132018238" Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.711473 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:07:25 crc kubenswrapper[4978]: I0225 07:07:25.899311 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:07:26 crc kubenswrapper[4978]: I0225 07:07:26.628799 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"09b73acd-bf19-49f5-b5e0-a60e180e9c1e","Type":"ContainerStarted","Data":"f68c2a7b016bdf0a6652e9324843ca4472b7b54fb5f39c860f5ab165234032df"} Feb 25 07:07:26 crc kubenswrapper[4978]: I0225 07:07:26.629192 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"09b73acd-bf19-49f5-b5e0-a60e180e9c1e","Type":"ContainerStarted","Data":"6c97b54e5d7163ddfe49a74471cf0b36387569b3312a3e69ff9b0a4c8c701862"} Feb 25 07:07:26 crc kubenswrapper[4978]: I0225 07:07:26.632766 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"667d3386-e724-42f3-940b-82af7780255b","Type":"ContainerStarted","Data":"32a5d429a99cd8a6f7e1deeb90aaac504efbba685415e1cd920ca02ab313cf16"} Feb 25 07:07:26 crc kubenswrapper[4978]: I0225 07:07:26.632789 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"667d3386-e724-42f3-940b-82af7780255b","Type":"ContainerStarted","Data":"442edee2912d4eb077b7d9115fd12860b305dec6c483690d2c36dde5dc5ffcc1"} Feb 25 07:07:26 crc kubenswrapper[4978]: I0225 07:07:26.636847 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5987f7d87c-fxhz5" event={"ID":"5ad94c00-0a28-4933-ad5c-27966cc8b437","Type":"ContainerStarted","Data":"58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db"} Feb 25 07:07:26 crc kubenswrapper[4978]: I0225 07:07:26.638470 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:26 crc kubenswrapper[4978]: I0225 07:07:26.643484 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7dd4fddc6c-nv49c" event={"ID":"3ee398cf-5bd3-4905-80f2-b66d67c57f3d","Type":"ContainerStarted","Data":"3b71dd25e21037082b24f81353af12fb7fe11432d9ea70560778f8494ba5ae98"} Feb 25 07:07:26 crc kubenswrapper[4978]: I0225 07:07:26.666419 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5987f7d87c-fxhz5" podStartSLOduration=2.666399966 podStartE2EDuration="2.666399966s" podCreationTimestamp="2026-02-25 07:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:26.65524901 +0000 UTC m=+1340.094505479" watchObservedRunningTime="2026-02-25 07:07:26.666399966 +0000 UTC m=+1340.105656425" Feb 25 07:07:27 crc kubenswrapper[4978]: I0225 07:07:27.663715 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"09b73acd-bf19-49f5-b5e0-a60e180e9c1e","Type":"ContainerStarted","Data":"b8960ac3480ae1d88146d962f2420be146e18df6db23c90812cb66575bdf4679"} Feb 25 07:07:27 crc kubenswrapper[4978]: I0225 07:07:27.667318 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"667d3386-e724-42f3-940b-82af7780255b","Type":"ContainerStarted","Data":"ef78e19bd4d46fac7902200f55b0093cbedffb4f12c8d2a96d94765f2b3f1e48"} Feb 25 07:07:27 crc kubenswrapper[4978]: I0225 07:07:27.688821 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.688804052 podStartE2EDuration="3.688804052s" podCreationTimestamp="2026-02-25 07:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:27.68740827 +0000 UTC m=+1341.126664769" watchObservedRunningTime="2026-02-25 07:07:27.688804052 +0000 UTC m=+1341.128060521" Feb 25 07:07:27 crc kubenswrapper[4978]: I0225 07:07:27.723441 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.723424775 podStartE2EDuration="3.723424775s" podCreationTimestamp="2026-02-25 07:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:27.714137016 +0000 UTC m=+1341.153393515" watchObservedRunningTime="2026-02-25 07:07:27.723424775 +0000 UTC m=+1341.162681234" Feb 25 07:07:29 crc kubenswrapper[4978]: I0225 07:07:29.685966 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xpwds" event={"ID":"4d7807bf-3cea-45ed-b534-29b26ebe2c12","Type":"ContainerStarted","Data":"04bf0ba3b369628fea98361e5da48ba3bc1bc5dde35d641c7aaf36fd44c0ff1c"} Feb 25 07:07:29 crc kubenswrapper[4978]: I0225 07:07:29.688671 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f812baa4-8d36-4a69-b0c1-49fb714beced","Type":"ContainerStarted","Data":"460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099"} Feb 25 07:07:29 crc kubenswrapper[4978]: I0225 07:07:29.721266 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-xpwds" podStartSLOduration=2.127591175 podStartE2EDuration="35.721244072s" podCreationTimestamp="2026-02-25 07:06:54 +0000 UTC" firstStartedPulling="2026-02-25 07:06:55.551777071 +0000 UTC m=+1308.991033520" lastFinishedPulling="2026-02-25 07:07:29.145429958 +0000 UTC m=+1342.584686417" observedRunningTime="2026-02-25 07:07:29.713265492 +0000 UTC m=+1343.152521951" watchObservedRunningTime="2026-02-25 07:07:29.721244072 +0000 UTC m=+1343.160500541" Feb 25 07:07:30 crc kubenswrapper[4978]: I0225 07:07:30.868637 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:30 crc kubenswrapper[4978]: I0225 07:07:30.945646 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd5bd7985-l4b4x"] Feb 25 07:07:30 crc kubenswrapper[4978]: I0225 07:07:30.945999 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" podUID="1eb078ad-fc4f-479f-b330-69f8b2fe62e8" containerName="dnsmasq-dns" containerID="cri-o://33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970" gracePeriod=10 Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.516242 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.626956 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-sb\") pod \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.627040 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-svc\") pod \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.627061 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs7j2\" (UniqueName: \"kubernetes.io/projected/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-kube-api-access-hs7j2\") pod \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.627261 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-config\") pod \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.627295 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-nb\") pod \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.627337 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-swift-storage-0\") pod \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\" (UID: \"1eb078ad-fc4f-479f-b330-69f8b2fe62e8\") " Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.647719 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-kube-api-access-hs7j2" (OuterVolumeSpecName: "kube-api-access-hs7j2") pod "1eb078ad-fc4f-479f-b330-69f8b2fe62e8" (UID: "1eb078ad-fc4f-479f-b330-69f8b2fe62e8"). InnerVolumeSpecName "kube-api-access-hs7j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.725715 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-config" (OuterVolumeSpecName: "config") pod "1eb078ad-fc4f-479f-b330-69f8b2fe62e8" (UID: "1eb078ad-fc4f-479f-b330-69f8b2fe62e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.728609 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tc5ks" event={"ID":"f18ee3ff-994c-46c4-8f0c-76b1ee45da50","Type":"ContainerStarted","Data":"883a527a9d84208558ca0007c0a8c77189f06bf0927bb703145ef20bcf5a9c9b"} Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.729710 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs7j2\" (UniqueName: \"kubernetes.io/projected/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-kube-api-access-hs7j2\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.729847 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.731007 4978 generic.go:334] "Generic (PLEG): container finished" podID="1eb078ad-fc4f-479f-b330-69f8b2fe62e8" containerID="33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970" exitCode=0 Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.731031 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" event={"ID":"1eb078ad-fc4f-479f-b330-69f8b2fe62e8","Type":"ContainerDied","Data":"33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970"} Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.731049 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" event={"ID":"1eb078ad-fc4f-479f-b330-69f8b2fe62e8","Type":"ContainerDied","Data":"9b33afadec79918c6788b4c2c7e34bd710efb3369f49347121c909ddf7562a7f"} Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.731066 4978 scope.go:117] "RemoveContainer" containerID="33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.731157 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd5bd7985-l4b4x" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.754353 4978 scope.go:117] "RemoveContainer" containerID="9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.777061 4978 scope.go:117] "RemoveContainer" containerID="33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970" Feb 25 07:07:31 crc kubenswrapper[4978]: E0225 07:07:31.777525 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970\": container with ID starting with 33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970 not found: ID does not exist" containerID="33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.777555 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970"} err="failed to get container status \"33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970\": rpc error: code = NotFound desc = could not find container \"33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970\": container with ID starting with 33ed2aac31a5a243fd7e12c22c45114271607b169b61f9147139e573b2c80970 not found: ID does not exist" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.777576 4978 scope.go:117] "RemoveContainer" containerID="9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.778818 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-tc5ks" podStartSLOduration=2.175124237 podStartE2EDuration="37.778797149s" podCreationTimestamp="2026-02-25 07:06:54 +0000 UTC" firstStartedPulling="2026-02-25 07:06:55.149184154 +0000 UTC m=+1308.588440613" lastFinishedPulling="2026-02-25 07:07:30.752857026 +0000 UTC m=+1344.192113525" observedRunningTime="2026-02-25 07:07:31.775271602 +0000 UTC m=+1345.214528071" watchObservedRunningTime="2026-02-25 07:07:31.778797149 +0000 UTC m=+1345.218053608" Feb 25 07:07:31 crc kubenswrapper[4978]: E0225 07:07:31.780208 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd\": container with ID starting with 9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd not found: ID does not exist" containerID="9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.780239 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd"} err="failed to get container status \"9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd\": rpc error: code = NotFound desc = could not find container \"9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd\": container with ID starting with 9ac56b2195d77be27c95e2340be83b447640210c9a915fd2fe4d39f42d0c87bd not found: ID does not exist" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.781802 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1eb078ad-fc4f-479f-b330-69f8b2fe62e8" (UID: "1eb078ad-fc4f-479f-b330-69f8b2fe62e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.795294 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1eb078ad-fc4f-479f-b330-69f8b2fe62e8" (UID: "1eb078ad-fc4f-479f-b330-69f8b2fe62e8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.796558 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1eb078ad-fc4f-479f-b330-69f8b2fe62e8" (UID: "1eb078ad-fc4f-479f-b330-69f8b2fe62e8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.813672 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1eb078ad-fc4f-479f-b330-69f8b2fe62e8" (UID: "1eb078ad-fc4f-479f-b330-69f8b2fe62e8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.831861 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.831896 4978 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.832491 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:31 crc kubenswrapper[4978]: I0225 07:07:31.833520 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1eb078ad-fc4f-479f-b330-69f8b2fe62e8-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:32 crc kubenswrapper[4978]: I0225 07:07:32.073446 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd5bd7985-l4b4x"] Feb 25 07:07:32 crc kubenswrapper[4978]: I0225 07:07:32.083053 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bd5bd7985-l4b4x"] Feb 25 07:07:32 crc kubenswrapper[4978]: I0225 07:07:32.753092 4978 generic.go:334] "Generic (PLEG): container finished" podID="4d7807bf-3cea-45ed-b534-29b26ebe2c12" containerID="04bf0ba3b369628fea98361e5da48ba3bc1bc5dde35d641c7aaf36fd44c0ff1c" exitCode=0 Feb 25 07:07:32 crc kubenswrapper[4978]: I0225 07:07:32.753467 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xpwds" event={"ID":"4d7807bf-3cea-45ed-b534-29b26ebe2c12","Type":"ContainerDied","Data":"04bf0ba3b369628fea98361e5da48ba3bc1bc5dde35d641c7aaf36fd44c0ff1c"} Feb 25 07:07:33 crc kubenswrapper[4978]: I0225 07:07:33.339159 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eb078ad-fc4f-479f-b330-69f8b2fe62e8" path="/var/lib/kubelet/pods/1eb078ad-fc4f-479f-b330-69f8b2fe62e8/volumes" Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.154470 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xpwds" Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.182894 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhl77\" (UniqueName: \"kubernetes.io/projected/4d7807bf-3cea-45ed-b534-29b26ebe2c12-kube-api-access-bhl77\") pod \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.182961 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-combined-ca-bundle\") pod \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.183002 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-db-sync-config-data\") pod \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\" (UID: \"4d7807bf-3cea-45ed-b534-29b26ebe2c12\") " Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.224560 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4d7807bf-3cea-45ed-b534-29b26ebe2c12" (UID: "4d7807bf-3cea-45ed-b534-29b26ebe2c12"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.224638 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d7807bf-3cea-45ed-b534-29b26ebe2c12-kube-api-access-bhl77" (OuterVolumeSpecName: "kube-api-access-bhl77") pod "4d7807bf-3cea-45ed-b534-29b26ebe2c12" (UID: "4d7807bf-3cea-45ed-b534-29b26ebe2c12"). InnerVolumeSpecName "kube-api-access-bhl77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.230598 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d7807bf-3cea-45ed-b534-29b26ebe2c12" (UID: "4d7807bf-3cea-45ed-b534-29b26ebe2c12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.289423 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhl77\" (UniqueName: \"kubernetes.io/projected/4d7807bf-3cea-45ed-b534-29b26ebe2c12-kube-api-access-bhl77\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.289489 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.289499 4978 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d7807bf-3cea-45ed-b534-29b26ebe2c12-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.774537 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xpwds" event={"ID":"4d7807bf-3cea-45ed-b534-29b26ebe2c12","Type":"ContainerDied","Data":"e767316480b28229f447bef48ca6576c4bc55b93b5f0975506f046f01c71b538"} Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.774623 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xpwds" Feb 25 07:07:34 crc kubenswrapper[4978]: I0225 07:07:34.781311 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e767316480b28229f447bef48ca6576c4bc55b93b5f0975506f046f01c71b538" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.002901 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.002946 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.045178 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.094233 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7696768c55-mnqh9"] Feb 25 07:07:35 crc kubenswrapper[4978]: E0225 07:07:35.094569 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb078ad-fc4f-479f-b330-69f8b2fe62e8" containerName="init" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.094585 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb078ad-fc4f-479f-b330-69f8b2fe62e8" containerName="init" Feb 25 07:07:35 crc kubenswrapper[4978]: E0225 07:07:35.094599 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb078ad-fc4f-479f-b330-69f8b2fe62e8" containerName="dnsmasq-dns" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.094606 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb078ad-fc4f-479f-b330-69f8b2fe62e8" containerName="dnsmasq-dns" Feb 25 07:07:35 crc kubenswrapper[4978]: E0225 07:07:35.094624 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7807bf-3cea-45ed-b534-29b26ebe2c12" containerName="barbican-db-sync" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.094630 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7807bf-3cea-45ed-b534-29b26ebe2c12" containerName="barbican-db-sync" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.094812 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb078ad-fc4f-479f-b330-69f8b2fe62e8" containerName="dnsmasq-dns" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.094837 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7807bf-3cea-45ed-b534-29b26ebe2c12" containerName="barbican-db-sync" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.098897 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.102221 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-zx5c6" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.103690 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.112090 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.113385 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.119128 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7696768c55-mnqh9"] Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.158028 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5b4d65dff6-x8qrr"] Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.160943 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.162556 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.182877 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5b4d65dff6-x8qrr"] Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.208289 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data-custom\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.208420 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.208445 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-combined-ca-bundle\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.208471 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltqc5\" (UniqueName: \"kubernetes.io/projected/a762aa77-9eb1-42eb-98e5-016cc3acac10-kube-api-access-ltqc5\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.208498 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a762aa77-9eb1-42eb-98e5-016cc3acac10-logs\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.231604 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-99b7f5fd5-gp9zm"] Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.234276 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.247540 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-99b7f5fd5-gp9zm"] Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.277193 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.277508 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.310635 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-combined-ca-bundle\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.310949 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data-custom\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.311015 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.311216 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-combined-ca-bundle\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.311235 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltqc5\" (UniqueName: \"kubernetes.io/projected/a762aa77-9eb1-42eb-98e5-016cc3acac10-kube-api-access-ltqc5\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.311267 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a762aa77-9eb1-42eb-98e5-016cc3acac10-logs\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.311293 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.311325 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6xjf\" (UniqueName: \"kubernetes.io/projected/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-kube-api-access-v6xjf\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.311381 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-logs\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.311403 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data-custom\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.312703 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a762aa77-9eb1-42eb-98e5-016cc3acac10-logs\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.320733 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.320923 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data-custom\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.328671 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-combined-ca-bundle\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.347161 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.351902 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltqc5\" (UniqueName: \"kubernetes.io/projected/a762aa77-9eb1-42eb-98e5-016cc3acac10-kube-api-access-ltqc5\") pod \"barbican-worker-7696768c55-mnqh9\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.356398 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413004 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzk9g\" (UniqueName: \"kubernetes.io/projected/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-kube-api-access-pzk9g\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413076 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-nb\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413113 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413156 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-swift-storage-0\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413190 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6xjf\" (UniqueName: \"kubernetes.io/projected/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-kube-api-access-v6xjf\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413210 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-sb\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413246 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-config\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413272 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-svc\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413311 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-logs\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413362 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-combined-ca-bundle\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.413396 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data-custom\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.414220 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-logs\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.419773 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.424397 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-combined-ca-bundle\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.427322 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data-custom\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.428404 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.450405 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6xjf\" (UniqueName: \"kubernetes.io/projected/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-kube-api-access-v6xjf\") pod \"barbican-keystone-listener-5b4d65dff6-x8qrr\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.468331 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5dc9c48444-drpjk"] Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.469771 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.475883 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.481442 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5dc9c48444-drpjk"] Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.493735 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.515214 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzk9g\" (UniqueName: \"kubernetes.io/projected/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-kube-api-access-pzk9g\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.515484 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-nb\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.515575 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-swift-storage-0\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.515656 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-sb\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.515735 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-config\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.515812 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-svc\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.516803 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-svc\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.517665 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-nb\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.518388 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-swift-storage-0\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.521541 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-config\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.530507 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-sb\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.541230 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzk9g\" (UniqueName: \"kubernetes.io/projected/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-kube-api-access-pzk9g\") pod \"dnsmasq-dns-99b7f5fd5-gp9zm\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.570698 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.617789 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-logs\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.617838 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.617891 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data-custom\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.617920 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-combined-ca-bundle\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.618078 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7hsv\" (UniqueName: \"kubernetes.io/projected/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-kube-api-access-m7hsv\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.719814 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7hsv\" (UniqueName: \"kubernetes.io/projected/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-kube-api-access-m7hsv\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.720096 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-logs\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.720127 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.720151 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data-custom\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.720173 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-combined-ca-bundle\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.721201 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-logs\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.729071 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data-custom\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.729525 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.731607 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-combined-ca-bundle\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.734881 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7hsv\" (UniqueName: \"kubernetes.io/projected/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-kube-api-access-m7hsv\") pod \"barbican-api-5dc9c48444-drpjk\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.782893 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.782937 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.782947 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.782956 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:35 crc kubenswrapper[4978]: I0225 07:07:35.811276 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:36 crc kubenswrapper[4978]: I0225 07:07:36.791357 4978 generic.go:334] "Generic (PLEG): container finished" podID="f18ee3ff-994c-46c4-8f0c-76b1ee45da50" containerID="883a527a9d84208558ca0007c0a8c77189f06bf0927bb703145ef20bcf5a9c9b" exitCode=0 Feb 25 07:07:36 crc kubenswrapper[4978]: I0225 07:07:36.791573 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tc5ks" event={"ID":"f18ee3ff-994c-46c4-8f0c-76b1ee45da50","Type":"ContainerDied","Data":"883a527a9d84208558ca0007c0a8c77189f06bf0927bb703145ef20bcf5a9c9b"} Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.688463 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.798730 4978 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.798748 4978 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.798781 4978 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.824117 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.844951 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-77bc5d768-qx88q"] Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.846648 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.848998 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.864297 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.868013 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-77bc5d768-qx88q"] Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.959992 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.960087 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-internal-tls-certs\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.960112 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2631ab-861d-40eb-9b52-b4fd269eb411-logs\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.960139 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-combined-ca-bundle\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.960284 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data-custom\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.960303 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vsbp\" (UniqueName: \"kubernetes.io/projected/3a2631ab-861d-40eb-9b52-b4fd269eb411-kube-api-access-5vsbp\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.960353 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-public-tls-certs\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:37 crc kubenswrapper[4978]: I0225 07:07:37.974173 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.027991 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.064703 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-internal-tls-certs\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.064752 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2631ab-861d-40eb-9b52-b4fd269eb411-logs\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.064787 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-combined-ca-bundle\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.064872 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data-custom\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.064893 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vsbp\" (UniqueName: \"kubernetes.io/projected/3a2631ab-861d-40eb-9b52-b4fd269eb411-kube-api-access-5vsbp\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.064929 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-public-tls-certs\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.064957 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.067766 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2631ab-861d-40eb-9b52-b4fd269eb411-logs\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.072115 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-public-tls-certs\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.074650 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.090339 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data-custom\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.092357 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-internal-tls-certs\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.096767 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vsbp\" (UniqueName: \"kubernetes.io/projected/3a2631ab-861d-40eb-9b52-b4fd269eb411-kube-api-access-5vsbp\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.109025 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-combined-ca-bundle\") pod \"barbican-api-77bc5d768-qx88q\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:38 crc kubenswrapper[4978]: I0225 07:07:38.229346 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.635649 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.717634 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9ttn\" (UniqueName: \"kubernetes.io/projected/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-kube-api-access-s9ttn\") pod \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.717956 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-config-data\") pod \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.717986 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-etc-machine-id\") pod \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.718168 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-combined-ca-bundle\") pod \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.718236 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-db-sync-config-data\") pod \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.718269 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-scripts\") pod \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\" (UID: \"f18ee3ff-994c-46c4-8f0c-76b1ee45da50\") " Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.718953 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f18ee3ff-994c-46c4-8f0c-76b1ee45da50" (UID: "f18ee3ff-994c-46c4-8f0c-76b1ee45da50"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.723503 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f18ee3ff-994c-46c4-8f0c-76b1ee45da50" (UID: "f18ee3ff-994c-46c4-8f0c-76b1ee45da50"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.723615 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-kube-api-access-s9ttn" (OuterVolumeSpecName: "kube-api-access-s9ttn") pod "f18ee3ff-994c-46c4-8f0c-76b1ee45da50" (UID: "f18ee3ff-994c-46c4-8f0c-76b1ee45da50"). InnerVolumeSpecName "kube-api-access-s9ttn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.725482 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-scripts" (OuterVolumeSpecName: "scripts") pod "f18ee3ff-994c-46c4-8f0c-76b1ee45da50" (UID: "f18ee3ff-994c-46c4-8f0c-76b1ee45da50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.744539 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f18ee3ff-994c-46c4-8f0c-76b1ee45da50" (UID: "f18ee3ff-994c-46c4-8f0c-76b1ee45da50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.803641 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-config-data" (OuterVolumeSpecName: "config-data") pod "f18ee3ff-994c-46c4-8f0c-76b1ee45da50" (UID: "f18ee3ff-994c-46c4-8f0c-76b1ee45da50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.820305 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.820342 4978 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.820353 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.820364 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9ttn\" (UniqueName: \"kubernetes.io/projected/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-kube-api-access-s9ttn\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.820386 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.820395 4978 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f18ee3ff-994c-46c4-8f0c-76b1ee45da50-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.832746 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-tc5ks" event={"ID":"f18ee3ff-994c-46c4-8f0c-76b1ee45da50","Type":"ContainerDied","Data":"a1c9a063536805161964b6f751b3329edfe8bd7212f701ca56b03986b200184a"} Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.832832 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1c9a063536805161964b6f751b3329edfe8bd7212f701ca56b03986b200184a" Feb 25 07:07:40 crc kubenswrapper[4978]: I0225 07:07:40.833358 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-tc5ks" Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.684504 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-77bc5d768-qx88q"] Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.746396 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-99b7f5fd5-gp9zm"] Feb 25 07:07:41 crc kubenswrapper[4978]: W0225 07:07:41.748728 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6653838e_0077_4fbb_aa6b_ec6ccdfb7078.slice/crio-28e2c6ced6ba5c140b380b6bba2cab7c1c7a983eb0da1bb585eb4308ed515b23 WatchSource:0}: Error finding container 28e2c6ced6ba5c140b380b6bba2cab7c1c7a983eb0da1bb585eb4308ed515b23: Status 404 returned error can't find the container with id 28e2c6ced6ba5c140b380b6bba2cab7c1c7a983eb0da1bb585eb4308ed515b23 Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.755254 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5dc9c48444-drpjk"] Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.846055 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" event={"ID":"6653838e-0077-4fbb-aa6b-ec6ccdfb7078","Type":"ContainerStarted","Data":"28e2c6ced6ba5c140b380b6bba2cab7c1c7a983eb0da1bb585eb4308ed515b23"} Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.855625 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dc9c48444-drpjk" event={"ID":"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98","Type":"ContainerStarted","Data":"3b4642f3e31e9f50cbb865f90bad61ab25a2c4c315b0aea81dc1f693aafab18b"} Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.859262 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77bc5d768-qx88q" event={"ID":"3a2631ab-861d-40eb-9b52-b4fd269eb411","Type":"ContainerStarted","Data":"3eda4c956f7bfa3516a7c81625512fe074672ee19730bb4f83f588aa74527021"} Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.869178 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f812baa4-8d36-4a69-b0c1-49fb714beced","Type":"ContainerStarted","Data":"3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d"} Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.869335 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="ceilometer-central-agent" containerID="cri-o://8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b" gracePeriod=30 Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.869471 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.869581 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="proxy-httpd" containerID="cri-o://3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d" gracePeriod=30 Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.869721 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="sg-core" containerID="cri-o://460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099" gracePeriod=30 Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.869770 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="ceilometer-notification-agent" containerID="cri-o://84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779" gracePeriod=30 Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.895166 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.0619920880000002 podStartE2EDuration="47.895152025s" podCreationTimestamp="2026-02-25 07:06:54 +0000 UTC" firstStartedPulling="2026-02-25 07:06:55.413405543 +0000 UTC m=+1308.852661992" lastFinishedPulling="2026-02-25 07:07:41.24656544 +0000 UTC m=+1354.685821929" observedRunningTime="2026-02-25 07:07:41.893444664 +0000 UTC m=+1355.332701123" watchObservedRunningTime="2026-02-25 07:07:41.895152025 +0000 UTC m=+1355.334408484" Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.929561 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:07:41 crc kubenswrapper[4978]: E0225 07:07:41.929872 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f18ee3ff-994c-46c4-8f0c-76b1ee45da50" containerName="cinder-db-sync" Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.929883 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f18ee3ff-994c-46c4-8f0c-76b1ee45da50" containerName="cinder-db-sync" Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.932529 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f18ee3ff-994c-46c4-8f0c-76b1ee45da50" containerName="cinder-db-sync" Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.933420 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.946321 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.946486 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.946610 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zzfvc" Feb 25 07:07:41 crc kubenswrapper[4978]: I0225 07:07:41.946741 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.012444 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-99b7f5fd5-gp9zm"] Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.047159 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.049974 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b110f494-d9b3-4e97-8431-c543f0822e61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.050046 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-scripts\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.050074 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.050097 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.052788 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5hcc\" (UniqueName: \"kubernetes.io/projected/b110f494-d9b3-4e97-8431-c543f0822e61-kube-api-access-l5hcc\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.052829 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.071948 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d4ffdd57-jtdkj"] Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.073824 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.109995 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7696768c55-mnqh9"] Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.119478 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d4ffdd57-jtdkj"] Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.127311 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5b4d65dff6-x8qrr"] Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.150951 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.152281 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154130 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b110f494-d9b3-4e97-8431-c543f0822e61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154188 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-svc\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154225 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckxj2\" (UniqueName: \"kubernetes.io/projected/32b6e3c5-3010-4286-b0aa-a2bd20517933-kube-api-access-ckxj2\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154243 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-config\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154267 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-scripts\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154291 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154309 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154348 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-sb\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154382 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5hcc\" (UniqueName: \"kubernetes.io/projected/b110f494-d9b3-4e97-8431-c543f0822e61-kube-api-access-l5hcc\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154405 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154434 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-nb\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154468 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-swift-storage-0\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.154548 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b110f494-d9b3-4e97-8431-c543f0822e61-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.158112 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.161166 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-scripts\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.162205 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.162416 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.165174 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.165448 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.195136 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5hcc\" (UniqueName: \"kubernetes.io/projected/b110f494-d9b3-4e97-8431-c543f0822e61-kube-api-access-l5hcc\") pod \"cinder-scheduler-0\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.255699 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-swift-storage-0\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.255803 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-svc\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.255832 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-logs\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.255861 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data-custom\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.255882 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckxj2\" (UniqueName: \"kubernetes.io/projected/32b6e3c5-3010-4286-b0aa-a2bd20517933-kube-api-access-ckxj2\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.255902 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-config\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.255946 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skfn2\" (UniqueName: \"kubernetes.io/projected/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-kube-api-access-skfn2\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.255972 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.255986 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.256006 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-sb\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.256036 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-etc-machine-id\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.256049 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-scripts\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.256071 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-nb\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.256601 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-swift-storage-0\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.256829 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-nb\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.257142 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-svc\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.257422 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-sb\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.257922 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-config\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.272395 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckxj2\" (UniqueName: \"kubernetes.io/projected/32b6e3c5-3010-4286-b0aa-a2bd20517933-kube-api-access-ckxj2\") pod \"dnsmasq-dns-7d4ffdd57-jtdkj\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.313637 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.365357 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skfn2\" (UniqueName: \"kubernetes.io/projected/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-kube-api-access-skfn2\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.365715 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.365738 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.365768 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-etc-machine-id\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.365785 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-scripts\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.365904 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-logs\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.365941 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data-custom\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.366781 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-etc-machine-id\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.368231 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-logs\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.369983 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data-custom\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.370713 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.371620 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.380112 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-scripts\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.382033 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skfn2\" (UniqueName: \"kubernetes.io/projected/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-kube-api-access-skfn2\") pod \"cinder-api-0\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.424267 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.506419 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.847849 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:07:42 crc kubenswrapper[4978]: W0225 07:07:42.863531 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb110f494_d9b3_4e97_8431_c543f0822e61.slice/crio-e812211709974fa17b6c4b2870521c3e2d971e97c3b4491f07b55b0e9a451ba7 WatchSource:0}: Error finding container e812211709974fa17b6c4b2870521c3e2d971e97c3b4491f07b55b0e9a451ba7: Status 404 returned error can't find the container with id e812211709974fa17b6c4b2870521c3e2d971e97c3b4491f07b55b0e9a451ba7 Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.886066 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" event={"ID":"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a","Type":"ContainerStarted","Data":"970e5ba4ea94db9181506a1fcfbd970c41876a8951f71cdc8d5839d4046099bf"} Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.913358 4978 generic.go:334] "Generic (PLEG): container finished" podID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerID="3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d" exitCode=0 Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.913477 4978 generic.go:334] "Generic (PLEG): container finished" podID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerID="460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099" exitCode=2 Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.913489 4978 generic.go:334] "Generic (PLEG): container finished" podID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerID="8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b" exitCode=0 Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.913577 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f812baa4-8d36-4a69-b0c1-49fb714beced","Type":"ContainerDied","Data":"3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d"} Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.913640 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f812baa4-8d36-4a69-b0c1-49fb714beced","Type":"ContainerDied","Data":"460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099"} Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.913656 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f812baa4-8d36-4a69-b0c1-49fb714beced","Type":"ContainerDied","Data":"8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b"} Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.924728 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7696768c55-mnqh9" event={"ID":"a762aa77-9eb1-42eb-98e5-016cc3acac10","Type":"ContainerStarted","Data":"5e7e5ca67e7e0604612baa166176d1957dd92e7cfb16bf5f97d6305b191dea2f"} Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.926931 4978 generic.go:334] "Generic (PLEG): container finished" podID="6653838e-0077-4fbb-aa6b-ec6ccdfb7078" containerID="b1144d59dd3d789df752c54e2d19d37b0a08b93ba7d79fecdca73335d9017010" exitCode=0 Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.927017 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" event={"ID":"6653838e-0077-4fbb-aa6b-ec6ccdfb7078","Type":"ContainerDied","Data":"b1144d59dd3d789df752c54e2d19d37b0a08b93ba7d79fecdca73335d9017010"} Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.977964 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dc9c48444-drpjk" event={"ID":"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98","Type":"ContainerStarted","Data":"daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc"} Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.978008 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dc9c48444-drpjk" event={"ID":"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98","Type":"ContainerStarted","Data":"5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97"} Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.979113 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:42 crc kubenswrapper[4978]: I0225 07:07:42.979142 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.018505 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77bc5d768-qx88q" event={"ID":"3a2631ab-861d-40eb-9b52-b4fd269eb411","Type":"ContainerStarted","Data":"e3bb5987c1f9fda6b9f36d86116e1f3a943d9693a3917d0c5c728a4dd3429744"} Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.018542 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77bc5d768-qx88q" event={"ID":"3a2631ab-861d-40eb-9b52-b4fd269eb411","Type":"ContainerStarted","Data":"3d765602f0864d6d998940f94e6b40eaa60dbf1a6c6bdb9427a47ca7993400a3"} Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.019531 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.019553 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.035396 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5dc9c48444-drpjk" podStartSLOduration=8.03535997 podStartE2EDuration="8.03535997s" podCreationTimestamp="2026-02-25 07:07:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:43.008747729 +0000 UTC m=+1356.448004188" watchObservedRunningTime="2026-02-25 07:07:43.03535997 +0000 UTC m=+1356.474616429" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.131104 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-77bc5d768-qx88q" podStartSLOduration=6.131090284 podStartE2EDuration="6.131090284s" podCreationTimestamp="2026-02-25 07:07:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:43.059256861 +0000 UTC m=+1356.498513320" watchObservedRunningTime="2026-02-25 07:07:43.131090284 +0000 UTC m=+1356.570346743" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.145580 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d4ffdd57-jtdkj"] Feb 25 07:07:43 crc kubenswrapper[4978]: W0225 07:07:43.147584 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32b6e3c5_3010_4286_b0aa_a2bd20517933.slice/crio-b385d710d3f2187ec84337f9dd4741031041c92d7cb4ca69bdee370613eec7db WatchSource:0}: Error finding container b385d710d3f2187ec84337f9dd4741031041c92d7cb4ca69bdee370613eec7db: Status 404 returned error can't find the container with id b385d710d3f2187ec84337f9dd4741031041c92d7cb4ca69bdee370613eec7db Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.363123 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.418140 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.619455 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-nb\") pod \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.619539 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-swift-storage-0\") pod \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.619569 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-sb\") pod \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.619607 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-config\") pod \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.619752 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzk9g\" (UniqueName: \"kubernetes.io/projected/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-kube-api-access-pzk9g\") pod \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.619793 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-svc\") pod \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\" (UID: \"6653838e-0077-4fbb-aa6b-ec6ccdfb7078\") " Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.625095 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-kube-api-access-pzk9g" (OuterVolumeSpecName: "kube-api-access-pzk9g") pod "6653838e-0077-4fbb-aa6b-ec6ccdfb7078" (UID: "6653838e-0077-4fbb-aa6b-ec6ccdfb7078"). InnerVolumeSpecName "kube-api-access-pzk9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.644224 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6653838e-0077-4fbb-aa6b-ec6ccdfb7078" (UID: "6653838e-0077-4fbb-aa6b-ec6ccdfb7078"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.648669 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6653838e-0077-4fbb-aa6b-ec6ccdfb7078" (UID: "6653838e-0077-4fbb-aa6b-ec6ccdfb7078"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.649156 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6653838e-0077-4fbb-aa6b-ec6ccdfb7078" (UID: "6653838e-0077-4fbb-aa6b-ec6ccdfb7078"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.655456 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6653838e-0077-4fbb-aa6b-ec6ccdfb7078" (UID: "6653838e-0077-4fbb-aa6b-ec6ccdfb7078"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.664887 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-config" (OuterVolumeSpecName: "config") pod "6653838e-0077-4fbb-aa6b-ec6ccdfb7078" (UID: "6653838e-0077-4fbb-aa6b-ec6ccdfb7078"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.721744 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.722002 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzk9g\" (UniqueName: \"kubernetes.io/projected/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-kube-api-access-pzk9g\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.722013 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.722023 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.722033 4978 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:43 crc kubenswrapper[4978]: I0225 07:07:43.722041 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6653838e-0077-4fbb-aa6b-ec6ccdfb7078-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.028107 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125","Type":"ContainerStarted","Data":"6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a"} Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.028150 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125","Type":"ContainerStarted","Data":"9b6ba52579b197654f60d5e2d84fb79277b4cd84d96ded7eae1fcba28db18a48"} Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.029065 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b110f494-d9b3-4e97-8431-c543f0822e61","Type":"ContainerStarted","Data":"e812211709974fa17b6c4b2870521c3e2d971e97c3b4491f07b55b0e9a451ba7"} Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.030384 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" event={"ID":"6653838e-0077-4fbb-aa6b-ec6ccdfb7078","Type":"ContainerDied","Data":"28e2c6ced6ba5c140b380b6bba2cab7c1c7a983eb0da1bb585eb4308ed515b23"} Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.030434 4978 scope.go:117] "RemoveContainer" containerID="b1144d59dd3d789df752c54e2d19d37b0a08b93ba7d79fecdca73335d9017010" Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.030512 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99b7f5fd5-gp9zm" Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.045097 4978 generic.go:334] "Generic (PLEG): container finished" podID="32b6e3c5-3010-4286-b0aa-a2bd20517933" containerID="cbffce9c96e1a8b231f5183e40f0fa24451e08a5a2711947712f77aa834fbc2d" exitCode=0 Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.045218 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" event={"ID":"32b6e3c5-3010-4286-b0aa-a2bd20517933","Type":"ContainerDied","Data":"cbffce9c96e1a8b231f5183e40f0fa24451e08a5a2711947712f77aa834fbc2d"} Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.045262 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" event={"ID":"32b6e3c5-3010-4286-b0aa-a2bd20517933","Type":"ContainerStarted","Data":"b385d710d3f2187ec84337f9dd4741031041c92d7cb4ca69bdee370613eec7db"} Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.140275 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-99b7f5fd5-gp9zm"] Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.181548 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-99b7f5fd5-gp9zm"] Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.191004 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:07:44 crc kubenswrapper[4978]: I0225 07:07:44.983118 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.045267 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-config-data\") pod \"f812baa4-8d36-4a69-b0c1-49fb714beced\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.045337 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-run-httpd\") pod \"f812baa4-8d36-4a69-b0c1-49fb714beced\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.045406 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-sg-core-conf-yaml\") pod \"f812baa4-8d36-4a69-b0c1-49fb714beced\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.045440 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-log-httpd\") pod \"f812baa4-8d36-4a69-b0c1-49fb714beced\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.045485 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-scripts\") pod \"f812baa4-8d36-4a69-b0c1-49fb714beced\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.045522 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-combined-ca-bundle\") pod \"f812baa4-8d36-4a69-b0c1-49fb714beced\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.045560 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kq7t\" (UniqueName: \"kubernetes.io/projected/f812baa4-8d36-4a69-b0c1-49fb714beced-kube-api-access-5kq7t\") pod \"f812baa4-8d36-4a69-b0c1-49fb714beced\" (UID: \"f812baa4-8d36-4a69-b0c1-49fb714beced\") " Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.045949 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f812baa4-8d36-4a69-b0c1-49fb714beced" (UID: "f812baa4-8d36-4a69-b0c1-49fb714beced"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.046284 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f812baa4-8d36-4a69-b0c1-49fb714beced" (UID: "f812baa4-8d36-4a69-b0c1-49fb714beced"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.052484 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-scripts" (OuterVolumeSpecName: "scripts") pod "f812baa4-8d36-4a69-b0c1-49fb714beced" (UID: "f812baa4-8d36-4a69-b0c1-49fb714beced"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.054844 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f812baa4-8d36-4a69-b0c1-49fb714beced-kube-api-access-5kq7t" (OuterVolumeSpecName: "kube-api-access-5kq7t") pod "f812baa4-8d36-4a69-b0c1-49fb714beced" (UID: "f812baa4-8d36-4a69-b0c1-49fb714beced"). InnerVolumeSpecName "kube-api-access-5kq7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.073766 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f812baa4-8d36-4a69-b0c1-49fb714beced" (UID: "f812baa4-8d36-4a69-b0c1-49fb714beced"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.076827 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" event={"ID":"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a","Type":"ContainerStarted","Data":"965888d2495c19353834efdc1dc9bee729c306c0310cc39f6260c432b9398f32"} Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.083933 4978 generic.go:334] "Generic (PLEG): container finished" podID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerID="84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779" exitCode=0 Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.083989 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f812baa4-8d36-4a69-b0c1-49fb714beced","Type":"ContainerDied","Data":"84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779"} Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.084012 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f812baa4-8d36-4a69-b0c1-49fb714beced","Type":"ContainerDied","Data":"684932be1fff80d666c4c7817523272c34daa5f728400430e15d1e71eb499300"} Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.084029 4978 scope.go:117] "RemoveContainer" containerID="3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.084120 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.090578 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7696768c55-mnqh9" event={"ID":"a762aa77-9eb1-42eb-98e5-016cc3acac10","Type":"ContainerStarted","Data":"47894865746e5efcf05b88359453cdd68cf29193d3fc760e152761c140f46884"} Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.101289 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" event={"ID":"32b6e3c5-3010-4286-b0aa-a2bd20517933","Type":"ContainerStarted","Data":"8b4f5ceb88e5882edab1442efa5555b2f11dddb6c403904c3906cd14a276a6f4"} Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.101863 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.125051 4978 scope.go:117] "RemoveContainer" containerID="460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.125635 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" podStartSLOduration=4.125617211 podStartE2EDuration="4.125617211s" podCreationTimestamp="2026-02-25 07:07:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:45.122202679 +0000 UTC m=+1358.561459148" watchObservedRunningTime="2026-02-25 07:07:45.125617211 +0000 UTC m=+1358.564873670" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.148237 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.148265 4978 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.148275 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f812baa4-8d36-4a69-b0c1-49fb714beced-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.148283 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.148294 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kq7t\" (UniqueName: \"kubernetes.io/projected/f812baa4-8d36-4a69-b0c1-49fb714beced-kube-api-access-5kq7t\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.164763 4978 scope.go:117] "RemoveContainer" containerID="84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.178276 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-config-data" (OuterVolumeSpecName: "config-data") pod "f812baa4-8d36-4a69-b0c1-49fb714beced" (UID: "f812baa4-8d36-4a69-b0c1-49fb714beced"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.189263 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f812baa4-8d36-4a69-b0c1-49fb714beced" (UID: "f812baa4-8d36-4a69-b0c1-49fb714beced"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.253512 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.253797 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f812baa4-8d36-4a69-b0c1-49fb714beced-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.259500 4978 scope.go:117] "RemoveContainer" containerID="8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.326809 4978 scope.go:117] "RemoveContainer" containerID="3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d" Feb 25 07:07:45 crc kubenswrapper[4978]: E0225 07:07:45.330472 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d\": container with ID starting with 3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d not found: ID does not exist" containerID="3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.330518 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d"} err="failed to get container status \"3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d\": rpc error: code = NotFound desc = could not find container \"3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d\": container with ID starting with 3f6a8c8d2b4589bbaae11a5ce22a1fa1f6ab507f6b44d4400f7d34b7c3d7e13d not found: ID does not exist" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.330544 4978 scope.go:117] "RemoveContainer" containerID="460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099" Feb 25 07:07:45 crc kubenswrapper[4978]: E0225 07:07:45.331073 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099\": container with ID starting with 460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099 not found: ID does not exist" containerID="460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.331092 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099"} err="failed to get container status \"460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099\": rpc error: code = NotFound desc = could not find container \"460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099\": container with ID starting with 460cec4cf8f0c0677f86d7f521d25119f4d0ad94fda6663418ba4c0b93134099 not found: ID does not exist" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.331108 4978 scope.go:117] "RemoveContainer" containerID="84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779" Feb 25 07:07:45 crc kubenswrapper[4978]: E0225 07:07:45.336459 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779\": container with ID starting with 84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779 not found: ID does not exist" containerID="84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.336493 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779"} err="failed to get container status \"84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779\": rpc error: code = NotFound desc = could not find container \"84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779\": container with ID starting with 84e0264c4907fcce8c82e740f785b139bb518565f643b935d98f428a58106779 not found: ID does not exist" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.336513 4978 scope.go:117] "RemoveContainer" containerID="8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b" Feb 25 07:07:45 crc kubenswrapper[4978]: E0225 07:07:45.344610 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b\": container with ID starting with 8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b not found: ID does not exist" containerID="8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.344659 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b"} err="failed to get container status \"8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b\": rpc error: code = NotFound desc = could not find container \"8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b\": container with ID starting with 8e9ef937ae259a1320071fda6848e727f9223ae1df2e67aaef98cbc61b172a8b not found: ID does not exist" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.347338 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6653838e-0077-4fbb-aa6b-ec6ccdfb7078" path="/var/lib/kubelet/pods/6653838e-0077-4fbb-aa6b-ec6ccdfb7078/volumes" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.419428 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.431315 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.438630 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:07:45 crc kubenswrapper[4978]: E0225 07:07:45.438944 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="ceilometer-central-agent" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.438962 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="ceilometer-central-agent" Feb 25 07:07:45 crc kubenswrapper[4978]: E0225 07:07:45.438973 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6653838e-0077-4fbb-aa6b-ec6ccdfb7078" containerName="init" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.438980 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6653838e-0077-4fbb-aa6b-ec6ccdfb7078" containerName="init" Feb 25 07:07:45 crc kubenswrapper[4978]: E0225 07:07:45.439002 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="proxy-httpd" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.439009 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="proxy-httpd" Feb 25 07:07:45 crc kubenswrapper[4978]: E0225 07:07:45.439023 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="ceilometer-notification-agent" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.439029 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="ceilometer-notification-agent" Feb 25 07:07:45 crc kubenswrapper[4978]: E0225 07:07:45.439039 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="sg-core" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.439046 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="sg-core" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.439204 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="sg-core" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.439221 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="proxy-httpd" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.439229 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="ceilometer-central-agent" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.439239 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6653838e-0077-4fbb-aa6b-ec6ccdfb7078" containerName="init" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.439250 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" containerName="ceilometer-notification-agent" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.440669 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.444714 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.445388 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.452468 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.581621 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-log-httpd\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.581817 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64zr7\" (UniqueName: \"kubernetes.io/projected/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-kube-api-access-64zr7\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.581854 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-config-data\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.581896 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.581946 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-run-httpd\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.581990 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.582011 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-scripts\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.683754 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64zr7\" (UniqueName: \"kubernetes.io/projected/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-kube-api-access-64zr7\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.684046 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-config-data\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.684074 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.684125 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-run-httpd\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.684151 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.684170 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-scripts\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.684220 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-log-httpd\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.684674 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-log-httpd\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.685283 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-run-httpd\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.690179 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.691863 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-config-data\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.695127 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-scripts\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.704988 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.707000 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64zr7\" (UniqueName: \"kubernetes.io/projected/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-kube-api-access-64zr7\") pod \"ceilometer-0\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " pod="openstack/ceilometer-0" Feb 25 07:07:45 crc kubenswrapper[4978]: I0225 07:07:45.807511 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.112576 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b110f494-d9b3-4e97-8431-c543f0822e61","Type":"ContainerStarted","Data":"8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf"} Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.117393 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" event={"ID":"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a","Type":"ContainerStarted","Data":"7f25fb6fb4d79c681a829daa912270875a167f3eab8e635f499b9abf6ae1a887"} Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.121560 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125","Type":"ContainerStarted","Data":"00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58"} Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.121721 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" containerName="cinder-api-log" containerID="cri-o://6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a" gracePeriod=30 Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.121980 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.122031 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" containerName="cinder-api" containerID="cri-o://00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58" gracePeriod=30 Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.134898 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7696768c55-mnqh9" event={"ID":"a762aa77-9eb1-42eb-98e5-016cc3acac10","Type":"ContainerStarted","Data":"97d4b20d6d6a152207b32f06fed0f8b590209ec24594d4831ee695bb8b21c4ee"} Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.140054 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" podStartSLOduration=8.635243579 podStartE2EDuration="11.140036857s" podCreationTimestamp="2026-02-25 07:07:35 +0000 UTC" firstStartedPulling="2026-02-25 07:07:42.07192727 +0000 UTC m=+1355.511183729" lastFinishedPulling="2026-02-25 07:07:44.576720538 +0000 UTC m=+1358.015977007" observedRunningTime="2026-02-25 07:07:46.133250413 +0000 UTC m=+1359.572506882" watchObservedRunningTime="2026-02-25 07:07:46.140036857 +0000 UTC m=+1359.579293316" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.160578 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.160559136 podStartE2EDuration="4.160559136s" podCreationTimestamp="2026-02-25 07:07:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:46.15274275 +0000 UTC m=+1359.591999209" watchObservedRunningTime="2026-02-25 07:07:46.160559136 +0000 UTC m=+1359.599815595" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.184128 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7696768c55-mnqh9" podStartSLOduration=8.683614376 podStartE2EDuration="11.184110194s" podCreationTimestamp="2026-02-25 07:07:35 +0000 UTC" firstStartedPulling="2026-02-25 07:07:42.074450456 +0000 UTC m=+1355.513706915" lastFinishedPulling="2026-02-25 07:07:44.574946254 +0000 UTC m=+1358.014202733" observedRunningTime="2026-02-25 07:07:46.179849537 +0000 UTC m=+1359.619105996" watchObservedRunningTime="2026-02-25 07:07:46.184110194 +0000 UTC m=+1359.623366653" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.261898 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:07:46 crc kubenswrapper[4978]: W0225 07:07:46.268578 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc903177_5cd3_489b_8a04_fe99e6a2d7fe.slice/crio-fd0e7b7fdeecbd3c9507423d95b83c90e7741fb0a6dce5dad5434d463237d380 WatchSource:0}: Error finding container fd0e7b7fdeecbd3c9507423d95b83c90e7741fb0a6dce5dad5434d463237d380: Status 404 returned error can't find the container with id fd0e7b7fdeecbd3c9507423d95b83c90e7741fb0a6dce5dad5434d463237d380 Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.541528 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.541873 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.722671 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.727873 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skfn2\" (UniqueName: \"kubernetes.io/projected/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-kube-api-access-skfn2\") pod \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.727920 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-etc-machine-id\") pod \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.728013 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-scripts\") pod \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.728054 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data\") pod \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.728081 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-logs\") pod \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.728112 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-combined-ca-bundle\") pod \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.728144 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data-custom\") pod \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\" (UID: \"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125\") " Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.729302 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" (UID: "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.729937 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-logs" (OuterVolumeSpecName: "logs") pod "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" (UID: "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.738330 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-kube-api-access-skfn2" (OuterVolumeSpecName: "kube-api-access-skfn2") pod "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" (UID: "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125"). InnerVolumeSpecName "kube-api-access-skfn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.746317 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" (UID: "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.746822 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-scripts" (OuterVolumeSpecName: "scripts") pod "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" (UID: "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.782384 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" (UID: "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.831492 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skfn2\" (UniqueName: \"kubernetes.io/projected/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-kube-api-access-skfn2\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.831519 4978 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.831529 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.831537 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.831545 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.831553 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.836790 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data" (OuterVolumeSpecName: "config-data") pod "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" (UID: "063c5c46-4f4f-4cfb-bea2-ca9a6aac8125"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:46 crc kubenswrapper[4978]: I0225 07:07:46.933625 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.148775 4978 generic.go:334] "Generic (PLEG): container finished" podID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" containerID="00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58" exitCode=0 Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.148831 4978 generic.go:334] "Generic (PLEG): container finished" podID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" containerID="6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a" exitCode=143 Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.148822 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125","Type":"ContainerDied","Data":"00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58"} Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.148875 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.148908 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125","Type":"ContainerDied","Data":"6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a"} Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.148927 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"063c5c46-4f4f-4cfb-bea2-ca9a6aac8125","Type":"ContainerDied","Data":"9b6ba52579b197654f60d5e2d84fb79277b4cd84d96ded7eae1fcba28db18a48"} Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.148951 4978 scope.go:117] "RemoveContainer" containerID="00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.155040 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b110f494-d9b3-4e97-8431-c543f0822e61","Type":"ContainerStarted","Data":"4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7"} Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.157332 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cc903177-5cd3-489b-8a04-fe99e6a2d7fe","Type":"ContainerStarted","Data":"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3"} Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.157359 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cc903177-5cd3-489b-8a04-fe99e6a2d7fe","Type":"ContainerStarted","Data":"fd0e7b7fdeecbd3c9507423d95b83c90e7741fb0a6dce5dad5434d463237d380"} Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.170224 4978 scope.go:117] "RemoveContainer" containerID="6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.184101 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.490592451 podStartE2EDuration="6.184062621s" podCreationTimestamp="2026-02-25 07:07:41 +0000 UTC" firstStartedPulling="2026-02-25 07:07:42.880381132 +0000 UTC m=+1356.319637591" lastFinishedPulling="2026-02-25 07:07:44.573851292 +0000 UTC m=+1358.013107761" observedRunningTime="2026-02-25 07:07:47.177578136 +0000 UTC m=+1360.616834625" watchObservedRunningTime="2026-02-25 07:07:47.184062621 +0000 UTC m=+1360.623319080" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.198362 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.205040 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.215132 4978 scope.go:117] "RemoveContainer" containerID="00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58" Feb 25 07:07:47 crc kubenswrapper[4978]: E0225 07:07:47.220588 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58\": container with ID starting with 00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58 not found: ID does not exist" containerID="00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.220633 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58"} err="failed to get container status \"00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58\": rpc error: code = NotFound desc = could not find container \"00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58\": container with ID starting with 00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58 not found: ID does not exist" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.220659 4978 scope.go:117] "RemoveContainer" containerID="6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a" Feb 25 07:07:47 crc kubenswrapper[4978]: E0225 07:07:47.221050 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a\": container with ID starting with 6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a not found: ID does not exist" containerID="6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.221070 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a"} err="failed to get container status \"6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a\": rpc error: code = NotFound desc = could not find container \"6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a\": container with ID starting with 6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a not found: ID does not exist" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.221084 4978 scope.go:117] "RemoveContainer" containerID="00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.221277 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58"} err="failed to get container status \"00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58\": rpc error: code = NotFound desc = could not find container \"00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58\": container with ID starting with 00fa0cca3d8bf22273f3bbeb0a9d3d3e33552589ccf72fee81f7b2c87d4c1a58 not found: ID does not exist" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.221296 4978 scope.go:117] "RemoveContainer" containerID="6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.221532 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a"} err="failed to get container status \"6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a\": rpc error: code = NotFound desc = could not find container \"6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a\": container with ID starting with 6df268d23c6dcb7e679c65cfe86046dc7b9d3650de7b956daf03056428e94d5a not found: ID does not exist" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.267677 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:07:47 crc kubenswrapper[4978]: E0225 07:07:47.268194 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" containerName="cinder-api-log" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.268215 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" containerName="cinder-api-log" Feb 25 07:07:47 crc kubenswrapper[4978]: E0225 07:07:47.268249 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" containerName="cinder-api" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.268261 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" containerName="cinder-api" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.268513 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" containerName="cinder-api" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.268552 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" containerName="cinder-api-log" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.269755 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.270615 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.272333 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.272596 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.272821 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.314602 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.344117 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="063c5c46-4f4f-4cfb-bea2-ca9a6aac8125" path="/var/lib/kubelet/pods/063c5c46-4f4f-4cfb-bea2-ca9a6aac8125/volumes" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.345638 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f812baa4-8d36-4a69-b0c1-49fb714beced" path="/var/lib/kubelet/pods/f812baa4-8d36-4a69-b0c1-49fb714beced/volumes" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.348006 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.348072 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/807eb4fa-5af0-450e-be50-f30c96731b46-logs\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.348117 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.348138 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-public-tls-certs\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.348164 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data-custom\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.348213 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.348232 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-scripts\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.348255 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85rdz\" (UniqueName: \"kubernetes.io/projected/807eb4fa-5af0-450e-be50-f30c96731b46-kube-api-access-85rdz\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.348292 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/807eb4fa-5af0-450e-be50-f30c96731b46-etc-machine-id\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450014 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450142 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/807eb4fa-5af0-450e-be50-f30c96731b46-logs\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450232 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450269 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-public-tls-certs\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450313 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data-custom\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450469 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450516 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-scripts\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450551 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85rdz\" (UniqueName: \"kubernetes.io/projected/807eb4fa-5af0-450e-be50-f30c96731b46-kube-api-access-85rdz\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450637 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/807eb4fa-5af0-450e-be50-f30c96731b46-etc-machine-id\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450819 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/807eb4fa-5af0-450e-be50-f30c96731b46-etc-machine-id\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.450997 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/807eb4fa-5af0-450e-be50-f30c96731b46-logs\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.455220 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.455337 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.455785 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-scripts\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.458108 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-public-tls-certs\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.458401 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.458931 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data-custom\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.474605 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85rdz\" (UniqueName: \"kubernetes.io/projected/807eb4fa-5af0-450e-be50-f30c96731b46-kube-api-access-85rdz\") pod \"cinder-api-0\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " pod="openstack/cinder-api-0" Feb 25 07:07:47 crc kubenswrapper[4978]: I0225 07:07:47.588114 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 07:07:48 crc kubenswrapper[4978]: I0225 07:07:48.044817 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:07:48 crc kubenswrapper[4978]: W0225 07:07:48.058353 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod807eb4fa_5af0_450e_be50_f30c96731b46.slice/crio-32c85db95df224855ea938a7986cce424a25a22f5af89d58ab60ae8b5161bd02 WatchSource:0}: Error finding container 32c85db95df224855ea938a7986cce424a25a22f5af89d58ab60ae8b5161bd02: Status 404 returned error can't find the container with id 32c85db95df224855ea938a7986cce424a25a22f5af89d58ab60ae8b5161bd02 Feb 25 07:07:48 crc kubenswrapper[4978]: I0225 07:07:48.185737 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cc903177-5cd3-489b-8a04-fe99e6a2d7fe","Type":"ContainerStarted","Data":"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc"} Feb 25 07:07:48 crc kubenswrapper[4978]: I0225 07:07:48.192180 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"807eb4fa-5af0-450e-be50-f30c96731b46","Type":"ContainerStarted","Data":"32c85db95df224855ea938a7986cce424a25a22f5af89d58ab60ae8b5161bd02"} Feb 25 07:07:49 crc kubenswrapper[4978]: I0225 07:07:49.207335 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"807eb4fa-5af0-450e-be50-f30c96731b46","Type":"ContainerStarted","Data":"05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc"} Feb 25 07:07:49 crc kubenswrapper[4978]: I0225 07:07:49.216266 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cc903177-5cd3-489b-8a04-fe99e6a2d7fe","Type":"ContainerStarted","Data":"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9"} Feb 25 07:07:49 crc kubenswrapper[4978]: I0225 07:07:49.782123 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:49 crc kubenswrapper[4978]: I0225 07:07:49.801237 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:07:49 crc kubenswrapper[4978]: I0225 07:07:49.897839 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5dc9c48444-drpjk"] Feb 25 07:07:49 crc kubenswrapper[4978]: I0225 07:07:49.898075 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5dc9c48444-drpjk" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api-log" containerID="cri-o://5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97" gracePeriod=30 Feb 25 07:07:49 crc kubenswrapper[4978]: I0225 07:07:49.898686 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5dc9c48444-drpjk" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api" containerID="cri-o://daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc" gracePeriod=30 Feb 25 07:07:49 crc kubenswrapper[4978]: I0225 07:07:49.917948 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5dc9c48444-drpjk" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": EOF" Feb 25 07:07:49 crc kubenswrapper[4978]: I0225 07:07:49.921233 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5dc9c48444-drpjk" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": EOF" Feb 25 07:07:50 crc kubenswrapper[4978]: I0225 07:07:50.227025 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"807eb4fa-5af0-450e-be50-f30c96731b46","Type":"ContainerStarted","Data":"e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644"} Feb 25 07:07:50 crc kubenswrapper[4978]: I0225 07:07:50.227328 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 25 07:07:50 crc kubenswrapper[4978]: I0225 07:07:50.228926 4978 generic.go:334] "Generic (PLEG): container finished" podID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerID="5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97" exitCode=143 Feb 25 07:07:50 crc kubenswrapper[4978]: I0225 07:07:50.229249 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dc9c48444-drpjk" event={"ID":"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98","Type":"ContainerDied","Data":"5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97"} Feb 25 07:07:50 crc kubenswrapper[4978]: I0225 07:07:50.258534 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.258518944 podStartE2EDuration="3.258518944s" podCreationTimestamp="2026-02-25 07:07:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:50.25777218 +0000 UTC m=+1363.697028649" watchObservedRunningTime="2026-02-25 07:07:50.258518944 +0000 UTC m=+1363.697775403" Feb 25 07:07:50 crc kubenswrapper[4978]: I0225 07:07:50.928996 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.155220 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5987f7d87c-fxhz5"] Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.157510 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5987f7d87c-fxhz5" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerName="neutron-httpd" containerID="cri-o://58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db" gracePeriod=30 Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.157959 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5987f7d87c-fxhz5" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerName="neutron-api" containerID="cri-o://6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a" gracePeriod=30 Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.169449 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-5987f7d87c-fxhz5" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.156:9696/\": read tcp 10.217.0.2:34502->10.217.0.156:9696: read: connection reset by peer" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.195167 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5cdb588b97-7jkd8"] Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.204801 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.229872 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5cdb588b97-7jkd8"] Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.277622 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cc903177-5cd3-489b-8a04-fe99e6a2d7fe","Type":"ContainerStarted","Data":"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2"} Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.277676 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.313427 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.44352871 podStartE2EDuration="6.313410818s" podCreationTimestamp="2026-02-25 07:07:45 +0000 UTC" firstStartedPulling="2026-02-25 07:07:46.277480067 +0000 UTC m=+1359.716736516" lastFinishedPulling="2026-02-25 07:07:50.147362165 +0000 UTC m=+1363.586618624" observedRunningTime="2026-02-25 07:07:51.309576358 +0000 UTC m=+1364.748832827" watchObservedRunningTime="2026-02-25 07:07:51.313410818 +0000 UTC m=+1364.752667277" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.321686 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-ovndb-tls-certs\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.321760 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2fdk\" (UniqueName: \"kubernetes.io/projected/d1ba0ae6-d125-48d2-a866-615d927dc525-kube-api-access-m2fdk\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.321804 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-public-tls-certs\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.321843 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-combined-ca-bundle\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.321870 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-httpd-config\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.321930 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-internal-tls-certs\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.321993 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-config\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.423859 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-config\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.423920 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-ovndb-tls-certs\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.423956 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2fdk\" (UniqueName: \"kubernetes.io/projected/d1ba0ae6-d125-48d2-a866-615d927dc525-kube-api-access-m2fdk\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.424006 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-public-tls-certs\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.424042 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-combined-ca-bundle\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.424059 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-httpd-config\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.424150 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-internal-tls-certs\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.431214 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-httpd-config\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.431819 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-ovndb-tls-certs\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.436241 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-internal-tls-certs\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.436948 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-public-tls-certs\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.437174 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-combined-ca-bundle\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.449133 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2fdk\" (UniqueName: \"kubernetes.io/projected/d1ba0ae6-d125-48d2-a866-615d927dc525-kube-api-access-m2fdk\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.459288 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-config\") pod \"neutron-5cdb588b97-7jkd8\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:51 crc kubenswrapper[4978]: I0225 07:07:51.524284 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:52 crc kubenswrapper[4978]: I0225 07:07:52.093438 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5cdb588b97-7jkd8"] Feb 25 07:07:52 crc kubenswrapper[4978]: W0225 07:07:52.102248 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1ba0ae6_d125_48d2_a866_615d927dc525.slice/crio-7f005098f780e1f9032ee5a1662dbc43eb64e66c0c2fa9b3c638c5e14258ff69 WatchSource:0}: Error finding container 7f005098f780e1f9032ee5a1662dbc43eb64e66c0c2fa9b3c638c5e14258ff69: Status 404 returned error can't find the container with id 7f005098f780e1f9032ee5a1662dbc43eb64e66c0c2fa9b3c638c5e14258ff69 Feb 25 07:07:52 crc kubenswrapper[4978]: I0225 07:07:52.287992 4978 generic.go:334] "Generic (PLEG): container finished" podID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerID="58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db" exitCode=0 Feb 25 07:07:52 crc kubenswrapper[4978]: I0225 07:07:52.288070 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5987f7d87c-fxhz5" event={"ID":"5ad94c00-0a28-4933-ad5c-27966cc8b437","Type":"ContainerDied","Data":"58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db"} Feb 25 07:07:52 crc kubenswrapper[4978]: I0225 07:07:52.289865 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cdb588b97-7jkd8" event={"ID":"d1ba0ae6-d125-48d2-a866-615d927dc525","Type":"ContainerStarted","Data":"7f005098f780e1f9032ee5a1662dbc43eb64e66c0c2fa9b3c638c5e14258ff69"} Feb 25 07:07:52 crc kubenswrapper[4978]: I0225 07:07:52.426393 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:07:52 crc kubenswrapper[4978]: I0225 07:07:52.498285 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78b5478d49-qg7sz"] Feb 25 07:07:52 crc kubenswrapper[4978]: I0225 07:07:52.498784 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" podUID="bd220760-4e92-4cac-b4f0-8eb2c38f67be" containerName="dnsmasq-dns" containerID="cri-o://5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b" gracePeriod=10 Feb 25 07:07:52 crc kubenswrapper[4978]: I0225 07:07:52.594966 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 25 07:07:52 crc kubenswrapper[4978]: I0225 07:07:52.648922 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.056055 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.155087 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-swift-storage-0\") pod \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.155131 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-nb\") pod \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.155235 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-config\") pod \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.155296 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9hp9\" (UniqueName: \"kubernetes.io/projected/bd220760-4e92-4cac-b4f0-8eb2c38f67be-kube-api-access-d9hp9\") pod \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.155317 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-sb\") pod \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.155351 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-svc\") pod \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\" (UID: \"bd220760-4e92-4cac-b4f0-8eb2c38f67be\") " Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.161551 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd220760-4e92-4cac-b4f0-8eb2c38f67be-kube-api-access-d9hp9" (OuterVolumeSpecName: "kube-api-access-d9hp9") pod "bd220760-4e92-4cac-b4f0-8eb2c38f67be" (UID: "bd220760-4e92-4cac-b4f0-8eb2c38f67be"). InnerVolumeSpecName "kube-api-access-d9hp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.203201 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bd220760-4e92-4cac-b4f0-8eb2c38f67be" (UID: "bd220760-4e92-4cac-b4f0-8eb2c38f67be"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.203904 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bd220760-4e92-4cac-b4f0-8eb2c38f67be" (UID: "bd220760-4e92-4cac-b4f0-8eb2c38f67be"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.209578 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bd220760-4e92-4cac-b4f0-8eb2c38f67be" (UID: "bd220760-4e92-4cac-b4f0-8eb2c38f67be"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.210800 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-config" (OuterVolumeSpecName: "config") pod "bd220760-4e92-4cac-b4f0-8eb2c38f67be" (UID: "bd220760-4e92-4cac-b4f0-8eb2c38f67be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.217047 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd220760-4e92-4cac-b4f0-8eb2c38f67be" (UID: "bd220760-4e92-4cac-b4f0-8eb2c38f67be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.256918 4978 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.257145 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.257235 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.257320 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9hp9\" (UniqueName: \"kubernetes.io/projected/bd220760-4e92-4cac-b4f0-8eb2c38f67be-kube-api-access-d9hp9\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.257387 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.257441 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd220760-4e92-4cac-b4f0-8eb2c38f67be-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.300123 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cdb588b97-7jkd8" event={"ID":"d1ba0ae6-d125-48d2-a866-615d927dc525","Type":"ContainerStarted","Data":"6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba"} Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.300197 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cdb588b97-7jkd8" event={"ID":"d1ba0ae6-d125-48d2-a866-615d927dc525","Type":"ContainerStarted","Data":"35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e"} Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.301238 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.304438 4978 generic.go:334] "Generic (PLEG): container finished" podID="bd220760-4e92-4cac-b4f0-8eb2c38f67be" containerID="5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b" exitCode=0 Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.304556 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" event={"ID":"bd220760-4e92-4cac-b4f0-8eb2c38f67be","Type":"ContainerDied","Data":"5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b"} Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.304589 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" event={"ID":"bd220760-4e92-4cac-b4f0-8eb2c38f67be","Type":"ContainerDied","Data":"5aa5223f9fc92e20c81d58a8c59d9fbb0939e1e284234687230fff236d873bb6"} Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.304608 4978 scope.go:117] "RemoveContainer" containerID="5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.304529 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78b5478d49-qg7sz" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.309735 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b110f494-d9b3-4e97-8431-c543f0822e61" containerName="cinder-scheduler" containerID="cri-o://8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf" gracePeriod=30 Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.309928 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b110f494-d9b3-4e97-8431-c543f0822e61" containerName="probe" containerID="cri-o://4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7" gracePeriod=30 Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.341554 4978 scope.go:117] "RemoveContainer" containerID="8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.352897 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5cdb588b97-7jkd8" podStartSLOduration=2.352873114 podStartE2EDuration="2.352873114s" podCreationTimestamp="2026-02-25 07:07:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:53.322646752 +0000 UTC m=+1366.761903221" watchObservedRunningTime="2026-02-25 07:07:53.352873114 +0000 UTC m=+1366.792129593" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.366384 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78b5478d49-qg7sz"] Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.379958 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78b5478d49-qg7sz"] Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.387181 4978 scope.go:117] "RemoveContainer" containerID="5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b" Feb 25 07:07:53 crc kubenswrapper[4978]: E0225 07:07:53.387679 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b\": container with ID starting with 5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b not found: ID does not exist" containerID="5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.387733 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b"} err="failed to get container status \"5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b\": rpc error: code = NotFound desc = could not find container \"5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b\": container with ID starting with 5b2def5a661aa62ad6f9fb41cd3ae46554016568b37a1c67cf4f687388a7858b not found: ID does not exist" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.387768 4978 scope.go:117] "RemoveContainer" containerID="8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f" Feb 25 07:07:53 crc kubenswrapper[4978]: E0225 07:07:53.388117 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f\": container with ID starting with 8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f not found: ID does not exist" containerID="8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f" Feb 25 07:07:53 crc kubenswrapper[4978]: I0225 07:07:53.388144 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f"} err="failed to get container status \"8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f\": rpc error: code = NotFound desc = could not find container \"8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f\": container with ID starting with 8d598a112bfbde6de7217231a5861a14b78cdf11f2f60ca787790850662b747f not found: ID does not exist" Feb 25 07:07:54 crc kubenswrapper[4978]: I0225 07:07:54.315140 4978 generic.go:334] "Generic (PLEG): container finished" podID="b110f494-d9b3-4e97-8431-c543f0822e61" containerID="4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7" exitCode=0 Feb 25 07:07:54 crc kubenswrapper[4978]: I0225 07:07:54.315191 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b110f494-d9b3-4e97-8431-c543f0822e61","Type":"ContainerDied","Data":"4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7"} Feb 25 07:07:54 crc kubenswrapper[4978]: I0225 07:07:54.439564 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-5987f7d87c-fxhz5" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.156:9696/\": dial tcp 10.217.0.156:9696: connect: connection refused" Feb 25 07:07:54 crc kubenswrapper[4978]: I0225 07:07:54.825340 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:54 crc kubenswrapper[4978]: I0225 07:07:54.827921 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-776b767d76-hl99l" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.133251 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-69794fc6d6-zqnw6"] Feb 25 07:07:55 crc kubenswrapper[4978]: E0225 07:07:55.133676 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd220760-4e92-4cac-b4f0-8eb2c38f67be" containerName="dnsmasq-dns" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.133695 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd220760-4e92-4cac-b4f0-8eb2c38f67be" containerName="dnsmasq-dns" Feb 25 07:07:55 crc kubenswrapper[4978]: E0225 07:07:55.133712 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd220760-4e92-4cac-b4f0-8eb2c38f67be" containerName="init" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.133719 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd220760-4e92-4cac-b4f0-8eb2c38f67be" containerName="init" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.133913 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd220760-4e92-4cac-b4f0-8eb2c38f67be" containerName="dnsmasq-dns" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.134882 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.172040 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69794fc6d6-zqnw6"] Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.296478 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-scripts\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.296525 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-config-data\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.296543 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-combined-ca-bundle\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.296593 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a16d36-c629-483e-9b15-fbd7b8913883-logs\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.296614 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-internal-tls-certs\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.296656 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-public-tls-certs\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.296674 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7g6j\" (UniqueName: \"kubernetes.io/projected/c6a16d36-c629-483e-9b15-fbd7b8913883-kube-api-access-w7g6j\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.340776 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd220760-4e92-4cac-b4f0-8eb2c38f67be" path="/var/lib/kubelet/pods/bd220760-4e92-4cac-b4f0-8eb2c38f67be/volumes" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.348872 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5dc9c48444-drpjk" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:50440->10.217.0.163:9311: read: connection reset by peer" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.349196 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5dc9c48444-drpjk" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:50452->10.217.0.163:9311: read: connection reset by peer" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.398041 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a16d36-c629-483e-9b15-fbd7b8913883-logs\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.398096 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-internal-tls-certs\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.398141 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-public-tls-certs\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.398160 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7g6j\" (UniqueName: \"kubernetes.io/projected/c6a16d36-c629-483e-9b15-fbd7b8913883-kube-api-access-w7g6j\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.398234 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-scripts\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.398250 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-config-data\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.398267 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-combined-ca-bundle\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.399219 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a16d36-c629-483e-9b15-fbd7b8913883-logs\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.403252 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-scripts\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.405813 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-internal-tls-certs\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.412000 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-public-tls-certs\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.412968 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-combined-ca-bundle\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.414809 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-config-data\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.421463 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7g6j\" (UniqueName: \"kubernetes.io/projected/c6a16d36-c629-483e-9b15-fbd7b8913883-kube-api-access-w7g6j\") pod \"placement-69794fc6d6-zqnw6\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.469838 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.794354 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.905814 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data\") pod \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.905981 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7hsv\" (UniqueName: \"kubernetes.io/projected/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-kube-api-access-m7hsv\") pod \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.906036 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-combined-ca-bundle\") pod \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.906070 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data-custom\") pod \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.906171 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-logs\") pod \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\" (UID: \"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98\") " Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.907031 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-logs" (OuterVolumeSpecName: "logs") pod "d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" (UID: "d4cf1be0-51f7-4a5c-a792-c3bc504a8a98"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.911909 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-kube-api-access-m7hsv" (OuterVolumeSpecName: "kube-api-access-m7hsv") pod "d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" (UID: "d4cf1be0-51f7-4a5c-a792-c3bc504a8a98"). InnerVolumeSpecName "kube-api-access-m7hsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.912440 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" (UID: "d4cf1be0-51f7-4a5c-a792-c3bc504a8a98"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.939563 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" (UID: "d4cf1be0-51f7-4a5c-a792-c3bc504a8a98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.962108 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data" (OuterVolumeSpecName: "config-data") pod "d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" (UID: "d4cf1be0-51f7-4a5c-a792-c3bc504a8a98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:55 crc kubenswrapper[4978]: I0225 07:07:55.994329 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69794fc6d6-zqnw6"] Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.007723 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.007756 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7hsv\" (UniqueName: \"kubernetes.io/projected/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-kube-api-access-m7hsv\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.007768 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.007776 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.007785 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.336832 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69794fc6d6-zqnw6" event={"ID":"c6a16d36-c629-483e-9b15-fbd7b8913883","Type":"ContainerStarted","Data":"03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6"} Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.336866 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69794fc6d6-zqnw6" event={"ID":"c6a16d36-c629-483e-9b15-fbd7b8913883","Type":"ContainerStarted","Data":"9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af"} Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.336877 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69794fc6d6-zqnw6" event={"ID":"c6a16d36-c629-483e-9b15-fbd7b8913883","Type":"ContainerStarted","Data":"52264c07775f9aaabcc27a16a6dbeb0d55deb95b0d2bc1cda2b362cb0db62e1d"} Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.337513 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.337534 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.339993 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5dc9c48444-drpjk" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.339966 4978 generic.go:334] "Generic (PLEG): container finished" podID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerID="daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc" exitCode=0 Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.340064 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dc9c48444-drpjk" event={"ID":"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98","Type":"ContainerDied","Data":"daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc"} Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.340439 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5dc9c48444-drpjk" event={"ID":"d4cf1be0-51f7-4a5c-a792-c3bc504a8a98","Type":"ContainerDied","Data":"3b4642f3e31e9f50cbb865f90bad61ab25a2c4c315b0aea81dc1f693aafab18b"} Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.340502 4978 scope.go:117] "RemoveContainer" containerID="daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.370454 4978 scope.go:117] "RemoveContainer" containerID="5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.385021 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-69794fc6d6-zqnw6" podStartSLOduration=1.384999755 podStartE2EDuration="1.384999755s" podCreationTimestamp="2026-02-25 07:07:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:07:56.372651046 +0000 UTC m=+1369.811907595" watchObservedRunningTime="2026-02-25 07:07:56.384999755 +0000 UTC m=+1369.824256224" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.408087 4978 scope.go:117] "RemoveContainer" containerID="daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc" Feb 25 07:07:56 crc kubenswrapper[4978]: E0225 07:07:56.410785 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc\": container with ID starting with daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc not found: ID does not exist" containerID="daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.410839 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc"} err="failed to get container status \"daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc\": rpc error: code = NotFound desc = could not find container \"daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc\": container with ID starting with daef4614f6448b986fb45d2df35af7ad76d768e08ba72ffb5b43e73b4f007dcc not found: ID does not exist" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.410866 4978 scope.go:117] "RemoveContainer" containerID="5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97" Feb 25 07:07:56 crc kubenswrapper[4978]: E0225 07:07:56.411505 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97\": container with ID starting with 5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97 not found: ID does not exist" containerID="5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.411569 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97"} err="failed to get container status \"5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97\": rpc error: code = NotFound desc = could not find container \"5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97\": container with ID starting with 5049382e9bed242e466946cd431421b6b55e713ba05c153dc48ad8a95f860f97 not found: ID does not exist" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.414661 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5dc9c48444-drpjk"] Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.421912 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5dc9c48444-drpjk"] Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.701194 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.825625 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.921566 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b110f494-d9b3-4e97-8431-c543f0822e61-etc-machine-id\") pod \"b110f494-d9b3-4e97-8431-c543f0822e61\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.921688 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data-custom\") pod \"b110f494-d9b3-4e97-8431-c543f0822e61\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.921718 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data\") pod \"b110f494-d9b3-4e97-8431-c543f0822e61\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.921741 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-combined-ca-bundle\") pod \"b110f494-d9b3-4e97-8431-c543f0822e61\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.921757 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-scripts\") pod \"b110f494-d9b3-4e97-8431-c543f0822e61\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.921788 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5hcc\" (UniqueName: \"kubernetes.io/projected/b110f494-d9b3-4e97-8431-c543f0822e61-kube-api-access-l5hcc\") pod \"b110f494-d9b3-4e97-8431-c543f0822e61\" (UID: \"b110f494-d9b3-4e97-8431-c543f0822e61\") " Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.923754 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b110f494-d9b3-4e97-8431-c543f0822e61-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b110f494-d9b3-4e97-8431-c543f0822e61" (UID: "b110f494-d9b3-4e97-8431-c543f0822e61"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.925973 4978 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b110f494-d9b3-4e97-8431-c543f0822e61-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.927522 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b110f494-d9b3-4e97-8431-c543f0822e61" (UID: "b110f494-d9b3-4e97-8431-c543f0822e61"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.928897 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b110f494-d9b3-4e97-8431-c543f0822e61-kube-api-access-l5hcc" (OuterVolumeSpecName: "kube-api-access-l5hcc") pod "b110f494-d9b3-4e97-8431-c543f0822e61" (UID: "b110f494-d9b3-4e97-8431-c543f0822e61"). InnerVolumeSpecName "kube-api-access-l5hcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.952657 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-scripts" (OuterVolumeSpecName: "scripts") pod "b110f494-d9b3-4e97-8431-c543f0822e61" (UID: "b110f494-d9b3-4e97-8431-c543f0822e61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:56 crc kubenswrapper[4978]: I0225 07:07:56.978336 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b110f494-d9b3-4e97-8431-c543f0822e61" (UID: "b110f494-d9b3-4e97-8431-c543f0822e61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.010357 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data" (OuterVolumeSpecName: "config-data") pod "b110f494-d9b3-4e97-8431-c543f0822e61" (UID: "b110f494-d9b3-4e97-8431-c543f0822e61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.027669 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.027705 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.027715 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.027724 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b110f494-d9b3-4e97-8431-c543f0822e61-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.027733 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5hcc\" (UniqueName: \"kubernetes.io/projected/b110f494-d9b3-4e97-8431-c543f0822e61-kube-api-access-l5hcc\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.356724 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" path="/var/lib/kubelet/pods/d4cf1be0-51f7-4a5c-a792-c3bc504a8a98/volumes" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.358547 4978 generic.go:334] "Generic (PLEG): container finished" podID="b110f494-d9b3-4e97-8431-c543f0822e61" containerID="8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf" exitCode=0 Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.359868 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b110f494-d9b3-4e97-8431-c543f0822e61","Type":"ContainerDied","Data":"8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf"} Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.359938 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b110f494-d9b3-4e97-8431-c543f0822e61","Type":"ContainerDied","Data":"e812211709974fa17b6c4b2870521c3e2d971e97c3b4491f07b55b0e9a451ba7"} Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.360252 4978 scope.go:117] "RemoveContainer" containerID="4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.360437 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.390542 4978 scope.go:117] "RemoveContainer" containerID="8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.427912 4978 scope.go:117] "RemoveContainer" containerID="4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7" Feb 25 07:07:57 crc kubenswrapper[4978]: E0225 07:07:57.428794 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7\": container with ID starting with 4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7 not found: ID does not exist" containerID="4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.428878 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7"} err="failed to get container status \"4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7\": rpc error: code = NotFound desc = could not find container \"4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7\": container with ID starting with 4ab474e7538463becec5e67c5010b2415bfe60987dbf66f8b529c7c99c7e50c7 not found: ID does not exist" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.428923 4978 scope.go:117] "RemoveContainer" containerID="8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf" Feb 25 07:07:57 crc kubenswrapper[4978]: E0225 07:07:57.429829 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf\": container with ID starting with 8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf not found: ID does not exist" containerID="8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.429879 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf"} err="failed to get container status \"8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf\": rpc error: code = NotFound desc = could not find container \"8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf\": container with ID starting with 8afbcc131c43c5b267f7cc092194d8bd3b5f704738dbf9f42e515fd52030e8bf not found: ID does not exist" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.466416 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.477558 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.484790 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:07:57 crc kubenswrapper[4978]: E0225 07:07:57.485217 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.485240 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api" Feb 25 07:07:57 crc kubenswrapper[4978]: E0225 07:07:57.485260 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api-log" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.485267 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api-log" Feb 25 07:07:57 crc kubenswrapper[4978]: E0225 07:07:57.485290 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b110f494-d9b3-4e97-8431-c543f0822e61" containerName="probe" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.485296 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b110f494-d9b3-4e97-8431-c543f0822e61" containerName="probe" Feb 25 07:07:57 crc kubenswrapper[4978]: E0225 07:07:57.485317 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b110f494-d9b3-4e97-8431-c543f0822e61" containerName="cinder-scheduler" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.485327 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b110f494-d9b3-4e97-8431-c543f0822e61" containerName="cinder-scheduler" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.485564 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api-log" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.485593 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b110f494-d9b3-4e97-8431-c543f0822e61" containerName="probe" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.485605 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4cf1be0-51f7-4a5c-a792-c3bc504a8a98" containerName="barbican-api" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.485617 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b110f494-d9b3-4e97-8431-c543f0822e61" containerName="cinder-scheduler" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.486551 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.489219 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.493334 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.646258 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.646355 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2fhp\" (UniqueName: \"kubernetes.io/projected/f8f7c6df-f16d-487d-98e8-cab2c89073b4-kube-api-access-m2fhp\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.646586 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.646638 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.646696 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-scripts\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.646746 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8f7c6df-f16d-487d-98e8-cab2c89073b4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.747860 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-scripts\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.747926 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8f7c6df-f16d-487d-98e8-cab2c89073b4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.748002 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.748042 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2fhp\" (UniqueName: \"kubernetes.io/projected/f8f7c6df-f16d-487d-98e8-cab2c89073b4-kube-api-access-m2fhp\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.748066 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8f7c6df-f16d-487d-98e8-cab2c89073b4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.748185 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.748219 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.759231 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.765438 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-scripts\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.767114 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.772259 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.773015 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2fhp\" (UniqueName: \"kubernetes.io/projected/f8f7c6df-f16d-487d-98e8-cab2c89073b4-kube-api-access-m2fhp\") pod \"cinder-scheduler-0\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.905877 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 07:07:57 crc kubenswrapper[4978]: I0225 07:07:57.971007 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.159853 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-ovndb-tls-certs\") pod \"5ad94c00-0a28-4933-ad5c-27966cc8b437\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.159934 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-internal-tls-certs\") pod \"5ad94c00-0a28-4933-ad5c-27966cc8b437\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.159972 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-public-tls-certs\") pod \"5ad94c00-0a28-4933-ad5c-27966cc8b437\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.160063 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-config\") pod \"5ad94c00-0a28-4933-ad5c-27966cc8b437\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.160118 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-httpd-config\") pod \"5ad94c00-0a28-4933-ad5c-27966cc8b437\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.160138 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gsf8\" (UniqueName: \"kubernetes.io/projected/5ad94c00-0a28-4933-ad5c-27966cc8b437-kube-api-access-4gsf8\") pod \"5ad94c00-0a28-4933-ad5c-27966cc8b437\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.160202 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-combined-ca-bundle\") pod \"5ad94c00-0a28-4933-ad5c-27966cc8b437\" (UID: \"5ad94c00-0a28-4933-ad5c-27966cc8b437\") " Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.168573 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5ad94c00-0a28-4933-ad5c-27966cc8b437" (UID: "5ad94c00-0a28-4933-ad5c-27966cc8b437"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.187111 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad94c00-0a28-4933-ad5c-27966cc8b437-kube-api-access-4gsf8" (OuterVolumeSpecName: "kube-api-access-4gsf8") pod "5ad94c00-0a28-4933-ad5c-27966cc8b437" (UID: "5ad94c00-0a28-4933-ad5c-27966cc8b437"). InnerVolumeSpecName "kube-api-access-4gsf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.235952 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-config" (OuterVolumeSpecName: "config") pod "5ad94c00-0a28-4933-ad5c-27966cc8b437" (UID: "5ad94c00-0a28-4933-ad5c-27966cc8b437"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.236201 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5ad94c00-0a28-4933-ad5c-27966cc8b437" (UID: "5ad94c00-0a28-4933-ad5c-27966cc8b437"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.239337 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ad94c00-0a28-4933-ad5c-27966cc8b437" (UID: "5ad94c00-0a28-4933-ad5c-27966cc8b437"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.249458 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5ad94c00-0a28-4933-ad5c-27966cc8b437" (UID: "5ad94c00-0a28-4933-ad5c-27966cc8b437"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.264832 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.264862 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.264872 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gsf8\" (UniqueName: \"kubernetes.io/projected/5ad94c00-0a28-4933-ad5c-27966cc8b437-kube-api-access-4gsf8\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.264883 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.264891 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.264900 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.273773 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5ad94c00-0a28-4933-ad5c-27966cc8b437" (UID: "5ad94c00-0a28-4933-ad5c-27966cc8b437"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.368949 4978 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ad94c00-0a28-4933-ad5c-27966cc8b437-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.372259 4978 generic.go:334] "Generic (PLEG): container finished" podID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerID="6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a" exitCode=0 Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.372387 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5987f7d87c-fxhz5" event={"ID":"5ad94c00-0a28-4933-ad5c-27966cc8b437","Type":"ContainerDied","Data":"6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a"} Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.372354 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5987f7d87c-fxhz5" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.372433 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5987f7d87c-fxhz5" event={"ID":"5ad94c00-0a28-4933-ad5c-27966cc8b437","Type":"ContainerDied","Data":"b86d30a69e41ee0a7893929787a3b26e8f0ad417dc776d5a7c7325639399e4c0"} Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.372455 4978 scope.go:117] "RemoveContainer" containerID="58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.413873 4978 scope.go:117] "RemoveContainer" containerID="6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.418423 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5987f7d87c-fxhz5"] Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.425758 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5987f7d87c-fxhz5"] Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.437629 4978 scope.go:117] "RemoveContainer" containerID="58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db" Feb 25 07:07:58 crc kubenswrapper[4978]: E0225 07:07:58.438161 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db\": container with ID starting with 58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db not found: ID does not exist" containerID="58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.438199 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db"} err="failed to get container status \"58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db\": rpc error: code = NotFound desc = could not find container \"58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db\": container with ID starting with 58f00bcaf581f3438e6a010b05484fd27107be8fae287f14234bb581f6a085db not found: ID does not exist" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.438240 4978 scope.go:117] "RemoveContainer" containerID="6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a" Feb 25 07:07:58 crc kubenswrapper[4978]: E0225 07:07:58.438516 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a\": container with ID starting with 6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a not found: ID does not exist" containerID="6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.438544 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a"} err="failed to get container status \"6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a\": rpc error: code = NotFound desc = could not find container \"6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a\": container with ID starting with 6afe87c57ef537fbc00c3253eb620bae55a026c26a0158715f2cbdc4a238fd6a not found: ID does not exist" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.454309 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.727892 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 25 07:07:58 crc kubenswrapper[4978]: E0225 07:07:58.728202 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerName="neutron-httpd" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.728219 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerName="neutron-httpd" Feb 25 07:07:58 crc kubenswrapper[4978]: E0225 07:07:58.728236 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerName="neutron-api" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.728242 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerName="neutron-api" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.729453 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerName="neutron-api" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.729473 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" containerName="neutron-httpd" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.730018 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.746226 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-fwr8j" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.746492 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.746666 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.754902 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.877550 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.877601 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config-secret\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.877630 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.877666 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcnx7\" (UniqueName: \"kubernetes.io/projected/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-kube-api-access-xcnx7\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.979230 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.979313 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config-secret\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.979351 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.979400 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcnx7\" (UniqueName: \"kubernetes.io/projected/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-kube-api-access-xcnx7\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.980846 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.991165 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config-secret\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:58 crc kubenswrapper[4978]: I0225 07:07:58.997197 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:59 crc kubenswrapper[4978]: I0225 07:07:59.009095 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcnx7\" (UniqueName: \"kubernetes.io/projected/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-kube-api-access-xcnx7\") pod \"openstackclient\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " pod="openstack/openstackclient" Feb 25 07:07:59 crc kubenswrapper[4978]: I0225 07:07:59.054964 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 07:07:59 crc kubenswrapper[4978]: I0225 07:07:59.366132 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad94c00-0a28-4933-ad5c-27966cc8b437" path="/var/lib/kubelet/pods/5ad94c00-0a28-4933-ad5c-27966cc8b437/volumes" Feb 25 07:07:59 crc kubenswrapper[4978]: I0225 07:07:59.367395 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b110f494-d9b3-4e97-8431-c543f0822e61" path="/var/lib/kubelet/pods/b110f494-d9b3-4e97-8431-c543f0822e61/volumes" Feb 25 07:07:59 crc kubenswrapper[4978]: I0225 07:07:59.388820 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8f7c6df-f16d-487d-98e8-cab2c89073b4","Type":"ContainerStarted","Data":"abc550e17c1b943bad86597853b93ec464de90bcf8090229b1e9cb5d9aeca0d7"} Feb 25 07:07:59 crc kubenswrapper[4978]: I0225 07:07:59.388866 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8f7c6df-f16d-487d-98e8-cab2c89073b4","Type":"ContainerStarted","Data":"097ec97b849c38518ace2fc78c37ad724fd0b060ecec088a0e316d7fa0266f50"} Feb 25 07:07:59 crc kubenswrapper[4978]: I0225 07:07:59.509241 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 07:07:59 crc kubenswrapper[4978]: I0225 07:07:59.524259 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.142637 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533388-rqj86"] Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.144152 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533388-rqj86" Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.147024 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.148473 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.151518 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533388-rqj86"] Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.152140 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.300910 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk54g\" (UniqueName: \"kubernetes.io/projected/a77f5934-71ee-4680-979b-94d1e4b9feef-kube-api-access-jk54g\") pod \"auto-csr-approver-29533388-rqj86\" (UID: \"a77f5934-71ee-4680-979b-94d1e4b9feef\") " pod="openshift-infra/auto-csr-approver-29533388-rqj86" Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.400700 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8f7c6df-f16d-487d-98e8-cab2c89073b4","Type":"ContainerStarted","Data":"8a5993b793c50295987a7b0055fc564f776c502b1b808c0832ceeaaacca5e945"} Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.401606 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ecf946d2-e6a8-4df6-ab76-b53d03db5d27","Type":"ContainerStarted","Data":"6fcc0e012b5a0f1402cdb18a9c9285f3fb41805ce4770cc47d37ddab6c260fc9"} Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.402380 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk54g\" (UniqueName: \"kubernetes.io/projected/a77f5934-71ee-4680-979b-94d1e4b9feef-kube-api-access-jk54g\") pod \"auto-csr-approver-29533388-rqj86\" (UID: \"a77f5934-71ee-4680-979b-94d1e4b9feef\") " pod="openshift-infra/auto-csr-approver-29533388-rqj86" Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.430134 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk54g\" (UniqueName: \"kubernetes.io/projected/a77f5934-71ee-4680-979b-94d1e4b9feef-kube-api-access-jk54g\") pod \"auto-csr-approver-29533388-rqj86\" (UID: \"a77f5934-71ee-4680-979b-94d1e4b9feef\") " pod="openshift-infra/auto-csr-approver-29533388-rqj86" Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.453855 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.4538360470000002 podStartE2EDuration="3.453836047s" podCreationTimestamp="2026-02-25 07:07:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:00.423276185 +0000 UTC m=+1373.862532654" watchObservedRunningTime="2026-02-25 07:08:00.453836047 +0000 UTC m=+1373.893092506" Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.463020 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533388-rqj86" Feb 25 07:08:00 crc kubenswrapper[4978]: I0225 07:08:00.968685 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533388-rqj86"] Feb 25 07:08:00 crc kubenswrapper[4978]: W0225 07:08:00.980355 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda77f5934_71ee_4680_979b_94d1e4b9feef.slice/crio-8ed98b3fc7d9cae390c6c61674fec9866f42d2f8dc4b308c53e4dfe78d43c2d8 WatchSource:0}: Error finding container 8ed98b3fc7d9cae390c6c61674fec9866f42d2f8dc4b308c53e4dfe78d43c2d8: Status 404 returned error can't find the container with id 8ed98b3fc7d9cae390c6c61674fec9866f42d2f8dc4b308c53e4dfe78d43c2d8 Feb 25 07:08:01 crc kubenswrapper[4978]: I0225 07:08:01.413415 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533388-rqj86" event={"ID":"a77f5934-71ee-4680-979b-94d1e4b9feef","Type":"ContainerStarted","Data":"8ed98b3fc7d9cae390c6c61674fec9866f42d2f8dc4b308c53e4dfe78d43c2d8"} Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.421333 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533388-rqj86" event={"ID":"a77f5934-71ee-4680-979b-94d1e4b9feef","Type":"ContainerStarted","Data":"fdd96869b4fb12af351379969d60b7fe078542fa9ee866fff1cb955e15ffee92"} Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.438772 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533388-rqj86" podStartSLOduration=1.6278205890000002 podStartE2EDuration="2.438755395s" podCreationTimestamp="2026-02-25 07:08:00 +0000 UTC" firstStartedPulling="2026-02-25 07:08:00.982875969 +0000 UTC m=+1374.422132428" lastFinishedPulling="2026-02-25 07:08:01.793810775 +0000 UTC m=+1375.233067234" observedRunningTime="2026-02-25 07:08:02.433128518 +0000 UTC m=+1375.872384977" watchObservedRunningTime="2026-02-25 07:08:02.438755395 +0000 UTC m=+1375.878011854" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.793849 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-785ccc58d9-p75wv"] Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.808548 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.816451 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.816909 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.835268 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.842728 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-785ccc58d9-p75wv"] Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.906510 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.952926 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd789\" (UniqueName: \"kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-kube-api-access-hd789\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.953187 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-run-httpd\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.953318 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-etc-swift\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.953478 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-public-tls-certs\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.953584 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-config-data\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.953758 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-internal-tls-certs\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.954130 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-log-httpd\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:02 crc kubenswrapper[4978]: I0225 07:08:02.954348 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-combined-ca-bundle\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.055622 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-public-tls-certs\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.055662 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-config-data\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.055722 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-internal-tls-certs\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.055760 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-log-httpd\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.055777 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-combined-ca-bundle\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.055821 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd789\" (UniqueName: \"kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-kube-api-access-hd789\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.055864 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-run-httpd\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.055902 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-etc-swift\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.057643 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-run-httpd\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.059033 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-log-httpd\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.063194 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-public-tls-certs\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.067148 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-internal-tls-certs\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.068733 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-config-data\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.073675 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-etc-swift\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.074419 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd789\" (UniqueName: \"kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-kube-api-access-hd789\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.075188 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-combined-ca-bundle\") pod \"swift-proxy-785ccc58d9-p75wv\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.135025 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.430806 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.431586 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="ceilometer-central-agent" containerID="cri-o://d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3" gracePeriod=30 Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.432609 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="sg-core" containerID="cri-o://0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9" gracePeriod=30 Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.432683 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="proxy-httpd" containerID="cri-o://45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2" gracePeriod=30 Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.432746 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="ceilometer-notification-agent" containerID="cri-o://24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc" gracePeriod=30 Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.447008 4978 generic.go:334] "Generic (PLEG): container finished" podID="a77f5934-71ee-4680-979b-94d1e4b9feef" containerID="fdd96869b4fb12af351379969d60b7fe078542fa9ee866fff1cb955e15ffee92" exitCode=0 Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.447058 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533388-rqj86" event={"ID":"a77f5934-71ee-4680-979b-94d1e4b9feef","Type":"ContainerDied","Data":"fdd96869b4fb12af351379969d60b7fe078542fa9ee866fff1cb955e15ffee92"} Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.540572 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.168:3000/\": read tcp 10.217.0.2:58660->10.217.0.168:3000: read: connection reset by peer" Feb 25 07:08:03 crc kubenswrapper[4978]: I0225 07:08:03.718503 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-785ccc58d9-p75wv"] Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.344244 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471584 4978 generic.go:334] "Generic (PLEG): container finished" podID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerID="45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2" exitCode=0 Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471619 4978 generic.go:334] "Generic (PLEG): container finished" podID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerID="0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9" exitCode=2 Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471627 4978 generic.go:334] "Generic (PLEG): container finished" podID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerID="24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc" exitCode=0 Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471634 4978 generic.go:334] "Generic (PLEG): container finished" podID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerID="d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3" exitCode=0 Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471635 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cc903177-5cd3-489b-8a04-fe99e6a2d7fe","Type":"ContainerDied","Data":"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2"} Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471677 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cc903177-5cd3-489b-8a04-fe99e6a2d7fe","Type":"ContainerDied","Data":"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9"} Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471690 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cc903177-5cd3-489b-8a04-fe99e6a2d7fe","Type":"ContainerDied","Data":"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc"} Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471700 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cc903177-5cd3-489b-8a04-fe99e6a2d7fe","Type":"ContainerDied","Data":"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3"} Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471686 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471718 4978 scope.go:117] "RemoveContainer" containerID="45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.471709 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cc903177-5cd3-489b-8a04-fe99e6a2d7fe","Type":"ContainerDied","Data":"fd0e7b7fdeecbd3c9507423d95b83c90e7741fb0a6dce5dad5434d463237d380"} Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.475273 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-785ccc58d9-p75wv" event={"ID":"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0","Type":"ContainerStarted","Data":"d9e6c02d47ab88e1a99d79a917daf7a487b0b7f226c09ae65c011a4c5088f429"} Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.475322 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-785ccc58d9-p75wv" event={"ID":"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0","Type":"ContainerStarted","Data":"e0503e6f2f40da27125f61fc3830d5697085e44e0e23aa2a64cee749b3cf958e"} Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.475333 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-785ccc58d9-p75wv" event={"ID":"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0","Type":"ContainerStarted","Data":"fdf036e6392d60dcd9126e8debfaca349560b2793a042d255bcad5aac174a529"} Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.475404 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.485807 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-run-httpd\") pod \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.485848 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-log-httpd\") pod \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.485949 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-sg-core-conf-yaml\") pod \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.486022 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-config-data\") pod \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.486104 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-scripts\") pod \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.486173 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64zr7\" (UniqueName: \"kubernetes.io/projected/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-kube-api-access-64zr7\") pod \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.486189 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-combined-ca-bundle\") pod \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\" (UID: \"cc903177-5cd3-489b-8a04-fe99e6a2d7fe\") " Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.486488 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cc903177-5cd3-489b-8a04-fe99e6a2d7fe" (UID: "cc903177-5cd3-489b-8a04-fe99e6a2d7fe"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.486569 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.486915 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cc903177-5cd3-489b-8a04-fe99e6a2d7fe" (UID: "cc903177-5cd3-489b-8a04-fe99e6a2d7fe"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.490422 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-kube-api-access-64zr7" (OuterVolumeSpecName: "kube-api-access-64zr7") pod "cc903177-5cd3-489b-8a04-fe99e6a2d7fe" (UID: "cc903177-5cd3-489b-8a04-fe99e6a2d7fe"). InnerVolumeSpecName "kube-api-access-64zr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.502453 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-scripts" (OuterVolumeSpecName: "scripts") pod "cc903177-5cd3-489b-8a04-fe99e6a2d7fe" (UID: "cc903177-5cd3-489b-8a04-fe99e6a2d7fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.505038 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-785ccc58d9-p75wv" podStartSLOduration=2.505022075 podStartE2EDuration="2.505022075s" podCreationTimestamp="2026-02-25 07:08:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:04.49377586 +0000 UTC m=+1377.933032319" watchObservedRunningTime="2026-02-25 07:08:04.505022075 +0000 UTC m=+1377.944278534" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.552033 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cc903177-5cd3-489b-8a04-fe99e6a2d7fe" (UID: "cc903177-5cd3-489b-8a04-fe99e6a2d7fe"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.578170 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc903177-5cd3-489b-8a04-fe99e6a2d7fe" (UID: "cc903177-5cd3-489b-8a04-fe99e6a2d7fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.588203 4978 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.588233 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.588243 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64zr7\" (UniqueName: \"kubernetes.io/projected/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-kube-api-access-64zr7\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.588254 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.588264 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.611728 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-config-data" (OuterVolumeSpecName: "config-data") pod "cc903177-5cd3-489b-8a04-fe99e6a2d7fe" (UID: "cc903177-5cd3-489b-8a04-fe99e6a2d7fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.689920 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc903177-5cd3-489b-8a04-fe99e6a2d7fe-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.804439 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.824017 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.833587 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:04 crc kubenswrapper[4978]: E0225 07:08:04.834091 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="proxy-httpd" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.834114 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="proxy-httpd" Feb 25 07:08:04 crc kubenswrapper[4978]: E0225 07:08:04.834167 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="sg-core" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.834181 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="sg-core" Feb 25 07:08:04 crc kubenswrapper[4978]: E0225 07:08:04.834198 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="ceilometer-central-agent" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.834210 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="ceilometer-central-agent" Feb 25 07:08:04 crc kubenswrapper[4978]: E0225 07:08:04.834233 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="ceilometer-notification-agent" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.834242 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="ceilometer-notification-agent" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.834485 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="ceilometer-central-agent" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.834512 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="ceilometer-notification-agent" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.834538 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="proxy-httpd" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.834550 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" containerName="sg-core" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.836707 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.840687 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.844421 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.844909 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.996965 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-config-data\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.997008 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-log-httpd\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.997042 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.997298 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.997327 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-run-httpd\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.997359 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4bnh\" (UniqueName: \"kubernetes.io/projected/e22ec80b-2066-4394-884c-7ce80ac4061b-kube-api-access-b4bnh\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:04 crc kubenswrapper[4978]: I0225 07:08:04.997391 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-scripts\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.112601 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-config-data\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.112669 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-log-httpd\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.112696 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.112748 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.112775 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-run-httpd\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.112802 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4bnh\" (UniqueName: \"kubernetes.io/projected/e22ec80b-2066-4394-884c-7ce80ac4061b-kube-api-access-b4bnh\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.112827 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-scripts\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.113195 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-log-httpd\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.113601 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-run-httpd\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.118020 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-config-data\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.118318 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-scripts\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.120476 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.121040 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.136120 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4bnh\" (UniqueName: \"kubernetes.io/projected/e22ec80b-2066-4394-884c-7ce80ac4061b-kube-api-access-b4bnh\") pod \"ceilometer-0\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.155988 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.344431 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc903177-5cd3-489b-8a04-fe99e6a2d7fe" path="/var/lib/kubelet/pods/cc903177-5cd3-489b-8a04-fe99e6a2d7fe/volumes" Feb 25 07:08:05 crc kubenswrapper[4978]: I0225 07:08:05.484352 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.672016 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-xq9mn"] Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.673344 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xq9mn" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.686151 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xq9mn"] Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.778416 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-wxrhg"] Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.779631 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wxrhg" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.812474 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wxrhg"] Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.842126 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8f2x\" (UniqueName: \"kubernetes.io/projected/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-kube-api-access-k8f2x\") pod \"nova-api-db-create-xq9mn\" (UID: \"f5a1b5fa-aba9-4232-b5b7-96365cebef1a\") " pod="openstack/nova-api-db-create-xq9mn" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.842345 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-operator-scripts\") pod \"nova-api-db-create-xq9mn\" (UID: \"f5a1b5fa-aba9-4232-b5b7-96365cebef1a\") " pod="openstack/nova-api-db-create-xq9mn" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.881562 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-9373-account-create-update-c5bk6"] Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.882694 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9373-account-create-update-c5bk6" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.884945 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.896722 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9373-account-create-update-c5bk6"] Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.944145 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-operator-scripts\") pod \"nova-api-db-create-xq9mn\" (UID: \"f5a1b5fa-aba9-4232-b5b7-96365cebef1a\") " pod="openstack/nova-api-db-create-xq9mn" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.944209 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8f2x\" (UniqueName: \"kubernetes.io/projected/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-kube-api-access-k8f2x\") pod \"nova-api-db-create-xq9mn\" (UID: \"f5a1b5fa-aba9-4232-b5b7-96365cebef1a\") " pod="openstack/nova-api-db-create-xq9mn" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.944259 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be1ecfff-67b5-4cc5-90f0-d833276a5b62-operator-scripts\") pod \"nova-cell0-db-create-wxrhg\" (UID: \"be1ecfff-67b5-4cc5-90f0-d833276a5b62\") " pod="openstack/nova-cell0-db-create-wxrhg" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.944311 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df8s9\" (UniqueName: \"kubernetes.io/projected/be1ecfff-67b5-4cc5-90f0-d833276a5b62-kube-api-access-df8s9\") pod \"nova-cell0-db-create-wxrhg\" (UID: \"be1ecfff-67b5-4cc5-90f0-d833276a5b62\") " pod="openstack/nova-cell0-db-create-wxrhg" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.945561 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-operator-scripts\") pod \"nova-api-db-create-xq9mn\" (UID: \"f5a1b5fa-aba9-4232-b5b7-96365cebef1a\") " pod="openstack/nova-api-db-create-xq9mn" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.973971 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8f2x\" (UniqueName: \"kubernetes.io/projected/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-kube-api-access-k8f2x\") pod \"nova-api-db-create-xq9mn\" (UID: \"f5a1b5fa-aba9-4232-b5b7-96365cebef1a\") " pod="openstack/nova-api-db-create-xq9mn" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.985427 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-swr2c"] Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.986979 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-swr2c" Feb 25 07:08:06 crc kubenswrapper[4978]: I0225 07:08:06.992770 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-swr2c"] Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.000437 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xq9mn" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.046040 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prbdq\" (UniqueName: \"kubernetes.io/projected/a872e452-e198-4c3d-a8d4-85473516feab-kube-api-access-prbdq\") pod \"nova-api-9373-account-create-update-c5bk6\" (UID: \"a872e452-e198-4c3d-a8d4-85473516feab\") " pod="openstack/nova-api-9373-account-create-update-c5bk6" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.046146 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be1ecfff-67b5-4cc5-90f0-d833276a5b62-operator-scripts\") pod \"nova-cell0-db-create-wxrhg\" (UID: \"be1ecfff-67b5-4cc5-90f0-d833276a5b62\") " pod="openstack/nova-cell0-db-create-wxrhg" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.046183 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a872e452-e198-4c3d-a8d4-85473516feab-operator-scripts\") pod \"nova-api-9373-account-create-update-c5bk6\" (UID: \"a872e452-e198-4c3d-a8d4-85473516feab\") " pod="openstack/nova-api-9373-account-create-update-c5bk6" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.046226 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df8s9\" (UniqueName: \"kubernetes.io/projected/be1ecfff-67b5-4cc5-90f0-d833276a5b62-kube-api-access-df8s9\") pod \"nova-cell0-db-create-wxrhg\" (UID: \"be1ecfff-67b5-4cc5-90f0-d833276a5b62\") " pod="openstack/nova-cell0-db-create-wxrhg" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.046769 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be1ecfff-67b5-4cc5-90f0-d833276a5b62-operator-scripts\") pod \"nova-cell0-db-create-wxrhg\" (UID: \"be1ecfff-67b5-4cc5-90f0-d833276a5b62\") " pod="openstack/nova-cell0-db-create-wxrhg" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.065766 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df8s9\" (UniqueName: \"kubernetes.io/projected/be1ecfff-67b5-4cc5-90f0-d833276a5b62-kube-api-access-df8s9\") pod \"nova-cell0-db-create-wxrhg\" (UID: \"be1ecfff-67b5-4cc5-90f0-d833276a5b62\") " pod="openstack/nova-cell0-db-create-wxrhg" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.083634 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-b71c-account-create-update-jjh4g"] Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.088597 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.095497 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.096124 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b71c-account-create-update-jjh4g"] Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.099450 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wxrhg" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.148080 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a872e452-e198-4c3d-a8d4-85473516feab-operator-scripts\") pod \"nova-api-9373-account-create-update-c5bk6\" (UID: \"a872e452-e198-4c3d-a8d4-85473516feab\") " pod="openstack/nova-api-9373-account-create-update-c5bk6" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.148180 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w24cb\" (UniqueName: \"kubernetes.io/projected/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-kube-api-access-w24cb\") pod \"nova-cell1-db-create-swr2c\" (UID: \"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45\") " pod="openstack/nova-cell1-db-create-swr2c" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.148237 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-operator-scripts\") pod \"nova-cell1-db-create-swr2c\" (UID: \"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45\") " pod="openstack/nova-cell1-db-create-swr2c" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.148276 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prbdq\" (UniqueName: \"kubernetes.io/projected/a872e452-e198-4c3d-a8d4-85473516feab-kube-api-access-prbdq\") pod \"nova-api-9373-account-create-update-c5bk6\" (UID: \"a872e452-e198-4c3d-a8d4-85473516feab\") " pod="openstack/nova-api-9373-account-create-update-c5bk6" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.155123 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a872e452-e198-4c3d-a8d4-85473516feab-operator-scripts\") pod \"nova-api-9373-account-create-update-c5bk6\" (UID: \"a872e452-e198-4c3d-a8d4-85473516feab\") " pod="openstack/nova-api-9373-account-create-update-c5bk6" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.189310 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prbdq\" (UniqueName: \"kubernetes.io/projected/a872e452-e198-4c3d-a8d4-85473516feab-kube-api-access-prbdq\") pod \"nova-api-9373-account-create-update-c5bk6\" (UID: \"a872e452-e198-4c3d-a8d4-85473516feab\") " pod="openstack/nova-api-9373-account-create-update-c5bk6" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.199184 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9373-account-create-update-c5bk6" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.250838 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-operator-scripts\") pod \"nova-cell1-db-create-swr2c\" (UID: \"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45\") " pod="openstack/nova-cell1-db-create-swr2c" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.250911 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-operator-scripts\") pod \"nova-cell0-b71c-account-create-update-jjh4g\" (UID: \"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1\") " pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.251089 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9mvl\" (UniqueName: \"kubernetes.io/projected/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-kube-api-access-m9mvl\") pod \"nova-cell0-b71c-account-create-update-jjh4g\" (UID: \"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1\") " pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.251127 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w24cb\" (UniqueName: \"kubernetes.io/projected/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-kube-api-access-w24cb\") pod \"nova-cell1-db-create-swr2c\" (UID: \"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45\") " pod="openstack/nova-cell1-db-create-swr2c" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.254860 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-operator-scripts\") pod \"nova-cell1-db-create-swr2c\" (UID: \"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45\") " pod="openstack/nova-cell1-db-create-swr2c" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.269572 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w24cb\" (UniqueName: \"kubernetes.io/projected/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-kube-api-access-w24cb\") pod \"nova-cell1-db-create-swr2c\" (UID: \"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45\") " pod="openstack/nova-cell1-db-create-swr2c" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.278346 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ac49-account-create-update-2rrpr"] Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.279992 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.284924 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.324208 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ac49-account-create-update-2rrpr"] Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.332929 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-swr2c" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.353821 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9mvl\" (UniqueName: \"kubernetes.io/projected/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-kube-api-access-m9mvl\") pod \"nova-cell0-b71c-account-create-update-jjh4g\" (UID: \"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1\") " pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.354315 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-operator-scripts\") pod \"nova-cell0-b71c-account-create-update-jjh4g\" (UID: \"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1\") " pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.355036 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-operator-scripts\") pod \"nova-cell0-b71c-account-create-update-jjh4g\" (UID: \"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1\") " pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.376564 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9mvl\" (UniqueName: \"kubernetes.io/projected/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-kube-api-access-m9mvl\") pod \"nova-cell0-b71c-account-create-update-jjh4g\" (UID: \"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1\") " pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.430696 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.457759 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-operator-scripts\") pod \"nova-cell1-ac49-account-create-update-2rrpr\" (UID: \"b5c931ba-06fd-41eb-9bdc-514ec3faef6e\") " pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.457920 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6tnr\" (UniqueName: \"kubernetes.io/projected/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-kube-api-access-v6tnr\") pod \"nova-cell1-ac49-account-create-update-2rrpr\" (UID: \"b5c931ba-06fd-41eb-9bdc-514ec3faef6e\") " pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.559830 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6tnr\" (UniqueName: \"kubernetes.io/projected/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-kube-api-access-v6tnr\") pod \"nova-cell1-ac49-account-create-update-2rrpr\" (UID: \"b5c931ba-06fd-41eb-9bdc-514ec3faef6e\") " pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.559916 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-operator-scripts\") pod \"nova-cell1-ac49-account-create-update-2rrpr\" (UID: \"b5c931ba-06fd-41eb-9bdc-514ec3faef6e\") " pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.560538 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-operator-scripts\") pod \"nova-cell1-ac49-account-create-update-2rrpr\" (UID: \"b5c931ba-06fd-41eb-9bdc-514ec3faef6e\") " pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.578706 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6tnr\" (UniqueName: \"kubernetes.io/projected/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-kube-api-access-v6tnr\") pod \"nova-cell1-ac49-account-create-update-2rrpr\" (UID: \"b5c931ba-06fd-41eb-9bdc-514ec3faef6e\") " pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" Feb 25 07:08:07 crc kubenswrapper[4978]: I0225 07:08:07.631145 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" Feb 25 07:08:08 crc kubenswrapper[4978]: I0225 07:08:08.155088 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:08 crc kubenswrapper[4978]: I0225 07:08:08.215059 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 25 07:08:09 crc kubenswrapper[4978]: I0225 07:08:09.278844 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:08:09 crc kubenswrapper[4978]: I0225 07:08:09.280345 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="667d3386-e724-42f3-940b-82af7780255b" containerName="glance-log" containerID="cri-o://32a5d429a99cd8a6f7e1deeb90aaac504efbba685415e1cd920ca02ab313cf16" gracePeriod=30 Feb 25 07:08:09 crc kubenswrapper[4978]: I0225 07:08:09.280579 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="667d3386-e724-42f3-940b-82af7780255b" containerName="glance-httpd" containerID="cri-o://ef78e19bd4d46fac7902200f55b0093cbedffb4f12c8d2a96d94765f2b3f1e48" gracePeriod=30 Feb 25 07:08:09 crc kubenswrapper[4978]: E0225 07:08:09.337378 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod667d3386_e724_42f3_940b_82af7780255b.slice/crio-32a5d429a99cd8a6f7e1deeb90aaac504efbba685415e1cd920ca02ab313cf16.scope\": RecentStats: unable to find data in memory cache]" Feb 25 07:08:09 crc kubenswrapper[4978]: I0225 07:08:09.538108 4978 generic.go:334] "Generic (PLEG): container finished" podID="667d3386-e724-42f3-940b-82af7780255b" containerID="32a5d429a99cd8a6f7e1deeb90aaac504efbba685415e1cd920ca02ab313cf16" exitCode=143 Feb 25 07:08:09 crc kubenswrapper[4978]: I0225 07:08:09.538166 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"667d3386-e724-42f3-940b-82af7780255b","Type":"ContainerDied","Data":"32a5d429a99cd8a6f7e1deeb90aaac504efbba685415e1cd920ca02ab313cf16"} Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.521658 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.676603 4978 scope.go:117] "RemoveContainer" containerID="0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.823910 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533388-rqj86" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.833278 4978 scope.go:117] "RemoveContainer" containerID="24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.883975 4978 scope.go:117] "RemoveContainer" containerID="d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.954999 4978 scope.go:117] "RemoveContainer" containerID="45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2" Feb 25 07:08:10 crc kubenswrapper[4978]: E0225 07:08:10.959763 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2\": container with ID starting with 45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2 not found: ID does not exist" containerID="45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.959798 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2"} err="failed to get container status \"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2\": rpc error: code = NotFound desc = could not find container \"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2\": container with ID starting with 45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.959822 4978 scope.go:117] "RemoveContainer" containerID="0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9" Feb 25 07:08:10 crc kubenswrapper[4978]: E0225 07:08:10.965138 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9\": container with ID starting with 0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9 not found: ID does not exist" containerID="0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.965364 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9"} err="failed to get container status \"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9\": rpc error: code = NotFound desc = could not find container \"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9\": container with ID starting with 0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.965394 4978 scope.go:117] "RemoveContainer" containerID="24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc" Feb 25 07:08:10 crc kubenswrapper[4978]: E0225 07:08:10.965843 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc\": container with ID starting with 24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc not found: ID does not exist" containerID="24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.965884 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc"} err="failed to get container status \"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc\": rpc error: code = NotFound desc = could not find container \"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc\": container with ID starting with 24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.965910 4978 scope.go:117] "RemoveContainer" containerID="d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3" Feb 25 07:08:10 crc kubenswrapper[4978]: E0225 07:08:10.966364 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3\": container with ID starting with d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3 not found: ID does not exist" containerID="d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.966395 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3"} err="failed to get container status \"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3\": rpc error: code = NotFound desc = could not find container \"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3\": container with ID starting with d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.966485 4978 scope.go:117] "RemoveContainer" containerID="45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.966750 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2"} err="failed to get container status \"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2\": rpc error: code = NotFound desc = could not find container \"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2\": container with ID starting with 45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.966778 4978 scope.go:117] "RemoveContainer" containerID="0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.966999 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9"} err="failed to get container status \"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9\": rpc error: code = NotFound desc = could not find container \"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9\": container with ID starting with 0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.967020 4978 scope.go:117] "RemoveContainer" containerID="24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.967210 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc"} err="failed to get container status \"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc\": rpc error: code = NotFound desc = could not find container \"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc\": container with ID starting with 24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.967228 4978 scope.go:117] "RemoveContainer" containerID="d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.967494 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3"} err="failed to get container status \"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3\": rpc error: code = NotFound desc = could not find container \"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3\": container with ID starting with d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.967514 4978 scope.go:117] "RemoveContainer" containerID="45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.967701 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2"} err="failed to get container status \"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2\": rpc error: code = NotFound desc = could not find container \"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2\": container with ID starting with 45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.967724 4978 scope.go:117] "RemoveContainer" containerID="0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.967877 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9"} err="failed to get container status \"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9\": rpc error: code = NotFound desc = could not find container \"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9\": container with ID starting with 0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.967893 4978 scope.go:117] "RemoveContainer" containerID="24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.968032 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc"} err="failed to get container status \"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc\": rpc error: code = NotFound desc = could not find container \"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc\": container with ID starting with 24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.968049 4978 scope.go:117] "RemoveContainer" containerID="d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.968231 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3"} err="failed to get container status \"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3\": rpc error: code = NotFound desc = could not find container \"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3\": container with ID starting with d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.968246 4978 scope.go:117] "RemoveContainer" containerID="45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.968531 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2"} err="failed to get container status \"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2\": rpc error: code = NotFound desc = could not find container \"45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2\": container with ID starting with 45832527b7dcbdecf613f429dc82087d212a9e75b78b6363433ac5ace19cb7e2 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.968547 4978 scope.go:117] "RemoveContainer" containerID="0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.968845 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9"} err="failed to get container status \"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9\": rpc error: code = NotFound desc = could not find container \"0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9\": container with ID starting with 0e705668a98184ecdac6c9f654cbff15328aed6ec10cb2853000260ceb1667b9 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.968861 4978 scope.go:117] "RemoveContainer" containerID="24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.972390 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc"} err="failed to get container status \"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc\": rpc error: code = NotFound desc = could not find container \"24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc\": container with ID starting with 24b1f2e5f6b20b4712f08cbb01ea9dda565cc3f0d14918f278cff49cc44265cc not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.972443 4978 scope.go:117] "RemoveContainer" containerID="d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.977675 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3"} err="failed to get container status \"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3\": rpc error: code = NotFound desc = could not find container \"d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3\": container with ID starting with d79f10968501df89597cee61f8e3730fe4d2e9cb71507ea4d40490aab9f6f9f3 not found: ID does not exist" Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.990754 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk54g\" (UniqueName: \"kubernetes.io/projected/a77f5934-71ee-4680-979b-94d1e4b9feef-kube-api-access-jk54g\") pod \"a77f5934-71ee-4680-979b-94d1e4b9feef\" (UID: \"a77f5934-71ee-4680-979b-94d1e4b9feef\") " Feb 25 07:08:10 crc kubenswrapper[4978]: I0225 07:08:10.994885 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a77f5934-71ee-4680-979b-94d1e4b9feef-kube-api-access-jk54g" (OuterVolumeSpecName: "kube-api-access-jk54g") pod "a77f5934-71ee-4680-979b-94d1e4b9feef" (UID: "a77f5934-71ee-4680-979b-94d1e4b9feef"). InnerVolumeSpecName "kube-api-access-jk54g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.092507 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk54g\" (UniqueName: \"kubernetes.io/projected/a77f5934-71ee-4680-979b-94d1e4b9feef-kube-api-access-jk54g\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.201441 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ac49-account-create-update-2rrpr"] Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.290226 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.313952 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b71c-account-create-update-jjh4g"] Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.321621 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9373-account-create-update-c5bk6"] Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.342010 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 25 07:08:11 crc kubenswrapper[4978]: W0225 07:08:11.342625 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcb2bbd3_7d86_4854_ab81_1e7ce97e8eb1.slice/crio-7cb996eb0d602262740401026b55dd74fd392520f28c818f2838ccac872afa6f WatchSource:0}: Error finding container 7cb996eb0d602262740401026b55dd74fd392520f28c818f2838ccac872afa6f: Status 404 returned error can't find the container with id 7cb996eb0d602262740401026b55dd74fd392520f28c818f2838ccac872afa6f Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.348161 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.467775 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-xq9mn"] Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.481002 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wxrhg"] Feb 25 07:08:11 crc kubenswrapper[4978]: W0225 07:08:11.485116 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5a1b5fa_aba9_4232_b5b7_96365cebef1a.slice/crio-01efc90514dc135417a5f411d44d73bfd6dbcf72a183142847cb5dd1b8faead7 WatchSource:0}: Error finding container 01efc90514dc135417a5f411d44d73bfd6dbcf72a183142847cb5dd1b8faead7: Status 404 returned error can't find the container with id 01efc90514dc135417a5f411d44d73bfd6dbcf72a183142847cb5dd1b8faead7 Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.503353 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-swr2c"] Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.509490 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.509706 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" containerName="glance-log" containerID="cri-o://f68c2a7b016bdf0a6652e9324843ca4472b7b54fb5f39c860f5ab165234032df" gracePeriod=30 Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.510104 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" containerName="glance-httpd" containerID="cri-o://b8960ac3480ae1d88146d962f2420be146e18df6db23c90812cb66575bdf4679" gracePeriod=30 Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.565085 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e22ec80b-2066-4394-884c-7ce80ac4061b","Type":"ContainerStarted","Data":"742259adffa91e4771f76c8962bb03b0c5d6d089dfd8be471d9533c52986027a"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.571825 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" event={"ID":"b5c931ba-06fd-41eb-9bdc-514ec3faef6e","Type":"ContainerStarted","Data":"d09e81920a08e75b18da1c30ea69b3f15f59cf1ccbbf32b83855123548ed4933"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.571867 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" event={"ID":"b5c931ba-06fd-41eb-9bdc-514ec3faef6e","Type":"ContainerStarted","Data":"fa2deb6eff28e9c3f8f0ff41593d5ebb91b6729aaea332b4c878bd2a449b0afa"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.577651 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-swr2c" event={"ID":"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45","Type":"ContainerStarted","Data":"858cea1c6c917896b61718eb5064eb5f9428bda78ffdde3bce8a01f1dd72b21c"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.579936 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wxrhg" event={"ID":"be1ecfff-67b5-4cc5-90f0-d833276a5b62","Type":"ContainerStarted","Data":"8d6aaec4738f311d2ff97f035fb630eecb67f67829d51e56241fa4251f82b79d"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.581719 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xq9mn" event={"ID":"f5a1b5fa-aba9-4232-b5b7-96365cebef1a","Type":"ContainerStarted","Data":"01efc90514dc135417a5f411d44d73bfd6dbcf72a183142847cb5dd1b8faead7"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.589490 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533388-rqj86" Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.589493 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533388-rqj86" event={"ID":"a77f5934-71ee-4680-979b-94d1e4b9feef","Type":"ContainerDied","Data":"8ed98b3fc7d9cae390c6c61674fec9866f42d2f8dc4b308c53e4dfe78d43c2d8"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.589532 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ed98b3fc7d9cae390c6c61674fec9866f42d2f8dc4b308c53e4dfe78d43c2d8" Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.592420 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" event={"ID":"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1","Type":"ContainerStarted","Data":"61c18e6841b098b42dd54f54e6c6ae0ce96ad977d7e9e58e7eafa71aa797add1"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.592448 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" event={"ID":"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1","Type":"ContainerStarted","Data":"7cb996eb0d602262740401026b55dd74fd392520f28c818f2838ccac872afa6f"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.598813 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" podStartSLOduration=4.598797122 podStartE2EDuration="4.598797122s" podCreationTimestamp="2026-02-25 07:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:11.585871705 +0000 UTC m=+1385.025128184" watchObservedRunningTime="2026-02-25 07:08:11.598797122 +0000 UTC m=+1385.038053581" Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.599600 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ecf946d2-e6a8-4df6-ab76-b53d03db5d27","Type":"ContainerStarted","Data":"b043db8fffc821daf3cf627f0556f152620f9dbbc236549845734482a005f759"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.619922 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" podStartSLOduration=4.619903417 podStartE2EDuration="4.619903417s" podCreationTimestamp="2026-02-25 07:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:11.613084922 +0000 UTC m=+1385.052341381" watchObservedRunningTime="2026-02-25 07:08:11.619903417 +0000 UTC m=+1385.059159876" Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.638805 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9373-account-create-update-c5bk6" event={"ID":"a872e452-e198-4c3d-a8d4-85473516feab","Type":"ContainerStarted","Data":"8f5846d5b393fc3031a5997a87d3f1c794afd8b6ffa7e3d26a66204512976d76"} Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.646176 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.457648157 podStartE2EDuration="13.646152112s" podCreationTimestamp="2026-02-25 07:07:58 +0000 UTC" firstStartedPulling="2026-02-25 07:07:59.517093432 +0000 UTC m=+1372.956349891" lastFinishedPulling="2026-02-25 07:08:10.705597397 +0000 UTC m=+1384.144853846" observedRunningTime="2026-02-25 07:08:11.629967623 +0000 UTC m=+1385.069224082" watchObservedRunningTime="2026-02-25 07:08:11.646152112 +0000 UTC m=+1385.085408571" Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.658243 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-9373-account-create-update-c5bk6" podStartSLOduration=5.658224823 podStartE2EDuration="5.658224823s" podCreationTimestamp="2026-02-25 07:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:11.649120846 +0000 UTC m=+1385.088377295" watchObservedRunningTime="2026-02-25 07:08:11.658224823 +0000 UTC m=+1385.097481282" Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.913249 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533382-pvr9s"] Feb 25 07:08:11 crc kubenswrapper[4978]: I0225 07:08:11.928038 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533382-pvr9s"] Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.653156 4978 generic.go:334] "Generic (PLEG): container finished" podID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" containerID="f68c2a7b016bdf0a6652e9324843ca4472b7b54fb5f39c860f5ab165234032df" exitCode=143 Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.653312 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"09b73acd-bf19-49f5-b5e0-a60e180e9c1e","Type":"ContainerDied","Data":"f68c2a7b016bdf0a6652e9324843ca4472b7b54fb5f39c860f5ab165234032df"} Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.656786 4978 generic.go:334] "Generic (PLEG): container finished" podID="667d3386-e724-42f3-940b-82af7780255b" containerID="ef78e19bd4d46fac7902200f55b0093cbedffb4f12c8d2a96d94765f2b3f1e48" exitCode=0 Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.656841 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"667d3386-e724-42f3-940b-82af7780255b","Type":"ContainerDied","Data":"ef78e19bd4d46fac7902200f55b0093cbedffb4f12c8d2a96d94765f2b3f1e48"} Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.659891 4978 generic.go:334] "Generic (PLEG): container finished" podID="be1ecfff-67b5-4cc5-90f0-d833276a5b62" containerID="d795d32225060d81627ee5c37db3d8d00175f8bc43ea39f4af9cbf6719fd2309" exitCode=0 Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.659932 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wxrhg" event={"ID":"be1ecfff-67b5-4cc5-90f0-d833276a5b62","Type":"ContainerDied","Data":"d795d32225060d81627ee5c37db3d8d00175f8bc43ea39f4af9cbf6719fd2309"} Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.679146 4978 generic.go:334] "Generic (PLEG): container finished" podID="5791f28c-d7ba-4aa4-99d5-319ecbd6bc45" containerID="e0624efb2273218ddfa7b0599230cfa818769d0d5c002395bf7506f0625b968f" exitCode=0 Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.679210 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-swr2c" event={"ID":"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45","Type":"ContainerDied","Data":"e0624efb2273218ddfa7b0599230cfa818769d0d5c002395bf7506f0625b968f"} Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.698530 4978 generic.go:334] "Generic (PLEG): container finished" podID="bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1" containerID="61c18e6841b098b42dd54f54e6c6ae0ce96ad977d7e9e58e7eafa71aa797add1" exitCode=0 Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.698597 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" event={"ID":"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1","Type":"ContainerDied","Data":"61c18e6841b098b42dd54f54e6c6ae0ce96ad977d7e9e58e7eafa71aa797add1"} Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.702118 4978 generic.go:334] "Generic (PLEG): container finished" podID="a872e452-e198-4c3d-a8d4-85473516feab" containerID="09f6efb9225a38d05749f4a3b1a4a38d04bdd5f6081c1c1cc01280de268c9d65" exitCode=0 Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.702299 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9373-account-create-update-c5bk6" event={"ID":"a872e452-e198-4c3d-a8d4-85473516feab","Type":"ContainerDied","Data":"09f6efb9225a38d05749f4a3b1a4a38d04bdd5f6081c1c1cc01280de268c9d65"} Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.705162 4978 generic.go:334] "Generic (PLEG): container finished" podID="b5c931ba-06fd-41eb-9bdc-514ec3faef6e" containerID="d09e81920a08e75b18da1c30ea69b3f15f59cf1ccbbf32b83855123548ed4933" exitCode=0 Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.705233 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" event={"ID":"b5c931ba-06fd-41eb-9bdc-514ec3faef6e","Type":"ContainerDied","Data":"d09e81920a08e75b18da1c30ea69b3f15f59cf1ccbbf32b83855123548ed4933"} Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.708107 4978 generic.go:334] "Generic (PLEG): container finished" podID="f5a1b5fa-aba9-4232-b5b7-96365cebef1a" containerID="6f42f16d8d601f4d2968ebd172a034ee1578a4272e4611cdbe696202023e8aad" exitCode=0 Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.708213 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xq9mn" event={"ID":"f5a1b5fa-aba9-4232-b5b7-96365cebef1a","Type":"ContainerDied","Data":"6f42f16d8d601f4d2968ebd172a034ee1578a4272e4611cdbe696202023e8aad"} Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.712397 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e22ec80b-2066-4394-884c-7ce80ac4061b","Type":"ContainerStarted","Data":"8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8"} Feb 25 07:08:12 crc kubenswrapper[4978]: I0225 07:08:12.935174 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.040046 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-combined-ca-bundle\") pod \"667d3386-e724-42f3-940b-82af7780255b\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.040137 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-config-data\") pod \"667d3386-e724-42f3-940b-82af7780255b\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.040159 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-scripts\") pod \"667d3386-e724-42f3-940b-82af7780255b\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.040191 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-httpd-run\") pod \"667d3386-e724-42f3-940b-82af7780255b\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.040239 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-public-tls-certs\") pod \"667d3386-e724-42f3-940b-82af7780255b\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.040307 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-logs\") pod \"667d3386-e724-42f3-940b-82af7780255b\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.040339 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q45fb\" (UniqueName: \"kubernetes.io/projected/667d3386-e724-42f3-940b-82af7780255b-kube-api-access-q45fb\") pod \"667d3386-e724-42f3-940b-82af7780255b\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.040379 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"667d3386-e724-42f3-940b-82af7780255b\" (UID: \"667d3386-e724-42f3-940b-82af7780255b\") " Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.040985 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-logs" (OuterVolumeSpecName: "logs") pod "667d3386-e724-42f3-940b-82af7780255b" (UID: "667d3386-e724-42f3-940b-82af7780255b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.041181 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "667d3386-e724-42f3-940b-82af7780255b" (UID: "667d3386-e724-42f3-940b-82af7780255b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.047947 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-scripts" (OuterVolumeSpecName: "scripts") pod "667d3386-e724-42f3-940b-82af7780255b" (UID: "667d3386-e724-42f3-940b-82af7780255b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.048064 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/667d3386-e724-42f3-940b-82af7780255b-kube-api-access-q45fb" (OuterVolumeSpecName: "kube-api-access-q45fb") pod "667d3386-e724-42f3-940b-82af7780255b" (UID: "667d3386-e724-42f3-940b-82af7780255b"). InnerVolumeSpecName "kube-api-access-q45fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.049570 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "667d3386-e724-42f3-940b-82af7780255b" (UID: "667d3386-e724-42f3-940b-82af7780255b"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.085526 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "667d3386-e724-42f3-940b-82af7780255b" (UID: "667d3386-e724-42f3-940b-82af7780255b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.093181 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "667d3386-e724-42f3-940b-82af7780255b" (UID: "667d3386-e724-42f3-940b-82af7780255b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.101010 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-config-data" (OuterVolumeSpecName: "config-data") pod "667d3386-e724-42f3-940b-82af7780255b" (UID: "667d3386-e724-42f3-940b-82af7780255b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.142010 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.142042 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.142051 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.142059 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.142070 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/667d3386-e724-42f3-940b-82af7780255b-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.142078 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q45fb\" (UniqueName: \"kubernetes.io/projected/667d3386-e724-42f3-940b-82af7780255b-kube-api-access-q45fb\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.142112 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.142121 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/667d3386-e724-42f3-940b-82af7780255b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.143042 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.170769 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.243415 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.342928 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cc2f335-e592-465e-93e9-ce8ba085dca6" path="/var/lib/kubelet/pods/1cc2f335-e592-465e-93e9-ce8ba085dca6/volumes" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.720136 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e22ec80b-2066-4394-884c-7ce80ac4061b","Type":"ContainerStarted","Data":"6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11"} Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.720175 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e22ec80b-2066-4394-884c-7ce80ac4061b","Type":"ContainerStarted","Data":"acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b"} Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.722369 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.722888 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"667d3386-e724-42f3-940b-82af7780255b","Type":"ContainerDied","Data":"442edee2912d4eb077b7d9115fd12860b305dec6c483690d2c36dde5dc5ffcc1"} Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.722918 4978 scope.go:117] "RemoveContainer" containerID="ef78e19bd4d46fac7902200f55b0093cbedffb4f12c8d2a96d94765f2b3f1e48" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.743892 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.751010 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.768675 4978 scope.go:117] "RemoveContainer" containerID="32a5d429a99cd8a6f7e1deeb90aaac504efbba685415e1cd920ca02ab313cf16" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.794022 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:08:13 crc kubenswrapper[4978]: E0225 07:08:13.794462 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a77f5934-71ee-4680-979b-94d1e4b9feef" containerName="oc" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.794477 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a77f5934-71ee-4680-979b-94d1e4b9feef" containerName="oc" Feb 25 07:08:13 crc kubenswrapper[4978]: E0225 07:08:13.794501 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="667d3386-e724-42f3-940b-82af7780255b" containerName="glance-httpd" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.794507 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="667d3386-e724-42f3-940b-82af7780255b" containerName="glance-httpd" Feb 25 07:08:13 crc kubenswrapper[4978]: E0225 07:08:13.794519 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="667d3386-e724-42f3-940b-82af7780255b" containerName="glance-log" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.794527 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="667d3386-e724-42f3-940b-82af7780255b" containerName="glance-log" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.794714 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a77f5934-71ee-4680-979b-94d1e4b9feef" containerName="oc" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.794736 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="667d3386-e724-42f3-940b-82af7780255b" containerName="glance-httpd" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.794744 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="667d3386-e724-42f3-940b-82af7780255b" containerName="glance-log" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.795703 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.797792 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.798043 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.807884 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.958468 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.958901 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-logs\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.959075 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwr48\" (UniqueName: \"kubernetes.io/projected/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-kube-api-access-wwr48\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.959190 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.959296 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.959452 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.959755 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-scripts\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:13 crc kubenswrapper[4978]: I0225 07:08:13.959878 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-config-data\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.061700 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwr48\" (UniqueName: \"kubernetes.io/projected/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-kube-api-access-wwr48\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.061742 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.061773 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.061822 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.061839 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-scripts\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.061859 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-config-data\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.061881 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.061939 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-logs\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.062474 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-logs\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.062714 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.063769 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.068473 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-config-data\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.069191 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.069638 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-scripts\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.070058 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.078314 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwr48\" (UniqueName: \"kubernetes.io/projected/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-kube-api-access-wwr48\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.121800 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.191366 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wxrhg" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.347258 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-swr2c" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.353365 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.367106 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df8s9\" (UniqueName: \"kubernetes.io/projected/be1ecfff-67b5-4cc5-90f0-d833276a5b62-kube-api-access-df8s9\") pod \"be1ecfff-67b5-4cc5-90f0-d833276a5b62\" (UID: \"be1ecfff-67b5-4cc5-90f0-d833276a5b62\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.367466 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be1ecfff-67b5-4cc5-90f0-d833276a5b62-operator-scripts\") pod \"be1ecfff-67b5-4cc5-90f0-d833276a5b62\" (UID: \"be1ecfff-67b5-4cc5-90f0-d833276a5b62\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.368140 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be1ecfff-67b5-4cc5-90f0-d833276a5b62-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "be1ecfff-67b5-4cc5-90f0-d833276a5b62" (UID: "be1ecfff-67b5-4cc5-90f0-d833276a5b62"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.369128 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be1ecfff-67b5-4cc5-90f0-d833276a5b62-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.371163 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.373944 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be1ecfff-67b5-4cc5-90f0-d833276a5b62-kube-api-access-df8s9" (OuterVolumeSpecName: "kube-api-access-df8s9") pod "be1ecfff-67b5-4cc5-90f0-d833276a5b62" (UID: "be1ecfff-67b5-4cc5-90f0-d833276a5b62"). InnerVolumeSpecName "kube-api-access-df8s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.376591 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9373-account-create-update-c5bk6" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.408432 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xq9mn" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.416583 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.470994 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a872e452-e198-4c3d-a8d4-85473516feab-operator-scripts\") pod \"a872e452-e198-4c3d-a8d4-85473516feab\" (UID: \"a872e452-e198-4c3d-a8d4-85473516feab\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.471078 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-operator-scripts\") pod \"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1\" (UID: \"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.471120 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prbdq\" (UniqueName: \"kubernetes.io/projected/a872e452-e198-4c3d-a8d4-85473516feab-kube-api-access-prbdq\") pod \"a872e452-e198-4c3d-a8d4-85473516feab\" (UID: \"a872e452-e198-4c3d-a8d4-85473516feab\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.471147 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9mvl\" (UniqueName: \"kubernetes.io/projected/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-kube-api-access-m9mvl\") pod \"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1\" (UID: \"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.471197 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w24cb\" (UniqueName: \"kubernetes.io/projected/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-kube-api-access-w24cb\") pod \"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45\" (UID: \"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.471218 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6tnr\" (UniqueName: \"kubernetes.io/projected/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-kube-api-access-v6tnr\") pod \"b5c931ba-06fd-41eb-9bdc-514ec3faef6e\" (UID: \"b5c931ba-06fd-41eb-9bdc-514ec3faef6e\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.471281 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-operator-scripts\") pod \"b5c931ba-06fd-41eb-9bdc-514ec3faef6e\" (UID: \"b5c931ba-06fd-41eb-9bdc-514ec3faef6e\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.471306 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-operator-scripts\") pod \"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45\" (UID: \"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.471720 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df8s9\" (UniqueName: \"kubernetes.io/projected/be1ecfff-67b5-4cc5-90f0-d833276a5b62-kube-api-access-df8s9\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.471708 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1" (UID: "bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.472047 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5791f28c-d7ba-4aa4-99d5-319ecbd6bc45" (UID: "5791f28c-d7ba-4aa4-99d5-319ecbd6bc45"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.472444 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5c931ba-06fd-41eb-9bdc-514ec3faef6e" (UID: "b5c931ba-06fd-41eb-9bdc-514ec3faef6e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.472538 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a872e452-e198-4c3d-a8d4-85473516feab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a872e452-e198-4c3d-a8d4-85473516feab" (UID: "a872e452-e198-4c3d-a8d4-85473516feab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.476517 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-kube-api-access-v6tnr" (OuterVolumeSpecName: "kube-api-access-v6tnr") pod "b5c931ba-06fd-41eb-9bdc-514ec3faef6e" (UID: "b5c931ba-06fd-41eb-9bdc-514ec3faef6e"). InnerVolumeSpecName "kube-api-access-v6tnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.476549 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-kube-api-access-w24cb" (OuterVolumeSpecName: "kube-api-access-w24cb") pod "5791f28c-d7ba-4aa4-99d5-319ecbd6bc45" (UID: "5791f28c-d7ba-4aa4-99d5-319ecbd6bc45"). InnerVolumeSpecName "kube-api-access-w24cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.476566 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-kube-api-access-m9mvl" (OuterVolumeSpecName: "kube-api-access-m9mvl") pod "bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1" (UID: "bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1"). InnerVolumeSpecName "kube-api-access-m9mvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.477516 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a872e452-e198-4c3d-a8d4-85473516feab-kube-api-access-prbdq" (OuterVolumeSpecName: "kube-api-access-prbdq") pod "a872e452-e198-4c3d-a8d4-85473516feab" (UID: "a872e452-e198-4c3d-a8d4-85473516feab"). InnerVolumeSpecName "kube-api-access-prbdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.572772 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8f2x\" (UniqueName: \"kubernetes.io/projected/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-kube-api-access-k8f2x\") pod \"f5a1b5fa-aba9-4232-b5b7-96365cebef1a\" (UID: \"f5a1b5fa-aba9-4232-b5b7-96365cebef1a\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.572954 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-operator-scripts\") pod \"f5a1b5fa-aba9-4232-b5b7-96365cebef1a\" (UID: \"f5a1b5fa-aba9-4232-b5b7-96365cebef1a\") " Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.573351 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a872e452-e198-4c3d-a8d4-85473516feab-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.573363 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.573388 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prbdq\" (UniqueName: \"kubernetes.io/projected/a872e452-e198-4c3d-a8d4-85473516feab-kube-api-access-prbdq\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.573399 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9mvl\" (UniqueName: \"kubernetes.io/projected/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1-kube-api-access-m9mvl\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.573407 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w24cb\" (UniqueName: \"kubernetes.io/projected/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-kube-api-access-w24cb\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.573416 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6tnr\" (UniqueName: \"kubernetes.io/projected/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-kube-api-access-v6tnr\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.573426 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5c931ba-06fd-41eb-9bdc-514ec3faef6e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.573434 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.574028 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5a1b5fa-aba9-4232-b5b7-96365cebef1a" (UID: "f5a1b5fa-aba9-4232-b5b7-96365cebef1a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.577615 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-kube-api-access-k8f2x" (OuterVolumeSpecName: "kube-api-access-k8f2x") pod "f5a1b5fa-aba9-4232-b5b7-96365cebef1a" (UID: "f5a1b5fa-aba9-4232-b5b7-96365cebef1a"). InnerVolumeSpecName "kube-api-access-k8f2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.677191 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8f2x\" (UniqueName: \"kubernetes.io/projected/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-kube-api-access-k8f2x\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.677540 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5a1b5fa-aba9-4232-b5b7-96365cebef1a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.732934 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.732944 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b71c-account-create-update-jjh4g" event={"ID":"bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1","Type":"ContainerDied","Data":"7cb996eb0d602262740401026b55dd74fd392520f28c818f2838ccac872afa6f"} Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.732986 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cb996eb0d602262740401026b55dd74fd392520f28c818f2838ccac872afa6f" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.735385 4978 generic.go:334] "Generic (PLEG): container finished" podID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" containerID="b8960ac3480ae1d88146d962f2420be146e18df6db23c90812cb66575bdf4679" exitCode=0 Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.735430 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"09b73acd-bf19-49f5-b5e0-a60e180e9c1e","Type":"ContainerDied","Data":"b8960ac3480ae1d88146d962f2420be146e18df6db23c90812cb66575bdf4679"} Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.737832 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9373-account-create-update-c5bk6" event={"ID":"a872e452-e198-4c3d-a8d4-85473516feab","Type":"ContainerDied","Data":"8f5846d5b393fc3031a5997a87d3f1c794afd8b6ffa7e3d26a66204512976d76"} Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.737855 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f5846d5b393fc3031a5997a87d3f1c794afd8b6ffa7e3d26a66204512976d76" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.737854 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9373-account-create-update-c5bk6" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.739038 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wxrhg" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.739040 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wxrhg" event={"ID":"be1ecfff-67b5-4cc5-90f0-d833276a5b62","Type":"ContainerDied","Data":"8d6aaec4738f311d2ff97f035fb630eecb67f67829d51e56241fa4251f82b79d"} Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.739069 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d6aaec4738f311d2ff97f035fb630eecb67f67829d51e56241fa4251f82b79d" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.741594 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.741591 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ac49-account-create-update-2rrpr" event={"ID":"b5c931ba-06fd-41eb-9bdc-514ec3faef6e","Type":"ContainerDied","Data":"fa2deb6eff28e9c3f8f0ff41593d5ebb91b6729aaea332b4c878bd2a449b0afa"} Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.741645 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa2deb6eff28e9c3f8f0ff41593d5ebb91b6729aaea332b4c878bd2a449b0afa" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.744364 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-xq9mn" event={"ID":"f5a1b5fa-aba9-4232-b5b7-96365cebef1a","Type":"ContainerDied","Data":"01efc90514dc135417a5f411d44d73bfd6dbcf72a183142847cb5dd1b8faead7"} Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.744455 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01efc90514dc135417a5f411d44d73bfd6dbcf72a183142847cb5dd1b8faead7" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.744525 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-xq9mn" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.748562 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-swr2c" event={"ID":"5791f28c-d7ba-4aa4-99d5-319ecbd6bc45","Type":"ContainerDied","Data":"858cea1c6c917896b61718eb5064eb5f9428bda78ffdde3bce8a01f1dd72b21c"} Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.748608 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="858cea1c6c917896b61718eb5064eb5f9428bda78ffdde3bce8a01f1dd72b21c" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.748660 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-swr2c" Feb 25 07:08:14 crc kubenswrapper[4978]: I0225 07:08:14.992845 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:08:15 crc kubenswrapper[4978]: W0225 07:08:15.010299 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d28d9f8_de4f_4223_bfa6_fdf398f133bb.slice/crio-29b79f431130815c4fa90c1335e3ead67afd184fefc3693c8996018550e93ddb WatchSource:0}: Error finding container 29b79f431130815c4fa90c1335e3ead67afd184fefc3693c8996018550e93ddb: Status 404 returned error can't find the container with id 29b79f431130815c4fa90c1335e3ead67afd184fefc3693c8996018550e93ddb Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.141848 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.288794 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-internal-tls-certs\") pod \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.288867 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-logs\") pod \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.289341 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-logs" (OuterVolumeSpecName: "logs") pod "09b73acd-bf19-49f5-b5e0-a60e180e9c1e" (UID: "09b73acd-bf19-49f5-b5e0-a60e180e9c1e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.289402 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.289509 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djv55\" (UniqueName: \"kubernetes.io/projected/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-kube-api-access-djv55\") pod \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.289535 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-scripts\") pod \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.289558 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-httpd-run\") pod \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.289594 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-config-data\") pod \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.289658 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-combined-ca-bundle\") pod \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\" (UID: \"09b73acd-bf19-49f5-b5e0-a60e180e9c1e\") " Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.290050 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.290828 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "09b73acd-bf19-49f5-b5e0-a60e180e9c1e" (UID: "09b73acd-bf19-49f5-b5e0-a60e180e9c1e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.292840 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "09b73acd-bf19-49f5-b5e0-a60e180e9c1e" (UID: "09b73acd-bf19-49f5-b5e0-a60e180e9c1e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.298649 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-kube-api-access-djv55" (OuterVolumeSpecName: "kube-api-access-djv55") pod "09b73acd-bf19-49f5-b5e0-a60e180e9c1e" (UID: "09b73acd-bf19-49f5-b5e0-a60e180e9c1e"). InnerVolumeSpecName "kube-api-access-djv55". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.304845 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-scripts" (OuterVolumeSpecName: "scripts") pod "09b73acd-bf19-49f5-b5e0-a60e180e9c1e" (UID: "09b73acd-bf19-49f5-b5e0-a60e180e9c1e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.339729 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09b73acd-bf19-49f5-b5e0-a60e180e9c1e" (UID: "09b73acd-bf19-49f5-b5e0-a60e180e9c1e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.346728 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="667d3386-e724-42f3-940b-82af7780255b" path="/var/lib/kubelet/pods/667d3386-e724-42f3-940b-82af7780255b/volumes" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.360882 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "09b73acd-bf19-49f5-b5e0-a60e180e9c1e" (UID: "09b73acd-bf19-49f5-b5e0-a60e180e9c1e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.373549 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-config-data" (OuterVolumeSpecName: "config-data") pod "09b73acd-bf19-49f5-b5e0-a60e180e9c1e" (UID: "09b73acd-bf19-49f5-b5e0-a60e180e9c1e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.391587 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.391630 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djv55\" (UniqueName: \"kubernetes.io/projected/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-kube-api-access-djv55\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.391642 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.391651 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.391660 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.391668 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.391676 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09b73acd-bf19-49f5-b5e0-a60e180e9c1e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.410382 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.493345 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.758299 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"09b73acd-bf19-49f5-b5e0-a60e180e9c1e","Type":"ContainerDied","Data":"6c97b54e5d7163ddfe49a74471cf0b36387569b3312a3e69ff9b0a4c8c701862"} Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.758615 4978 scope.go:117] "RemoveContainer" containerID="b8960ac3480ae1d88146d962f2420be146e18df6db23c90812cb66575bdf4679" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.758348 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.764107 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e22ec80b-2066-4394-884c-7ce80ac4061b","Type":"ContainerStarted","Data":"20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527"} Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.764388 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.764493 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="ceilometer-central-agent" containerID="cri-o://8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8" gracePeriod=30 Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.764546 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="sg-core" containerID="cri-o://6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11" gracePeriod=30 Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.764580 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="ceilometer-notification-agent" containerID="cri-o://acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b" gracePeriod=30 Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.764599 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="proxy-httpd" containerID="cri-o://20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527" gracePeriod=30 Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.767884 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7d28d9f8-de4f-4223-bfa6-fdf398f133bb","Type":"ContainerStarted","Data":"29b79f431130815c4fa90c1335e3ead67afd184fefc3693c8996018550e93ddb"} Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.788538 4978 scope.go:117] "RemoveContainer" containerID="f68c2a7b016bdf0a6652e9324843ca4472b7b54fb5f39c860f5ab165234032df" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.795649 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.049968614 podStartE2EDuration="11.795595393s" podCreationTimestamp="2026-02-25 07:08:04 +0000 UTC" firstStartedPulling="2026-02-25 07:08:11.296631511 +0000 UTC m=+1384.735887970" lastFinishedPulling="2026-02-25 07:08:15.04225829 +0000 UTC m=+1388.481514749" observedRunningTime="2026-02-25 07:08:15.788863401 +0000 UTC m=+1389.228119870" watchObservedRunningTime="2026-02-25 07:08:15.795595393 +0000 UTC m=+1389.234851852" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.809348 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.824281 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.845307 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:08:15 crc kubenswrapper[4978]: E0225 07:08:15.847741 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" containerName="glance-log" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.847766 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" containerName="glance-log" Feb 25 07:08:15 crc kubenswrapper[4978]: E0225 07:08:15.847781 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a1b5fa-aba9-4232-b5b7-96365cebef1a" containerName="mariadb-database-create" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.847787 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a1b5fa-aba9-4232-b5b7-96365cebef1a" containerName="mariadb-database-create" Feb 25 07:08:15 crc kubenswrapper[4978]: E0225 07:08:15.847802 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5c931ba-06fd-41eb-9bdc-514ec3faef6e" containerName="mariadb-account-create-update" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.847810 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5c931ba-06fd-41eb-9bdc-514ec3faef6e" containerName="mariadb-account-create-update" Feb 25 07:08:15 crc kubenswrapper[4978]: E0225 07:08:15.847823 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5791f28c-d7ba-4aa4-99d5-319ecbd6bc45" containerName="mariadb-database-create" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.847829 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5791f28c-d7ba-4aa4-99d5-319ecbd6bc45" containerName="mariadb-database-create" Feb 25 07:08:15 crc kubenswrapper[4978]: E0225 07:08:15.847849 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be1ecfff-67b5-4cc5-90f0-d833276a5b62" containerName="mariadb-database-create" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.847855 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="be1ecfff-67b5-4cc5-90f0-d833276a5b62" containerName="mariadb-database-create" Feb 25 07:08:15 crc kubenswrapper[4978]: E0225 07:08:15.847877 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1" containerName="mariadb-account-create-update" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.847885 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1" containerName="mariadb-account-create-update" Feb 25 07:08:15 crc kubenswrapper[4978]: E0225 07:08:15.847904 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a872e452-e198-4c3d-a8d4-85473516feab" containerName="mariadb-account-create-update" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.847910 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a872e452-e198-4c3d-a8d4-85473516feab" containerName="mariadb-account-create-update" Feb 25 07:08:15 crc kubenswrapper[4978]: E0225 07:08:15.847940 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" containerName="glance-httpd" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.847946 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" containerName="glance-httpd" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.848336 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5791f28c-d7ba-4aa4-99d5-319ecbd6bc45" containerName="mariadb-database-create" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.848359 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" containerName="glance-httpd" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.848396 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" containerName="glance-log" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.848412 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="be1ecfff-67b5-4cc5-90f0-d833276a5b62" containerName="mariadb-database-create" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.848429 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a1b5fa-aba9-4232-b5b7-96365cebef1a" containerName="mariadb-database-create" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.848447 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a872e452-e198-4c3d-a8d4-85473516feab" containerName="mariadb-account-create-update" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.848469 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5c931ba-06fd-41eb-9bdc-514ec3faef6e" containerName="mariadb-account-create-update" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.848481 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1" containerName="mariadb-account-create-update" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.850780 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.854190 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.855621 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 25 07:08:15 crc kubenswrapper[4978]: I0225 07:08:15.917461 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.003197 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.003240 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-logs\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.003268 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.003322 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-scripts\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.003354 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.003433 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-config-data\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.003467 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.003515 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2hkq\" (UniqueName: \"kubernetes.io/projected/58c891d9-fbca-406c-bd1a-ecb13f9d7444-kube-api-access-g2hkq\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.105533 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-config-data\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.105613 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.105651 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2hkq\" (UniqueName: \"kubernetes.io/projected/58c891d9-fbca-406c-bd1a-ecb13f9d7444-kube-api-access-g2hkq\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.105725 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.105749 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-logs\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.105784 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.105839 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-scripts\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.105868 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.106368 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.106443 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-logs\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.106663 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.110686 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-config-data\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.115202 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-scripts\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.115670 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.115898 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.127336 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2hkq\" (UniqueName: \"kubernetes.io/projected/58c891d9-fbca-406c-bd1a-ecb13f9d7444-kube-api-access-g2hkq\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.149228 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.333030 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.540245 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.540286 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.540326 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.540966 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1d31d1a63c47b381a67200dff4a026477c582f264c877adf6984b366b252ba2"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.541010 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://b1d31d1a63c47b381a67200dff4a026477c582f264c877adf6984b366b252ba2" gracePeriod=600 Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.784429 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="b1d31d1a63c47b381a67200dff4a026477c582f264c877adf6984b366b252ba2" exitCode=0 Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.784756 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"b1d31d1a63c47b381a67200dff4a026477c582f264c877adf6984b366b252ba2"} Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.784788 4978 scope.go:117] "RemoveContainer" containerID="03d7d67e86736ca0a34d745efcdc9b40b54189fb150b16a7aa6067c654318766" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.790403 4978 generic.go:334] "Generic (PLEG): container finished" podID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerID="20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527" exitCode=0 Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.790421 4978 generic.go:334] "Generic (PLEG): container finished" podID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerID="6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11" exitCode=2 Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.790428 4978 generic.go:334] "Generic (PLEG): container finished" podID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerID="acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b" exitCode=0 Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.790462 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e22ec80b-2066-4394-884c-7ce80ac4061b","Type":"ContainerDied","Data":"20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527"} Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.790488 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e22ec80b-2066-4394-884c-7ce80ac4061b","Type":"ContainerDied","Data":"6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11"} Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.790500 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e22ec80b-2066-4394-884c-7ce80ac4061b","Type":"ContainerDied","Data":"acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b"} Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.794251 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7d28d9f8-de4f-4223-bfa6-fdf398f133bb","Type":"ContainerStarted","Data":"231167ca97b4be8ade22723cde0ed17906f6163a3cedd000bd07b8e2dddaaa83"} Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.794274 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7d28d9f8-de4f-4223-bfa6-fdf398f133bb","Type":"ContainerStarted","Data":"251a79da9fba47a65d4db2f16a1ee10da3ad7f02ffcf859eb40da544a9209781"} Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.828971 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.831382 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.831372025 podStartE2EDuration="3.831372025s" podCreationTimestamp="2026-02-25 07:08:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:16.820987299 +0000 UTC m=+1390.260243768" watchObservedRunningTime="2026-02-25 07:08:16.831372025 +0000 UTC m=+1390.270628484" Feb 25 07:08:16 crc kubenswrapper[4978]: I0225 07:08:16.936909 4978 scope.go:117] "RemoveContainer" containerID="6a5aed0cd78d1335f2c61b8685492f41cff7d1fc97c36517a6514e5c9e63dfe8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.343529 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09b73acd-bf19-49f5-b5e0-a60e180e9c1e" path="/var/lib/kubelet/pods/09b73acd-bf19-49f5-b5e0-a60e180e9c1e/volumes" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.398984 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tfnq8"] Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.400189 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.407603 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tfnq8"] Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.408929 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.409457 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-cx8wx" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.409617 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.529508 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-config-data\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.529573 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfpgb\" (UniqueName: \"kubernetes.io/projected/b2885313-ac3c-44af-9fb9-941071137b56-kube-api-access-lfpgb\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.529653 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.529703 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-scripts\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.630799 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfpgb\" (UniqueName: \"kubernetes.io/projected/b2885313-ac3c-44af-9fb9-941071137b56-kube-api-access-lfpgb\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.630902 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.630951 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-scripts\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.630987 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-config-data\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.636607 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-config-data\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.636812 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-scripts\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.636928 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.650085 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfpgb\" (UniqueName: \"kubernetes.io/projected/b2885313-ac3c-44af-9fb9-941071137b56-kube-api-access-lfpgb\") pod \"nova-cell0-conductor-db-sync-tfnq8\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.732043 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.807494 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5"} Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.810907 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"58c891d9-fbca-406c-bd1a-ecb13f9d7444","Type":"ContainerStarted","Data":"23b908a2d2c8c5b19239c4d15a74e513b855d7ec6155c2aab43b358150b36fe5"} Feb 25 07:08:17 crc kubenswrapper[4978]: I0225 07:08:17.810936 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"58c891d9-fbca-406c-bd1a-ecb13f9d7444","Type":"ContainerStarted","Data":"591db7360b851f95617b4ff54b6a72b81df0b8fd45afff36b1f06a0c114de670"} Feb 25 07:08:18 crc kubenswrapper[4978]: I0225 07:08:18.210056 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tfnq8"] Feb 25 07:08:18 crc kubenswrapper[4978]: W0225 07:08:18.211533 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2885313_ac3c_44af_9fb9_941071137b56.slice/crio-7b3b40c3a5306523937203d3dc1a968dd0a4c8da5e76cef74f35d71d0c97ba73 WatchSource:0}: Error finding container 7b3b40c3a5306523937203d3dc1a968dd0a4c8da5e76cef74f35d71d0c97ba73: Status 404 returned error can't find the container with id 7b3b40c3a5306523937203d3dc1a968dd0a4c8da5e76cef74f35d71d0c97ba73 Feb 25 07:08:18 crc kubenswrapper[4978]: I0225 07:08:18.824799 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"58c891d9-fbca-406c-bd1a-ecb13f9d7444","Type":"ContainerStarted","Data":"4b497e57dce20bdb93e7a82daed8baa0e9acd2b6048c9e18954a4341399c8f5b"} Feb 25 07:08:18 crc kubenswrapper[4978]: I0225 07:08:18.827553 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tfnq8" event={"ID":"b2885313-ac3c-44af-9fb9-941071137b56","Type":"ContainerStarted","Data":"7b3b40c3a5306523937203d3dc1a968dd0a4c8da5e76cef74f35d71d0c97ba73"} Feb 25 07:08:18 crc kubenswrapper[4978]: I0225 07:08:18.860707 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.860690641 podStartE2EDuration="3.860690641s" podCreationTimestamp="2026-02-25 07:08:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:18.846909917 +0000 UTC m=+1392.286166386" watchObservedRunningTime="2026-02-25 07:08:18.860690641 +0000 UTC m=+1392.299947100" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.411796 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.464710 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-sg-core-conf-yaml\") pod \"e22ec80b-2066-4394-884c-7ce80ac4061b\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.464852 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-combined-ca-bundle\") pod \"e22ec80b-2066-4394-884c-7ce80ac4061b\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.464918 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-scripts\") pod \"e22ec80b-2066-4394-884c-7ce80ac4061b\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.464994 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4bnh\" (UniqueName: \"kubernetes.io/projected/e22ec80b-2066-4394-884c-7ce80ac4061b-kube-api-access-b4bnh\") pod \"e22ec80b-2066-4394-884c-7ce80ac4061b\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.465062 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-run-httpd\") pod \"e22ec80b-2066-4394-884c-7ce80ac4061b\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.465130 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-config-data\") pod \"e22ec80b-2066-4394-884c-7ce80ac4061b\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.465178 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-log-httpd\") pod \"e22ec80b-2066-4394-884c-7ce80ac4061b\" (UID: \"e22ec80b-2066-4394-884c-7ce80ac4061b\") " Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.465958 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e22ec80b-2066-4394-884c-7ce80ac4061b" (UID: "e22ec80b-2066-4394-884c-7ce80ac4061b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.466005 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e22ec80b-2066-4394-884c-7ce80ac4061b" (UID: "e22ec80b-2066-4394-884c-7ce80ac4061b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.474478 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-scripts" (OuterVolumeSpecName: "scripts") pod "e22ec80b-2066-4394-884c-7ce80ac4061b" (UID: "e22ec80b-2066-4394-884c-7ce80ac4061b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.474693 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e22ec80b-2066-4394-884c-7ce80ac4061b-kube-api-access-b4bnh" (OuterVolumeSpecName: "kube-api-access-b4bnh") pod "e22ec80b-2066-4394-884c-7ce80ac4061b" (UID: "e22ec80b-2066-4394-884c-7ce80ac4061b"). InnerVolumeSpecName "kube-api-access-b4bnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.506685 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e22ec80b-2066-4394-884c-7ce80ac4061b" (UID: "e22ec80b-2066-4394-884c-7ce80ac4061b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.550525 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e22ec80b-2066-4394-884c-7ce80ac4061b" (UID: "e22ec80b-2066-4394-884c-7ce80ac4061b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.564453 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-config-data" (OuterVolumeSpecName: "config-data") pod "e22ec80b-2066-4394-884c-7ce80ac4061b" (UID: "e22ec80b-2066-4394-884c-7ce80ac4061b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.567238 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4bnh\" (UniqueName: \"kubernetes.io/projected/e22ec80b-2066-4394-884c-7ce80ac4061b-kube-api-access-b4bnh\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.567265 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.567274 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.567285 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e22ec80b-2066-4394-884c-7ce80ac4061b-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.567293 4978 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.567302 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.567309 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22ec80b-2066-4394-884c-7ce80ac4061b-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.843798 4978 generic.go:334] "Generic (PLEG): container finished" podID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerID="8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8" exitCode=0 Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.843926 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e22ec80b-2066-4394-884c-7ce80ac4061b","Type":"ContainerDied","Data":"8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8"} Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.844745 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e22ec80b-2066-4394-884c-7ce80ac4061b","Type":"ContainerDied","Data":"742259adffa91e4771f76c8962bb03b0c5d6d089dfd8be471d9533c52986027a"} Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.844788 4978 scope.go:117] "RemoveContainer" containerID="20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.843952 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.871703 4978 scope.go:117] "RemoveContainer" containerID="6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.887295 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.906566 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.911832 4978 scope.go:117] "RemoveContainer" containerID="acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.931027 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:19 crc kubenswrapper[4978]: E0225 07:08:19.931658 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="sg-core" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.931688 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="sg-core" Feb 25 07:08:19 crc kubenswrapper[4978]: E0225 07:08:19.931712 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="ceilometer-central-agent" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.931725 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="ceilometer-central-agent" Feb 25 07:08:19 crc kubenswrapper[4978]: E0225 07:08:19.931743 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="proxy-httpd" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.931757 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="proxy-httpd" Feb 25 07:08:19 crc kubenswrapper[4978]: E0225 07:08:19.931827 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="ceilometer-notification-agent" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.931844 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="ceilometer-notification-agent" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.932138 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="ceilometer-central-agent" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.932169 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="ceilometer-notification-agent" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.932195 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="proxy-httpd" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.932223 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" containerName="sg-core" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.935260 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.937199 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.940506 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.942353 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.957331 4978 scope.go:117] "RemoveContainer" containerID="8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.972983 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.973031 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-log-httpd\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.973056 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-config-data\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.973089 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-run-httpd\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.973113 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-scripts\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.973132 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5lqk\" (UniqueName: \"kubernetes.io/projected/104e9af0-2418-4ad7-9ab9-1c28240a4112-kube-api-access-j5lqk\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.973147 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.991374 4978 scope.go:117] "RemoveContainer" containerID="20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527" Feb 25 07:08:19 crc kubenswrapper[4978]: E0225 07:08:19.991857 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527\": container with ID starting with 20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527 not found: ID does not exist" containerID="20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.991896 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527"} err="failed to get container status \"20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527\": rpc error: code = NotFound desc = could not find container \"20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527\": container with ID starting with 20d5a65d4950b08e8e563a72dd6e3be6f2d115d50cf7c34720b78116c6c34527 not found: ID does not exist" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.991921 4978 scope.go:117] "RemoveContainer" containerID="6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11" Feb 25 07:08:19 crc kubenswrapper[4978]: E0225 07:08:19.992776 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11\": container with ID starting with 6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11 not found: ID does not exist" containerID="6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.992819 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11"} err="failed to get container status \"6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11\": rpc error: code = NotFound desc = could not find container \"6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11\": container with ID starting with 6fa83cb32d361f74198bb60dde563d7cd1505a0cbb849bc0e3eb13aee65dcf11 not found: ID does not exist" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.992845 4978 scope.go:117] "RemoveContainer" containerID="acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b" Feb 25 07:08:19 crc kubenswrapper[4978]: E0225 07:08:19.993251 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b\": container with ID starting with acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b not found: ID does not exist" containerID="acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.993281 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b"} err="failed to get container status \"acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b\": rpc error: code = NotFound desc = could not find container \"acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b\": container with ID starting with acc5df5f4c8c9f97a0581e32f6bddc85ec7b769bbccce4505dd7036ea60c228b not found: ID does not exist" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.993295 4978 scope.go:117] "RemoveContainer" containerID="8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8" Feb 25 07:08:19 crc kubenswrapper[4978]: E0225 07:08:19.993550 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8\": container with ID starting with 8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8 not found: ID does not exist" containerID="8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8" Feb 25 07:08:19 crc kubenswrapper[4978]: I0225 07:08:19.993572 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8"} err="failed to get container status \"8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8\": rpc error: code = NotFound desc = could not find container \"8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8\": container with ID starting with 8ce2e44e86d40bf9190c8ec47c90aac24c147bb7475ace949c175ca647efa0b8 not found: ID does not exist" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.074271 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.074321 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-log-httpd\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.074344 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-config-data\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.074396 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-run-httpd\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.074421 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-scripts\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.074437 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5lqk\" (UniqueName: \"kubernetes.io/projected/104e9af0-2418-4ad7-9ab9-1c28240a4112-kube-api-access-j5lqk\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.074458 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.074838 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-log-httpd\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.075283 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-run-httpd\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.078449 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.079476 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-scripts\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.082930 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-config-data\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.088539 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.094499 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5lqk\" (UniqueName: \"kubernetes.io/projected/104e9af0-2418-4ad7-9ab9-1c28240a4112-kube-api-access-j5lqk\") pod \"ceilometer-0\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.257405 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.674856 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:20 crc kubenswrapper[4978]: W0225 07:08:20.691950 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod104e9af0_2418_4ad7_9ab9_1c28240a4112.slice/crio-709059ab6c0fe1aaec3acb6e5bd4a6a9a0ecfd812df3d9bd92a97712cf549c50 WatchSource:0}: Error finding container 709059ab6c0fe1aaec3acb6e5bd4a6a9a0ecfd812df3d9bd92a97712cf549c50: Status 404 returned error can't find the container with id 709059ab6c0fe1aaec3acb6e5bd4a6a9a0ecfd812df3d9bd92a97712cf549c50 Feb 25 07:08:20 crc kubenswrapper[4978]: I0225 07:08:20.858437 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"104e9af0-2418-4ad7-9ab9-1c28240a4112","Type":"ContainerStarted","Data":"709059ab6c0fe1aaec3acb6e5bd4a6a9a0ecfd812df3d9bd92a97712cf549c50"} Feb 25 07:08:21 crc kubenswrapper[4978]: I0225 07:08:21.340975 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e22ec80b-2066-4394-884c-7ce80ac4061b" path="/var/lib/kubelet/pods/e22ec80b-2066-4394-884c-7ce80ac4061b/volumes" Feb 25 07:08:21 crc kubenswrapper[4978]: I0225 07:08:21.536489 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:08:21 crc kubenswrapper[4978]: I0225 07:08:21.601788 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5964547b68-crjwr"] Feb 25 07:08:21 crc kubenswrapper[4978]: I0225 07:08:21.602068 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5964547b68-crjwr" podUID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" containerName="neutron-api" containerID="cri-o://af6d5e300eab1e02eed4638bd374ff091a24a5f26d9debc40f54880c4cadf6b8" gracePeriod=30 Feb 25 07:08:21 crc kubenswrapper[4978]: I0225 07:08:21.603292 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5964547b68-crjwr" podUID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" containerName="neutron-httpd" containerID="cri-o://0d1ec2a95bdf608b1bae7e76043322c9d54dc4357493235e1965da313e3493ee" gracePeriod=30 Feb 25 07:08:21 crc kubenswrapper[4978]: I0225 07:08:21.870499 4978 generic.go:334] "Generic (PLEG): container finished" podID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" containerID="0d1ec2a95bdf608b1bae7e76043322c9d54dc4357493235e1965da313e3493ee" exitCode=0 Feb 25 07:08:21 crc kubenswrapper[4978]: I0225 07:08:21.870540 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5964547b68-crjwr" event={"ID":"9ee02b80-8075-47b3-8fe8-52b8a2e17371","Type":"ContainerDied","Data":"0d1ec2a95bdf608b1bae7e76043322c9d54dc4357493235e1965da313e3493ee"} Feb 25 07:08:23 crc kubenswrapper[4978]: I0225 07:08:23.415936 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:24 crc kubenswrapper[4978]: I0225 07:08:24.417636 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 07:08:24 crc kubenswrapper[4978]: I0225 07:08:24.417690 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 07:08:24 crc kubenswrapper[4978]: I0225 07:08:24.460706 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 07:08:24 crc kubenswrapper[4978]: I0225 07:08:24.472168 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 07:08:24 crc kubenswrapper[4978]: I0225 07:08:24.926047 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 07:08:24 crc kubenswrapper[4978]: I0225 07:08:24.926100 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 07:08:25 crc kubenswrapper[4978]: I0225 07:08:25.944762 4978 generic.go:334] "Generic (PLEG): container finished" podID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" containerID="af6d5e300eab1e02eed4638bd374ff091a24a5f26d9debc40f54880c4cadf6b8" exitCode=0 Feb 25 07:08:25 crc kubenswrapper[4978]: I0225 07:08:25.944918 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5964547b68-crjwr" event={"ID":"9ee02b80-8075-47b3-8fe8-52b8a2e17371","Type":"ContainerDied","Data":"af6d5e300eab1e02eed4638bd374ff091a24a5f26d9debc40f54880c4cadf6b8"} Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.281697 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.333589 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.333631 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.363011 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.377047 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.392988 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-ovndb-tls-certs\") pod \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.393036 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v8cw\" (UniqueName: \"kubernetes.io/projected/9ee02b80-8075-47b3-8fe8-52b8a2e17371-kube-api-access-9v8cw\") pod \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.393087 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-combined-ca-bundle\") pod \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.393875 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-httpd-config\") pod \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.393906 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-config\") pod \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\" (UID: \"9ee02b80-8075-47b3-8fe8-52b8a2e17371\") " Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.410535 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ee02b80-8075-47b3-8fe8-52b8a2e17371-kube-api-access-9v8cw" (OuterVolumeSpecName: "kube-api-access-9v8cw") pod "9ee02b80-8075-47b3-8fe8-52b8a2e17371" (UID: "9ee02b80-8075-47b3-8fe8-52b8a2e17371"). InnerVolumeSpecName "kube-api-access-9v8cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.410604 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "9ee02b80-8075-47b3-8fe8-52b8a2e17371" (UID: "9ee02b80-8075-47b3-8fe8-52b8a2e17371"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.458998 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-config" (OuterVolumeSpecName: "config") pod "9ee02b80-8075-47b3-8fe8-52b8a2e17371" (UID: "9ee02b80-8075-47b3-8fe8-52b8a2e17371"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.462488 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ee02b80-8075-47b3-8fe8-52b8a2e17371" (UID: "9ee02b80-8075-47b3-8fe8-52b8a2e17371"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.493579 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.495522 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v8cw\" (UniqueName: \"kubernetes.io/projected/9ee02b80-8075-47b3-8fe8-52b8a2e17371-kube-api-access-9v8cw\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.495550 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.495559 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.495568 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.503036 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.503399 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "9ee02b80-8075-47b3-8fe8-52b8a2e17371" (UID: "9ee02b80-8075-47b3-8fe8-52b8a2e17371"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.585597 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-776b767d76-hl99l"] Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.585834 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-776b767d76-hl99l" podUID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" containerName="placement-log" containerID="cri-o://1fb7bf2f0b0317ab071ae00aecb34bc39e311241db22967f13d25e3a933d30e5" gracePeriod=30 Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.586211 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-776b767d76-hl99l" podUID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" containerName="placement-api" containerID="cri-o://3738d8be721c087bd8f44c86484a7539eb563b4afdd76d2543dcc52eb0e30860" gracePeriod=30 Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.598004 4978 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ee02b80-8075-47b3-8fe8-52b8a2e17371-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.953894 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5964547b68-crjwr" event={"ID":"9ee02b80-8075-47b3-8fe8-52b8a2e17371","Type":"ContainerDied","Data":"33bea0fd49f502a85f8efa8d5c6ffccd81b68b31c773f26ffa1d0c16dbec49ac"} Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.953917 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5964547b68-crjwr" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.954153 4978 scope.go:117] "RemoveContainer" containerID="0d1ec2a95bdf608b1bae7e76043322c9d54dc4357493235e1965da313e3493ee" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.955534 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"104e9af0-2418-4ad7-9ab9-1c28240a4112","Type":"ContainerStarted","Data":"111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50"} Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.955578 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"104e9af0-2418-4ad7-9ab9-1c28240a4112","Type":"ContainerStarted","Data":"6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06"} Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.959513 4978 generic.go:334] "Generic (PLEG): container finished" podID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" containerID="1fb7bf2f0b0317ab071ae00aecb34bc39e311241db22967f13d25e3a933d30e5" exitCode=143 Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.959575 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-776b767d76-hl99l" event={"ID":"be86cdbd-c5d3-465a-96fb-426b76a8b6bb","Type":"ContainerDied","Data":"1fb7bf2f0b0317ab071ae00aecb34bc39e311241db22967f13d25e3a933d30e5"} Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.978622 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tfnq8" event={"ID":"b2885313-ac3c-44af-9fb9-941071137b56","Type":"ContainerStarted","Data":"d6443061bb18b3e732682398607026b9a8bfff4e6eda5690b8cd984c7e598a37"} Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.979479 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.979495 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.987282 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5964547b68-crjwr"] Feb 25 07:08:26 crc kubenswrapper[4978]: I0225 07:08:26.995118 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5964547b68-crjwr"] Feb 25 07:08:27 crc kubenswrapper[4978]: I0225 07:08:27.001276 4978 scope.go:117] "RemoveContainer" containerID="af6d5e300eab1e02eed4638bd374ff091a24a5f26d9debc40f54880c4cadf6b8" Feb 25 07:08:27 crc kubenswrapper[4978]: I0225 07:08:27.018534 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-tfnq8" podStartSLOduration=2.345920574 podStartE2EDuration="10.018515481s" podCreationTimestamp="2026-02-25 07:08:17 +0000 UTC" firstStartedPulling="2026-02-25 07:08:18.214228923 +0000 UTC m=+1391.653485382" lastFinishedPulling="2026-02-25 07:08:25.88682383 +0000 UTC m=+1399.326080289" observedRunningTime="2026-02-25 07:08:27.004546431 +0000 UTC m=+1400.443802890" watchObservedRunningTime="2026-02-25 07:08:27.018515481 +0000 UTC m=+1400.457771930" Feb 25 07:08:27 crc kubenswrapper[4978]: I0225 07:08:27.054936 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 07:08:27 crc kubenswrapper[4978]: I0225 07:08:27.055030 4978 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 25 07:08:27 crc kubenswrapper[4978]: I0225 07:08:27.355614 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" path="/var/lib/kubelet/pods/9ee02b80-8075-47b3-8fe8-52b8a2e17371/volumes" Feb 25 07:08:27 crc kubenswrapper[4978]: I0225 07:08:27.478452 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 07:08:28 crc kubenswrapper[4978]: I0225 07:08:28.002160 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"104e9af0-2418-4ad7-9ab9-1c28240a4112","Type":"ContainerStarted","Data":"7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf"} Feb 25 07:08:29 crc kubenswrapper[4978]: I0225 07:08:29.000545 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:29 crc kubenswrapper[4978]: I0225 07:08:29.001029 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 07:08:29 crc kubenswrapper[4978]: E0225 07:08:29.896499 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbe86cdbd_c5d3_465a_96fb_426b76a8b6bb.slice/crio-conmon-3738d8be721c087bd8f44c86484a7539eb563b4afdd76d2543dcc52eb0e30860.scope\": RecentStats: unable to find data in memory cache]" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.038893 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"104e9af0-2418-4ad7-9ab9-1c28240a4112","Type":"ContainerStarted","Data":"ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53"} Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.039548 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="ceilometer-central-agent" containerID="cri-o://6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06" gracePeriod=30 Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.039804 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.040163 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="sg-core" containerID="cri-o://7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf" gracePeriod=30 Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.040219 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="ceilometer-notification-agent" containerID="cri-o://111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50" gracePeriod=30 Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.040247 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="proxy-httpd" containerID="cri-o://ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53" gracePeriod=30 Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.050983 4978 generic.go:334] "Generic (PLEG): container finished" podID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" containerID="3738d8be721c087bd8f44c86484a7539eb563b4afdd76d2543dcc52eb0e30860" exitCode=0 Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.051725 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-776b767d76-hl99l" event={"ID":"be86cdbd-c5d3-465a-96fb-426b76a8b6bb","Type":"ContainerDied","Data":"3738d8be721c087bd8f44c86484a7539eb563b4afdd76d2543dcc52eb0e30860"} Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.097234 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.8708819930000002 podStartE2EDuration="11.097216949s" podCreationTimestamp="2026-02-25 07:08:19 +0000 UTC" firstStartedPulling="2026-02-25 07:08:20.69628565 +0000 UTC m=+1394.135542129" lastFinishedPulling="2026-02-25 07:08:28.922620616 +0000 UTC m=+1402.361877085" observedRunningTime="2026-02-25 07:08:30.069068203 +0000 UTC m=+1403.508324662" watchObservedRunningTime="2026-02-25 07:08:30.097216949 +0000 UTC m=+1403.536473398" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.228351 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-776b767d76-hl99l" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.281279 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-combined-ca-bundle\") pod \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.281345 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-scripts\") pod \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.281486 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-logs\") pod \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.281568 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-internal-tls-certs\") pod \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.281603 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qrgh\" (UniqueName: \"kubernetes.io/projected/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-kube-api-access-4qrgh\") pod \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.282551 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-config-data\") pod \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.282618 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-public-tls-certs\") pod \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\" (UID: \"be86cdbd-c5d3-465a-96fb-426b76a8b6bb\") " Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.283718 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-logs" (OuterVolumeSpecName: "logs") pod "be86cdbd-c5d3-465a-96fb-426b76a8b6bb" (UID: "be86cdbd-c5d3-465a-96fb-426b76a8b6bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.295298 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-scripts" (OuterVolumeSpecName: "scripts") pod "be86cdbd-c5d3-465a-96fb-426b76a8b6bb" (UID: "be86cdbd-c5d3-465a-96fb-426b76a8b6bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.303959 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-kube-api-access-4qrgh" (OuterVolumeSpecName: "kube-api-access-4qrgh") pod "be86cdbd-c5d3-465a-96fb-426b76a8b6bb" (UID: "be86cdbd-c5d3-465a-96fb-426b76a8b6bb"). InnerVolumeSpecName "kube-api-access-4qrgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.350270 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-config-data" (OuterVolumeSpecName: "config-data") pod "be86cdbd-c5d3-465a-96fb-426b76a8b6bb" (UID: "be86cdbd-c5d3-465a-96fb-426b76a8b6bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.376550 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be86cdbd-c5d3-465a-96fb-426b76a8b6bb" (UID: "be86cdbd-c5d3-465a-96fb-426b76a8b6bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.385106 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.385134 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.385144 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.385155 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qrgh\" (UniqueName: \"kubernetes.io/projected/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-kube-api-access-4qrgh\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.385165 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.401810 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "be86cdbd-c5d3-465a-96fb-426b76a8b6bb" (UID: "be86cdbd-c5d3-465a-96fb-426b76a8b6bb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.419817 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "be86cdbd-c5d3-465a-96fb-426b76a8b6bb" (UID: "be86cdbd-c5d3-465a-96fb-426b76a8b6bb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.486498 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:30 crc kubenswrapper[4978]: I0225 07:08:30.486532 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/be86cdbd-c5d3-465a-96fb-426b76a8b6bb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.060970 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-776b767d76-hl99l" event={"ID":"be86cdbd-c5d3-465a-96fb-426b76a8b6bb","Type":"ContainerDied","Data":"566e36ddc4b6b30b8144c94b686a183e417c2e1130150217e9c40be2cd8047d8"} Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.061040 4978 scope.go:117] "RemoveContainer" containerID="3738d8be721c087bd8f44c86484a7539eb563b4afdd76d2543dcc52eb0e30860" Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.061203 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-776b767d76-hl99l" Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.071741 4978 generic.go:334] "Generic (PLEG): container finished" podID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerID="ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53" exitCode=0 Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.071773 4978 generic.go:334] "Generic (PLEG): container finished" podID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerID="7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf" exitCode=2 Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.071788 4978 generic.go:334] "Generic (PLEG): container finished" podID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerID="111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50" exitCode=0 Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.071824 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"104e9af0-2418-4ad7-9ab9-1c28240a4112","Type":"ContainerDied","Data":"ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53"} Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.071871 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"104e9af0-2418-4ad7-9ab9-1c28240a4112","Type":"ContainerDied","Data":"7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf"} Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.071891 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"104e9af0-2418-4ad7-9ab9-1c28240a4112","Type":"ContainerDied","Data":"111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50"} Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.092499 4978 scope.go:117] "RemoveContainer" containerID="1fb7bf2f0b0317ab071ae00aecb34bc39e311241db22967f13d25e3a933d30e5" Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.180820 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-776b767d76-hl99l"] Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.189126 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-776b767d76-hl99l"] Feb 25 07:08:31 crc kubenswrapper[4978]: I0225 07:08:31.340400 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" path="/var/lib/kubelet/pods/be86cdbd-c5d3-465a-96fb-426b76a8b6bb/volumes" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.702414 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.865314 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-scripts\") pod \"104e9af0-2418-4ad7-9ab9-1c28240a4112\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.865449 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5lqk\" (UniqueName: \"kubernetes.io/projected/104e9af0-2418-4ad7-9ab9-1c28240a4112-kube-api-access-j5lqk\") pod \"104e9af0-2418-4ad7-9ab9-1c28240a4112\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.865481 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-config-data\") pod \"104e9af0-2418-4ad7-9ab9-1c28240a4112\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.865542 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-log-httpd\") pod \"104e9af0-2418-4ad7-9ab9-1c28240a4112\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.865568 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-run-httpd\") pod \"104e9af0-2418-4ad7-9ab9-1c28240a4112\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.865597 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-sg-core-conf-yaml\") pod \"104e9af0-2418-4ad7-9ab9-1c28240a4112\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.865614 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-combined-ca-bundle\") pod \"104e9af0-2418-4ad7-9ab9-1c28240a4112\" (UID: \"104e9af0-2418-4ad7-9ab9-1c28240a4112\") " Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.866225 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "104e9af0-2418-4ad7-9ab9-1c28240a4112" (UID: "104e9af0-2418-4ad7-9ab9-1c28240a4112"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.866495 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "104e9af0-2418-4ad7-9ab9-1c28240a4112" (UID: "104e9af0-2418-4ad7-9ab9-1c28240a4112"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.866559 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.872195 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-scripts" (OuterVolumeSpecName: "scripts") pod "104e9af0-2418-4ad7-9ab9-1c28240a4112" (UID: "104e9af0-2418-4ad7-9ab9-1c28240a4112"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.872451 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/104e9af0-2418-4ad7-9ab9-1c28240a4112-kube-api-access-j5lqk" (OuterVolumeSpecName: "kube-api-access-j5lqk") pod "104e9af0-2418-4ad7-9ab9-1c28240a4112" (UID: "104e9af0-2418-4ad7-9ab9-1c28240a4112"). InnerVolumeSpecName "kube-api-access-j5lqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.892333 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "104e9af0-2418-4ad7-9ab9-1c28240a4112" (UID: "104e9af0-2418-4ad7-9ab9-1c28240a4112"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.955078 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "104e9af0-2418-4ad7-9ab9-1c28240a4112" (UID: "104e9af0-2418-4ad7-9ab9-1c28240a4112"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.967615 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5lqk\" (UniqueName: \"kubernetes.io/projected/104e9af0-2418-4ad7-9ab9-1c28240a4112-kube-api-access-j5lqk\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.967651 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/104e9af0-2418-4ad7-9ab9-1c28240a4112-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.967662 4978 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.967670 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.967678 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:34 crc kubenswrapper[4978]: I0225 07:08:34.971854 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-config-data" (OuterVolumeSpecName: "config-data") pod "104e9af0-2418-4ad7-9ab9-1c28240a4112" (UID: "104e9af0-2418-4ad7-9ab9-1c28240a4112"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.069878 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104e9af0-2418-4ad7-9ab9-1c28240a4112-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.135193 4978 generic.go:334] "Generic (PLEG): container finished" podID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerID="6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06" exitCode=0 Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.135242 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"104e9af0-2418-4ad7-9ab9-1c28240a4112","Type":"ContainerDied","Data":"6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06"} Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.135274 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"104e9af0-2418-4ad7-9ab9-1c28240a4112","Type":"ContainerDied","Data":"709059ab6c0fe1aaec3acb6e5bd4a6a9a0ecfd812df3d9bd92a97712cf549c50"} Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.135301 4978 scope.go:117] "RemoveContainer" containerID="ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.135523 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.158915 4978 scope.go:117] "RemoveContainer" containerID="7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.175978 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.182317 4978 scope.go:117] "RemoveContainer" containerID="111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.194435 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.201778 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.202666 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" containerName="neutron-httpd" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.202686 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" containerName="neutron-httpd" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.202701 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="ceilometer-notification-agent" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.202707 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="ceilometer-notification-agent" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.202716 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" containerName="placement-log" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.202723 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" containerName="placement-log" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.202735 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="proxy-httpd" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.202741 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="proxy-httpd" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.202755 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" containerName="placement-api" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.202762 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" containerName="placement-api" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.202777 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="ceilometer-central-agent" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.202785 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="ceilometer-central-agent" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.202794 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" containerName="neutron-api" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.202800 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" containerName="neutron-api" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.202810 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="sg-core" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.202816 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="sg-core" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.203010 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" containerName="neutron-httpd" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.203024 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="ceilometer-notification-agent" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.203025 4978 scope.go:117] "RemoveContainer" containerID="6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.203035 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="proxy-httpd" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.203139 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="sg-core" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.203149 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ee02b80-8075-47b3-8fe8-52b8a2e17371" containerName="neutron-api" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.203159 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" containerName="ceilometer-central-agent" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.203172 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" containerName="placement-log" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.203185 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="be86cdbd-c5d3-465a-96fb-426b76a8b6bb" containerName="placement-api" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.205117 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.232147 4978 scope.go:117] "RemoveContainer" containerID="ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.235556 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53\": container with ID starting with ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53 not found: ID does not exist" containerID="ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.235594 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53"} err="failed to get container status \"ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53\": rpc error: code = NotFound desc = could not find container \"ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53\": container with ID starting with ced957fa31ba0a81f13cac47381188a9020a2a332eece5c6a527558968f9bb53 not found: ID does not exist" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.235616 4978 scope.go:117] "RemoveContainer" containerID="7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.235885 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf\": container with ID starting with 7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf not found: ID does not exist" containerID="7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.235905 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf"} err="failed to get container status \"7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf\": rpc error: code = NotFound desc = could not find container \"7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf\": container with ID starting with 7e40bd421d8f42eb3cd35ecc6a810e812795ba043515afd283bcba9df0d3c1bf not found: ID does not exist" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.235919 4978 scope.go:117] "RemoveContainer" containerID="111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.236128 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50\": container with ID starting with 111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50 not found: ID does not exist" containerID="111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.236151 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50"} err="failed to get container status \"111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50\": rpc error: code = NotFound desc = could not find container \"111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50\": container with ID starting with 111f19aa420248a835324f3cc555ae9182b140101f3448d9c275e4c345a66b50 not found: ID does not exist" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.236163 4978 scope.go:117] "RemoveContainer" containerID="6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06" Feb 25 07:08:35 crc kubenswrapper[4978]: E0225 07:08:35.236421 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06\": container with ID starting with 6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06 not found: ID does not exist" containerID="6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.236436 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06"} err="failed to get container status \"6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06\": rpc error: code = NotFound desc = could not find container \"6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06\": container with ID starting with 6773b7ff6dae57b7ae88139a2b187069500d1d03f4dab3fe7e8356987e2bef06 not found: ID does not exist" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.236914 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.237429 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.237456 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.338305 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="104e9af0-2418-4ad7-9ab9-1c28240a4112" path="/var/lib/kubelet/pods/104e9af0-2418-4ad7-9ab9-1c28240a4112/volumes" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.384726 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-log-httpd\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.384778 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-run-httpd\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.384801 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn5lx\" (UniqueName: \"kubernetes.io/projected/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-kube-api-access-zn5lx\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.384913 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.384940 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-scripts\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.384961 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-config-data\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.385072 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.486327 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.486928 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-log-httpd\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.487007 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-run-httpd\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.487052 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn5lx\" (UniqueName: \"kubernetes.io/projected/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-kube-api-access-zn5lx\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.487168 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.487218 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-scripts\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.487250 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-config-data\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.487434 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-log-httpd\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.487733 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-run-httpd\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.492249 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-scripts\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.492282 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-config-data\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.505486 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.508146 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.510918 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn5lx\" (UniqueName: \"kubernetes.io/projected/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-kube-api-access-zn5lx\") pod \"ceilometer-0\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.526073 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:35 crc kubenswrapper[4978]: I0225 07:08:35.986171 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:35 crc kubenswrapper[4978]: W0225 07:08:35.987206 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1aa9e7cf_a55a_40a3_a45a_843fc399f49c.slice/crio-c0a9ab3b507461edc1375215fc985382beda5c30da4fc704bfeeb5da1d1cdc2b WatchSource:0}: Error finding container c0a9ab3b507461edc1375215fc985382beda5c30da4fc704bfeeb5da1d1cdc2b: Status 404 returned error can't find the container with id c0a9ab3b507461edc1375215fc985382beda5c30da4fc704bfeeb5da1d1cdc2b Feb 25 07:08:36 crc kubenswrapper[4978]: I0225 07:08:36.153361 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1aa9e7cf-a55a-40a3-a45a-843fc399f49c","Type":"ContainerStarted","Data":"c0a9ab3b507461edc1375215fc985382beda5c30da4fc704bfeeb5da1d1cdc2b"} Feb 25 07:08:36 crc kubenswrapper[4978]: I0225 07:08:36.529668 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:37 crc kubenswrapper[4978]: I0225 07:08:37.164266 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1aa9e7cf-a55a-40a3-a45a-843fc399f49c","Type":"ContainerStarted","Data":"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6"} Feb 25 07:08:37 crc kubenswrapper[4978]: I0225 07:08:37.165702 4978 generic.go:334] "Generic (PLEG): container finished" podID="b2885313-ac3c-44af-9fb9-941071137b56" containerID="d6443061bb18b3e732682398607026b9a8bfff4e6eda5690b8cd984c7e598a37" exitCode=0 Feb 25 07:08:37 crc kubenswrapper[4978]: I0225 07:08:37.165726 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tfnq8" event={"ID":"b2885313-ac3c-44af-9fb9-941071137b56","Type":"ContainerDied","Data":"d6443061bb18b3e732682398607026b9a8bfff4e6eda5690b8cd984c7e598a37"} Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.174607 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1aa9e7cf-a55a-40a3-a45a-843fc399f49c","Type":"ContainerStarted","Data":"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e"} Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.567205 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.749268 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-combined-ca-bundle\") pod \"b2885313-ac3c-44af-9fb9-941071137b56\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.749360 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfpgb\" (UniqueName: \"kubernetes.io/projected/b2885313-ac3c-44af-9fb9-941071137b56-kube-api-access-lfpgb\") pod \"b2885313-ac3c-44af-9fb9-941071137b56\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.749518 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-config-data\") pod \"b2885313-ac3c-44af-9fb9-941071137b56\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.749719 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-scripts\") pod \"b2885313-ac3c-44af-9fb9-941071137b56\" (UID: \"b2885313-ac3c-44af-9fb9-941071137b56\") " Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.755863 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2885313-ac3c-44af-9fb9-941071137b56-kube-api-access-lfpgb" (OuterVolumeSpecName: "kube-api-access-lfpgb") pod "b2885313-ac3c-44af-9fb9-941071137b56" (UID: "b2885313-ac3c-44af-9fb9-941071137b56"). InnerVolumeSpecName "kube-api-access-lfpgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.762541 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-scripts" (OuterVolumeSpecName: "scripts") pod "b2885313-ac3c-44af-9fb9-941071137b56" (UID: "b2885313-ac3c-44af-9fb9-941071137b56"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.789565 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-config-data" (OuterVolumeSpecName: "config-data") pod "b2885313-ac3c-44af-9fb9-941071137b56" (UID: "b2885313-ac3c-44af-9fb9-941071137b56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.796474 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2885313-ac3c-44af-9fb9-941071137b56" (UID: "b2885313-ac3c-44af-9fb9-941071137b56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.853795 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.855242 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.856363 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2885313-ac3c-44af-9fb9-941071137b56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:38 crc kubenswrapper[4978]: I0225 07:08:38.856436 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfpgb\" (UniqueName: \"kubernetes.io/projected/b2885313-ac3c-44af-9fb9-941071137b56-kube-api-access-lfpgb\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.189421 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tfnq8" event={"ID":"b2885313-ac3c-44af-9fb9-941071137b56","Type":"ContainerDied","Data":"7b3b40c3a5306523937203d3dc1a968dd0a4c8da5e76cef74f35d71d0c97ba73"} Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.189488 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b3b40c3a5306523937203d3dc1a968dd0a4c8da5e76cef74f35d71d0c97ba73" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.189514 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tfnq8" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.351030 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 07:08:39 crc kubenswrapper[4978]: E0225 07:08:39.351419 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2885313-ac3c-44af-9fb9-941071137b56" containerName="nova-cell0-conductor-db-sync" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.351435 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2885313-ac3c-44af-9fb9-941071137b56" containerName="nova-cell0-conductor-db-sync" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.351599 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2885313-ac3c-44af-9fb9-941071137b56" containerName="nova-cell0-conductor-db-sync" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.352156 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.354140 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-cx8wx" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.354482 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.370542 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.470950 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.471598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.471757 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt5fk\" (UniqueName: \"kubernetes.io/projected/50c62987-be2b-4deb-a88c-107db5d00415-kube-api-access-tt5fk\") pod \"nova-cell0-conductor-0\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.574054 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt5fk\" (UniqueName: \"kubernetes.io/projected/50c62987-be2b-4deb-a88c-107db5d00415-kube-api-access-tt5fk\") pod \"nova-cell0-conductor-0\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.574145 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.574193 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.578925 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.580862 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.614130 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt5fk\" (UniqueName: \"kubernetes.io/projected/50c62987-be2b-4deb-a88c-107db5d00415-kube-api-access-tt5fk\") pod \"nova-cell0-conductor-0\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:39 crc kubenswrapper[4978]: I0225 07:08:39.673376 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:40 crc kubenswrapper[4978]: I0225 07:08:40.184719 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 07:08:40 crc kubenswrapper[4978]: W0225 07:08:40.193983 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50c62987_be2b_4deb_a88c_107db5d00415.slice/crio-dab00b558315dd670ef2dc1afe389f10485644ac083dd18bde12f210ae9add55 WatchSource:0}: Error finding container dab00b558315dd670ef2dc1afe389f10485644ac083dd18bde12f210ae9add55: Status 404 returned error can't find the container with id dab00b558315dd670ef2dc1afe389f10485644ac083dd18bde12f210ae9add55 Feb 25 07:08:41 crc kubenswrapper[4978]: I0225 07:08:41.224448 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"50c62987-be2b-4deb-a88c-107db5d00415","Type":"ContainerStarted","Data":"91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f"} Feb 25 07:08:41 crc kubenswrapper[4978]: I0225 07:08:41.225000 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"50c62987-be2b-4deb-a88c-107db5d00415","Type":"ContainerStarted","Data":"dab00b558315dd670ef2dc1afe389f10485644ac083dd18bde12f210ae9add55"} Feb 25 07:08:41 crc kubenswrapper[4978]: I0225 07:08:41.225043 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:41 crc kubenswrapper[4978]: I0225 07:08:41.228452 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1aa9e7cf-a55a-40a3-a45a-843fc399f49c","Type":"ContainerStarted","Data":"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b"} Feb 25 07:08:41 crc kubenswrapper[4978]: I0225 07:08:41.248739 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.248719259 podStartE2EDuration="2.248719259s" podCreationTimestamp="2026-02-25 07:08:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:41.245473277 +0000 UTC m=+1414.684729746" watchObservedRunningTime="2026-02-25 07:08:41.248719259 +0000 UTC m=+1414.687975718" Feb 25 07:08:43 crc kubenswrapper[4978]: I0225 07:08:43.259022 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1aa9e7cf-a55a-40a3-a45a-843fc399f49c","Type":"ContainerStarted","Data":"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05"} Feb 25 07:08:43 crc kubenswrapper[4978]: I0225 07:08:43.259691 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 07:08:43 crc kubenswrapper[4978]: I0225 07:08:43.259218 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="ceilometer-central-agent" containerID="cri-o://7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6" gracePeriod=30 Feb 25 07:08:43 crc kubenswrapper[4978]: I0225 07:08:43.259866 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="proxy-httpd" containerID="cri-o://781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05" gracePeriod=30 Feb 25 07:08:43 crc kubenswrapper[4978]: I0225 07:08:43.260438 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="ceilometer-notification-agent" containerID="cri-o://e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e" gracePeriod=30 Feb 25 07:08:43 crc kubenswrapper[4978]: I0225 07:08:43.260529 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="sg-core" containerID="cri-o://4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b" gracePeriod=30 Feb 25 07:08:43 crc kubenswrapper[4978]: I0225 07:08:43.296138 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.5856929530000001 podStartE2EDuration="8.296109254s" podCreationTimestamp="2026-02-25 07:08:35 +0000 UTC" firstStartedPulling="2026-02-25 07:08:35.9895892 +0000 UTC m=+1409.428845659" lastFinishedPulling="2026-02-25 07:08:42.700005461 +0000 UTC m=+1416.139261960" observedRunningTime="2026-02-25 07:08:43.292876862 +0000 UTC m=+1416.732133381" watchObservedRunningTime="2026-02-25 07:08:43.296109254 +0000 UTC m=+1416.735365753" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.258984 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273661 4978 generic.go:334] "Generic (PLEG): container finished" podID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerID="781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05" exitCode=0 Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273722 4978 generic.go:334] "Generic (PLEG): container finished" podID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerID="4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b" exitCode=2 Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273740 4978 generic.go:334] "Generic (PLEG): container finished" podID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerID="e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e" exitCode=0 Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273751 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273803 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1aa9e7cf-a55a-40a3-a45a-843fc399f49c","Type":"ContainerDied","Data":"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05"} Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273759 4978 generic.go:334] "Generic (PLEG): container finished" podID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerID="7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6" exitCode=0 Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273886 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1aa9e7cf-a55a-40a3-a45a-843fc399f49c","Type":"ContainerDied","Data":"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b"} Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273908 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1aa9e7cf-a55a-40a3-a45a-843fc399f49c","Type":"ContainerDied","Data":"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e"} Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273928 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1aa9e7cf-a55a-40a3-a45a-843fc399f49c","Type":"ContainerDied","Data":"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6"} Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273947 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1aa9e7cf-a55a-40a3-a45a-843fc399f49c","Type":"ContainerDied","Data":"c0a9ab3b507461edc1375215fc985382beda5c30da4fc704bfeeb5da1d1cdc2b"} Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.273968 4978 scope.go:117] "RemoveContainer" containerID="781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.282573 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-run-httpd\") pod \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.282655 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn5lx\" (UniqueName: \"kubernetes.io/projected/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-kube-api-access-zn5lx\") pod \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.282699 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-log-httpd\") pod \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.282784 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-config-data\") pod \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.282824 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-sg-core-conf-yaml\") pod \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.282925 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-combined-ca-bundle\") pod \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.283056 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-scripts\") pod \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\" (UID: \"1aa9e7cf-a55a-40a3-a45a-843fc399f49c\") " Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.285725 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1aa9e7cf-a55a-40a3-a45a-843fc399f49c" (UID: "1aa9e7cf-a55a-40a3-a45a-843fc399f49c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.288088 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1aa9e7cf-a55a-40a3-a45a-843fc399f49c" (UID: "1aa9e7cf-a55a-40a3-a45a-843fc399f49c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.292828 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-kube-api-access-zn5lx" (OuterVolumeSpecName: "kube-api-access-zn5lx") pod "1aa9e7cf-a55a-40a3-a45a-843fc399f49c" (UID: "1aa9e7cf-a55a-40a3-a45a-843fc399f49c"). InnerVolumeSpecName "kube-api-access-zn5lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.295612 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-scripts" (OuterVolumeSpecName: "scripts") pod "1aa9e7cf-a55a-40a3-a45a-843fc399f49c" (UID: "1aa9e7cf-a55a-40a3-a45a-843fc399f49c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.308865 4978 scope.go:117] "RemoveContainer" containerID="4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.340937 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1aa9e7cf-a55a-40a3-a45a-843fc399f49c" (UID: "1aa9e7cf-a55a-40a3-a45a-843fc399f49c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.344966 4978 scope.go:117] "RemoveContainer" containerID="e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.361607 4978 scope.go:117] "RemoveContainer" containerID="7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.380122 4978 scope.go:117] "RemoveContainer" containerID="781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05" Feb 25 07:08:44 crc kubenswrapper[4978]: E0225 07:08:44.380659 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05\": container with ID starting with 781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05 not found: ID does not exist" containerID="781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.380690 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05"} err="failed to get container status \"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05\": rpc error: code = NotFound desc = could not find container \"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05\": container with ID starting with 781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05 not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.380712 4978 scope.go:117] "RemoveContainer" containerID="4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b" Feb 25 07:08:44 crc kubenswrapper[4978]: E0225 07:08:44.381422 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b\": container with ID starting with 4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b not found: ID does not exist" containerID="4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.381446 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b"} err="failed to get container status \"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b\": rpc error: code = NotFound desc = could not find container \"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b\": container with ID starting with 4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.381459 4978 scope.go:117] "RemoveContainer" containerID="e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e" Feb 25 07:08:44 crc kubenswrapper[4978]: E0225 07:08:44.382017 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e\": container with ID starting with e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e not found: ID does not exist" containerID="e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.382066 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e"} err="failed to get container status \"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e\": rpc error: code = NotFound desc = could not find container \"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e\": container with ID starting with e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.382098 4978 scope.go:117] "RemoveContainer" containerID="7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6" Feb 25 07:08:44 crc kubenswrapper[4978]: E0225 07:08:44.382511 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6\": container with ID starting with 7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6 not found: ID does not exist" containerID="7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.382539 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6"} err="failed to get container status \"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6\": rpc error: code = NotFound desc = could not find container \"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6\": container with ID starting with 7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6 not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.382553 4978 scope.go:117] "RemoveContainer" containerID="781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.382842 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05"} err="failed to get container status \"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05\": rpc error: code = NotFound desc = could not find container \"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05\": container with ID starting with 781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05 not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.382858 4978 scope.go:117] "RemoveContainer" containerID="4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.383451 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b"} err="failed to get container status \"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b\": rpc error: code = NotFound desc = could not find container \"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b\": container with ID starting with 4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.383469 4978 scope.go:117] "RemoveContainer" containerID="e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.385313 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e"} err="failed to get container status \"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e\": rpc error: code = NotFound desc = could not find container \"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e\": container with ID starting with e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.385447 4978 scope.go:117] "RemoveContainer" containerID="7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.386306 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.386414 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn5lx\" (UniqueName: \"kubernetes.io/projected/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-kube-api-access-zn5lx\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.386486 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.386549 4978 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.386606 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.387119 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6"} err="failed to get container status \"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6\": rpc error: code = NotFound desc = could not find container \"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6\": container with ID starting with 7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6 not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.387207 4978 scope.go:117] "RemoveContainer" containerID="781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.387706 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05"} err="failed to get container status \"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05\": rpc error: code = NotFound desc = could not find container \"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05\": container with ID starting with 781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05 not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.387787 4978 scope.go:117] "RemoveContainer" containerID="4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.388277 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b"} err="failed to get container status \"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b\": rpc error: code = NotFound desc = could not find container \"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b\": container with ID starting with 4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.388548 4978 scope.go:117] "RemoveContainer" containerID="e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.388849 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e"} err="failed to get container status \"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e\": rpc error: code = NotFound desc = could not find container \"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e\": container with ID starting with e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.388927 4978 scope.go:117] "RemoveContainer" containerID="7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.389450 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6"} err="failed to get container status \"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6\": rpc error: code = NotFound desc = could not find container \"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6\": container with ID starting with 7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6 not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.389557 4978 scope.go:117] "RemoveContainer" containerID="781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.389815 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05"} err="failed to get container status \"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05\": rpc error: code = NotFound desc = could not find container \"781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05\": container with ID starting with 781624d2435b48c80ec12506b21dfe8797a6c6d520f214bb3e4a74786aff7c05 not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.389887 4978 scope.go:117] "RemoveContainer" containerID="4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.390394 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b"} err="failed to get container status \"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b\": rpc error: code = NotFound desc = could not find container \"4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b\": container with ID starting with 4ee146cbcb02adb4389eda0ca6acc3eb31c708db80816265349b494da0704e3b not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.390491 4978 scope.go:117] "RemoveContainer" containerID="e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.390909 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e"} err="failed to get container status \"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e\": rpc error: code = NotFound desc = could not find container \"e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e\": container with ID starting with e27542919f46407b81f1f16185d517a2ff8e9291eab49d8f0920eaba79fd073e not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.390987 4978 scope.go:117] "RemoveContainer" containerID="7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.391285 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6"} err="failed to get container status \"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6\": rpc error: code = NotFound desc = could not find container \"7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6\": container with ID starting with 7661d0b3bdd86315f6a56fa7048540e3fcd85bc55455e39637c6e8f63c7f4ac6 not found: ID does not exist" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.398699 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-config-data" (OuterVolumeSpecName: "config-data") pod "1aa9e7cf-a55a-40a3-a45a-843fc399f49c" (UID: "1aa9e7cf-a55a-40a3-a45a-843fc399f49c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.418363 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1aa9e7cf-a55a-40a3-a45a-843fc399f49c" (UID: "1aa9e7cf-a55a-40a3-a45a-843fc399f49c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.488520 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.488557 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1aa9e7cf-a55a-40a3-a45a-843fc399f49c-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.610852 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.619706 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.637299 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:44 crc kubenswrapper[4978]: E0225 07:08:44.637753 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="ceilometer-notification-agent" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.637779 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="ceilometer-notification-agent" Feb 25 07:08:44 crc kubenswrapper[4978]: E0225 07:08:44.637807 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="ceilometer-central-agent" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.637819 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="ceilometer-central-agent" Feb 25 07:08:44 crc kubenswrapper[4978]: E0225 07:08:44.637840 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="proxy-httpd" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.637851 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="proxy-httpd" Feb 25 07:08:44 crc kubenswrapper[4978]: E0225 07:08:44.637878 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="sg-core" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.637889 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="sg-core" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.638145 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="proxy-httpd" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.638170 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="ceilometer-central-agent" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.638192 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="ceilometer-notification-agent" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.638205 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" containerName="sg-core" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.640168 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.643902 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.647852 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.658149 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.796117 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-run-httpd\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.796191 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-scripts\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.796218 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-config-data\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.796340 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfbvh\" (UniqueName: \"kubernetes.io/projected/07069098-bc61-49f2-bdab-55e8e72f4faa-kube-api-access-qfbvh\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.796396 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-log-httpd\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.796448 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.796483 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.898287 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.898524 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-run-httpd\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.898689 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-scripts\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.898846 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-config-data\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.898963 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfbvh\" (UniqueName: \"kubernetes.io/projected/07069098-bc61-49f2-bdab-55e8e72f4faa-kube-api-access-qfbvh\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.899033 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-log-httpd\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.899088 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-run-httpd\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.899119 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.899550 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-log-httpd\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.904668 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-config-data\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.905215 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.919251 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.919896 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-scripts\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:44 crc kubenswrapper[4978]: I0225 07:08:44.928745 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfbvh\" (UniqueName: \"kubernetes.io/projected/07069098-bc61-49f2-bdab-55e8e72f4faa-kube-api-access-qfbvh\") pod \"ceilometer-0\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " pod="openstack/ceilometer-0" Feb 25 07:08:45 crc kubenswrapper[4978]: I0225 07:08:45.005697 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:08:45 crc kubenswrapper[4978]: I0225 07:08:45.336364 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1aa9e7cf-a55a-40a3-a45a-843fc399f49c" path="/var/lib/kubelet/pods/1aa9e7cf-a55a-40a3-a45a-843fc399f49c/volumes" Feb 25 07:08:45 crc kubenswrapper[4978]: I0225 07:08:45.475310 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:08:45 crc kubenswrapper[4978]: W0225 07:08:45.488852 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07069098_bc61_49f2_bdab_55e8e72f4faa.slice/crio-9bb5e9afa4e86fd8fe93d6ef9285750c8cb1729a3a67348dc857b7a484792a13 WatchSource:0}: Error finding container 9bb5e9afa4e86fd8fe93d6ef9285750c8cb1729a3a67348dc857b7a484792a13: Status 404 returned error can't find the container with id 9bb5e9afa4e86fd8fe93d6ef9285750c8cb1729a3a67348dc857b7a484792a13 Feb 25 07:08:46 crc kubenswrapper[4978]: I0225 07:08:46.295521 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07069098-bc61-49f2-bdab-55e8e72f4faa","Type":"ContainerStarted","Data":"0fab927ded83dc237f13d657ba5e00c5469912aaca90ef0539d5006a4a702098"} Feb 25 07:08:46 crc kubenswrapper[4978]: I0225 07:08:46.296283 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07069098-bc61-49f2-bdab-55e8e72f4faa","Type":"ContainerStarted","Data":"9bb5e9afa4e86fd8fe93d6ef9285750c8cb1729a3a67348dc857b7a484792a13"} Feb 25 07:08:47 crc kubenswrapper[4978]: I0225 07:08:47.311295 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07069098-bc61-49f2-bdab-55e8e72f4faa","Type":"ContainerStarted","Data":"f077e044a528cd13bb7a6f089c9729303595fa1f4b7b0316534c5443b54448c1"} Feb 25 07:08:48 crc kubenswrapper[4978]: I0225 07:08:48.333433 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07069098-bc61-49f2-bdab-55e8e72f4faa","Type":"ContainerStarted","Data":"7e58ad9997f17eb86b9ef45aba593e1784829f72024c363155d853f453ed3c18"} Feb 25 07:08:49 crc kubenswrapper[4978]: I0225 07:08:49.348557 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07069098-bc61-49f2-bdab-55e8e72f4faa","Type":"ContainerStarted","Data":"00360f4efb86fb57db32d1190a034b7291ca7ec862281fdebb829d0d236c5560"} Feb 25 07:08:49 crc kubenswrapper[4978]: I0225 07:08:49.348881 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 07:08:49 crc kubenswrapper[4978]: I0225 07:08:49.388125 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.912473058 podStartE2EDuration="5.388094078s" podCreationTimestamp="2026-02-25 07:08:44 +0000 UTC" firstStartedPulling="2026-02-25 07:08:45.491643092 +0000 UTC m=+1418.930899581" lastFinishedPulling="2026-02-25 07:08:48.967264112 +0000 UTC m=+1422.406520601" observedRunningTime="2026-02-25 07:08:49.377627529 +0000 UTC m=+1422.816884078" watchObservedRunningTime="2026-02-25 07:08:49.388094078 +0000 UTC m=+1422.827350577" Feb 25 07:08:49 crc kubenswrapper[4978]: I0225 07:08:49.709829 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.240153 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-kr7xw"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.241225 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.246479 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.246535 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.252221 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-kr7xw"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.337282 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwcfb\" (UniqueName: \"kubernetes.io/projected/a2cdb535-d30e-445f-97f9-259185cb4223-kube-api-access-xwcfb\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.337345 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.337438 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-scripts\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.337523 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-config-data\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.424183 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.425728 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.436331 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.439324 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-config-data\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.439389 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-config-data\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.439428 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e93f8648-b297-4a22-93fa-61d3ffe59bff-logs\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.439490 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwcfb\" (UniqueName: \"kubernetes.io/projected/a2cdb535-d30e-445f-97f9-259185cb4223-kube-api-access-xwcfb\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.439517 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.439538 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.439598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws8f4\" (UniqueName: \"kubernetes.io/projected/e93f8648-b297-4a22-93fa-61d3ffe59bff-kube-api-access-ws8f4\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.439635 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-scripts\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.448649 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.459281 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-scripts\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.464065 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-config-data\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.470732 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.493090 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.494225 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.500919 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwcfb\" (UniqueName: \"kubernetes.io/projected/a2cdb535-d30e-445f-97f9-259185cb4223-kube-api-access-xwcfb\") pod \"nova-cell0-cell-mapping-kr7xw\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.504771 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.513012 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.538746 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.540260 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.541410 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-config-data\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.541439 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e93f8648-b297-4a22-93fa-61d3ffe59bff-logs\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.541482 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.541506 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.541526 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.541546 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmzjp\" (UniqueName: \"kubernetes.io/projected/4ef26333-cc4a-4e32-bc18-72ffa6735f71-kube-api-access-hmzjp\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.541574 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws8f4\" (UniqueName: \"kubernetes.io/projected/e93f8648-b297-4a22-93fa-61d3ffe59bff-kube-api-access-ws8f4\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.544830 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e93f8648-b297-4a22-93fa-61d3ffe59bff-logs\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.545028 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.557478 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.557918 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.558709 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.561246 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.561483 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws8f4\" (UniqueName: \"kubernetes.io/projected/e93f8648-b297-4a22-93fa-61d3ffe59bff-kube-api-access-ws8f4\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.561903 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-config-data\") pod \"nova-api-0\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.569840 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.601112 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.635496 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.643572 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.643624 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.643643 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.643668 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmzjp\" (UniqueName: \"kubernetes.io/projected/4ef26333-cc4a-4e32-bc18-72ffa6735f71-kube-api-access-hmzjp\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.643688 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.643711 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-config-data\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.643809 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-config-data\") pod \"nova-scheduler-0\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.643842 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-logs\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.643877 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz6xr\" (UniqueName: \"kubernetes.io/projected/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-kube-api-access-pz6xr\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.643903 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxd9t\" (UniqueName: \"kubernetes.io/projected/1e877097-9faf-4497-9328-a46e828f79eb-kube-api-access-rxd9t\") pod \"nova-scheduler-0\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.648862 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.650976 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.662346 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmzjp\" (UniqueName: \"kubernetes.io/projected/4ef26333-cc4a-4e32-bc18-72ffa6735f71-kube-api-access-hmzjp\") pod \"nova-cell1-novncproxy-0\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.671007 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.692154 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-689695c99c-nmcjm"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.693805 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.703662 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-689695c99c-nmcjm"] Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.716631 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.748779 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxd9t\" (UniqueName: \"kubernetes.io/projected/1e877097-9faf-4497-9328-a46e828f79eb-kube-api-access-rxd9t\") pod \"nova-scheduler-0\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.749091 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgbc9\" (UniqueName: \"kubernetes.io/projected/617c5921-2132-4867-b0ef-5f10a7c1c938-kube-api-access-dgbc9\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.749458 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-config\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.749547 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-svc\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.749566 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.749615 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.749639 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-config-data\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.749681 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-config-data\") pod \"nova-scheduler-0\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.749698 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-sb\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.753292 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-logs\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.753323 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-swift-storage-0\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.753358 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-nb\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.753425 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz6xr\" (UniqueName: \"kubernetes.io/projected/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-kube-api-access-pz6xr\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.753674 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-logs\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.756199 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.757442 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-config-data\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.759793 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.759890 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-config-data\") pod \"nova-scheduler-0\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.778974 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz6xr\" (UniqueName: \"kubernetes.io/projected/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-kube-api-access-pz6xr\") pod \"nova-metadata-0\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " pod="openstack/nova-metadata-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.779820 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxd9t\" (UniqueName: \"kubernetes.io/projected/1e877097-9faf-4497-9328-a46e828f79eb-kube-api-access-rxd9t\") pod \"nova-scheduler-0\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.788392 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.855037 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-svc\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.855156 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-sb\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.855187 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-swift-storage-0\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.855214 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-nb\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.855280 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgbc9\" (UniqueName: \"kubernetes.io/projected/617c5921-2132-4867-b0ef-5f10a7c1c938-kube-api-access-dgbc9\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.855303 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-config\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.856926 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-config\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.857822 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-swift-storage-0\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.859721 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-svc\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.859845 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-nb\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.860094 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-sb\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.875660 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgbc9\" (UniqueName: \"kubernetes.io/projected/617c5921-2132-4867-b0ef-5f10a7c1c938-kube-api-access-dgbc9\") pod \"dnsmasq-dns-689695c99c-nmcjm\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:50 crc kubenswrapper[4978]: I0225 07:08:50.902190 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-kr7xw"] Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.044522 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.110735 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.281472 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:08:51 crc kubenswrapper[4978]: W0225 07:08:51.346640 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ef26333_cc4a_4e32_bc18_72ffa6735f71.slice/crio-af5a64bd80bb02a120422a57d3c91be4516d2b6a5f0e184a9daaeefe93e4f073 WatchSource:0}: Error finding container af5a64bd80bb02a120422a57d3c91be4516d2b6a5f0e184a9daaeefe93e4f073: Status 404 returned error can't find the container with id af5a64bd80bb02a120422a57d3c91be4516d2b6a5f0e184a9daaeefe93e4f073 Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.347007 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.424991 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-b7zzg"] Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.426065 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.428669 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kr7xw" event={"ID":"a2cdb535-d30e-445f-97f9-259185cb4223","Type":"ContainerStarted","Data":"348413bef14e6c205395d2c485c730cf829af44894b23680e9874614a44661ea"} Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.428712 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kr7xw" event={"ID":"a2cdb535-d30e-445f-97f9-259185cb4223","Type":"ContainerStarted","Data":"0fd8b31242343b99e59754ece96eaecddb138090b572edd5a8729ae7338b8fd2"} Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.428852 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.429084 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.430322 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ef26333-cc4a-4e32-bc18-72ffa6735f71","Type":"ContainerStarted","Data":"af5a64bd80bb02a120422a57d3c91be4516d2b6a5f0e184a9daaeefe93e4f073"} Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.432883 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e93f8648-b297-4a22-93fa-61d3ffe59bff","Type":"ContainerStarted","Data":"ff97ad5acf565fd7d1506bf40a98318eeb3a3d22d04d1f88a3f1e1a79c33b0e9"} Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.438101 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-b7zzg"] Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.453626 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.471111 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plmp2\" (UniqueName: \"kubernetes.io/projected/12c4ce63-060f-4f53-8e1a-a70b6ba97255-kube-api-access-plmp2\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.471258 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-scripts\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.471511 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-config-data\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.471610 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.479782 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-kr7xw" podStartSLOduration=1.479763157 podStartE2EDuration="1.479763157s" podCreationTimestamp="2026-02-25 07:08:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:51.470247857 +0000 UTC m=+1424.909504336" watchObservedRunningTime="2026-02-25 07:08:51.479763157 +0000 UTC m=+1424.919019616" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.575702 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-config-data\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.575782 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.575845 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plmp2\" (UniqueName: \"kubernetes.io/projected/12c4ce63-060f-4f53-8e1a-a70b6ba97255-kube-api-access-plmp2\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.575894 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-scripts\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.580282 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.582871 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-config-data\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.583859 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-scripts\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.591000 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.595582 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plmp2\" (UniqueName: \"kubernetes.io/projected/12c4ce63-060f-4f53-8e1a-a70b6ba97255-kube-api-access-plmp2\") pod \"nova-cell1-conductor-db-sync-b7zzg\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:51 crc kubenswrapper[4978]: W0225 07:08:51.678873 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod617c5921_2132_4867_b0ef_5f10a7c1c938.slice/crio-f8e2b1a2a9d0f0c7eb7b2873583675d29458e5d253699766e9daa9b96e3c0855 WatchSource:0}: Error finding container f8e2b1a2a9d0f0c7eb7b2873583675d29458e5d253699766e9daa9b96e3c0855: Status 404 returned error can't find the container with id f8e2b1a2a9d0f0c7eb7b2873583675d29458e5d253699766e9daa9b96e3c0855 Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.681590 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-689695c99c-nmcjm"] Feb 25 07:08:51 crc kubenswrapper[4978]: I0225 07:08:51.744838 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:08:52 crc kubenswrapper[4978]: I0225 07:08:52.205414 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-b7zzg"] Feb 25 07:08:52 crc kubenswrapper[4978]: I0225 07:08:52.444630 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-b7zzg" event={"ID":"12c4ce63-060f-4f53-8e1a-a70b6ba97255","Type":"ContainerStarted","Data":"e36db710c1ab7f0d8cb9355b7504726128ca6de40de4ab2bc21a32b4e767925d"} Feb 25 07:08:52 crc kubenswrapper[4978]: I0225 07:08:52.444668 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-b7zzg" event={"ID":"12c4ce63-060f-4f53-8e1a-a70b6ba97255","Type":"ContainerStarted","Data":"fbc6761becc794725b0ccbf16c9c1bf328e772f99cc5f19e9130378489671099"} Feb 25 07:08:52 crc kubenswrapper[4978]: I0225 07:08:52.466985 4978 generic.go:334] "Generic (PLEG): container finished" podID="617c5921-2132-4867-b0ef-5f10a7c1c938" containerID="e2d52313a54ebf436eac7928a9eab7d77dd7826b23c86f93edd92380125e47dd" exitCode=0 Feb 25 07:08:52 crc kubenswrapper[4978]: I0225 07:08:52.468219 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" event={"ID":"617c5921-2132-4867-b0ef-5f10a7c1c938","Type":"ContainerDied","Data":"e2d52313a54ebf436eac7928a9eab7d77dd7826b23c86f93edd92380125e47dd"} Feb 25 07:08:52 crc kubenswrapper[4978]: I0225 07:08:52.468282 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" event={"ID":"617c5921-2132-4867-b0ef-5f10a7c1c938","Type":"ContainerStarted","Data":"f8e2b1a2a9d0f0c7eb7b2873583675d29458e5d253699766e9daa9b96e3c0855"} Feb 25 07:08:52 crc kubenswrapper[4978]: I0225 07:08:52.490753 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94","Type":"ContainerStarted","Data":"132d6f212760973682ca74ace381f063d7e7f5a62402e8c5257357be0b161bd7"} Feb 25 07:08:52 crc kubenswrapper[4978]: I0225 07:08:52.506395 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e877097-9faf-4497-9328-a46e828f79eb","Type":"ContainerStarted","Data":"d2014a38192231d7483d5ab7cbfa79ed27b6f085d9127dbed18a9ba37ffce361"} Feb 25 07:08:52 crc kubenswrapper[4978]: I0225 07:08:52.513841 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-b7zzg" podStartSLOduration=1.513819136 podStartE2EDuration="1.513819136s" podCreationTimestamp="2026-02-25 07:08:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:52.473194257 +0000 UTC m=+1425.912450706" watchObservedRunningTime="2026-02-25 07:08:52.513819136 +0000 UTC m=+1425.953075595" Feb 25 07:08:53 crc kubenswrapper[4978]: I0225 07:08:53.516705 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" event={"ID":"617c5921-2132-4867-b0ef-5f10a7c1c938","Type":"ContainerStarted","Data":"e227b0a33aecace45163696f40de57e5900930437a82fb19dd2b00654d1ceff6"} Feb 25 07:08:53 crc kubenswrapper[4978]: I0225 07:08:53.540471 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" podStartSLOduration=3.54044982 podStartE2EDuration="3.54044982s" podCreationTimestamp="2026-02-25 07:08:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:53.532314564 +0000 UTC m=+1426.971571043" watchObservedRunningTime="2026-02-25 07:08:53.54044982 +0000 UTC m=+1426.979706279" Feb 25 07:08:54 crc kubenswrapper[4978]: I0225 07:08:54.417648 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:08:54 crc kubenswrapper[4978]: I0225 07:08:54.429149 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:08:54 crc kubenswrapper[4978]: I0225 07:08:54.523844 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.533817 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ef26333-cc4a-4e32-bc18-72ffa6735f71","Type":"ContainerStarted","Data":"9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a"} Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.534146 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="4ef26333-cc4a-4e32-bc18-72ffa6735f71" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a" gracePeriod=30 Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.535739 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e93f8648-b297-4a22-93fa-61d3ffe59bff","Type":"ContainerStarted","Data":"79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5"} Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.535765 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e93f8648-b297-4a22-93fa-61d3ffe59bff","Type":"ContainerStarted","Data":"99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7"} Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.538853 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94","Type":"ContainerStarted","Data":"635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67"} Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.538899 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94","Type":"ContainerStarted","Data":"0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823"} Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.539012 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" containerName="nova-metadata-log" containerID="cri-o://0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823" gracePeriod=30 Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.539092 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" containerName="nova-metadata-metadata" containerID="cri-o://635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67" gracePeriod=30 Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.543005 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e877097-9faf-4497-9328-a46e828f79eb","Type":"ContainerStarted","Data":"90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268"} Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.570011 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.259775578 podStartE2EDuration="5.569993623s" podCreationTimestamp="2026-02-25 07:08:50 +0000 UTC" firstStartedPulling="2026-02-25 07:08:51.358978345 +0000 UTC m=+1424.798234804" lastFinishedPulling="2026-02-25 07:08:54.66919639 +0000 UTC m=+1428.108452849" observedRunningTime="2026-02-25 07:08:55.552706989 +0000 UTC m=+1428.991963448" watchObservedRunningTime="2026-02-25 07:08:55.569993623 +0000 UTC m=+1429.009250072" Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.583624 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.496219682 podStartE2EDuration="5.583604232s" podCreationTimestamp="2026-02-25 07:08:50 +0000 UTC" firstStartedPulling="2026-02-25 07:08:51.586239809 +0000 UTC m=+1425.025496268" lastFinishedPulling="2026-02-25 07:08:54.673624339 +0000 UTC m=+1428.112880818" observedRunningTime="2026-02-25 07:08:55.574206815 +0000 UTC m=+1429.013463294" watchObservedRunningTime="2026-02-25 07:08:55.583604232 +0000 UTC m=+1429.022860691" Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.592854 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.356066661 podStartE2EDuration="5.592836683s" podCreationTimestamp="2026-02-25 07:08:50 +0000 UTC" firstStartedPulling="2026-02-25 07:08:51.453586033 +0000 UTC m=+1424.892842492" lastFinishedPulling="2026-02-25 07:08:54.690356045 +0000 UTC m=+1428.129612514" observedRunningTime="2026-02-25 07:08:55.589131166 +0000 UTC m=+1429.028387635" watchObservedRunningTime="2026-02-25 07:08:55.592836683 +0000 UTC m=+1429.032093142" Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.610908 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.246016526 podStartE2EDuration="5.61089069s" podCreationTimestamp="2026-02-25 07:08:50 +0000 UTC" firstStartedPulling="2026-02-25 07:08:51.294985511 +0000 UTC m=+1424.734241960" lastFinishedPulling="2026-02-25 07:08:54.659859665 +0000 UTC m=+1428.099116124" observedRunningTime="2026-02-25 07:08:55.604687805 +0000 UTC m=+1429.043944294" watchObservedRunningTime="2026-02-25 07:08:55.61089069 +0000 UTC m=+1429.050147169" Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.718005 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:08:55 crc kubenswrapper[4978]: I0225 07:08:55.789467 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.045507 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.045549 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.194616 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.372072 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz6xr\" (UniqueName: \"kubernetes.io/projected/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-kube-api-access-pz6xr\") pod \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.372163 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-combined-ca-bundle\") pod \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.372251 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-config-data\") pod \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.372317 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-logs\") pod \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\" (UID: \"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94\") " Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.373622 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-logs" (OuterVolumeSpecName: "logs") pod "02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" (UID: "02f6ec24-b512-4176-8ba6-0cd4d7a4ac94"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.378121 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-kube-api-access-pz6xr" (OuterVolumeSpecName: "kube-api-access-pz6xr") pod "02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" (UID: "02f6ec24-b512-4176-8ba6-0cd4d7a4ac94"). InnerVolumeSpecName "kube-api-access-pz6xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.404128 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" (UID: "02f6ec24-b512-4176-8ba6-0cd4d7a4ac94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.411653 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-config-data" (OuterVolumeSpecName: "config-data") pod "02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" (UID: "02f6ec24-b512-4176-8ba6-0cd4d7a4ac94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.475884 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.475970 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz6xr\" (UniqueName: \"kubernetes.io/projected/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-kube-api-access-pz6xr\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.475991 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.476042 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.556871 4978 generic.go:334] "Generic (PLEG): container finished" podID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" containerID="635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67" exitCode=0 Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.556901 4978 generic.go:334] "Generic (PLEG): container finished" podID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" containerID="0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823" exitCode=143 Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.556961 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94","Type":"ContainerDied","Data":"635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67"} Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.557039 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94","Type":"ContainerDied","Data":"0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823"} Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.557062 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.557088 4978 scope.go:117] "RemoveContainer" containerID="635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.557066 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"02f6ec24-b512-4176-8ba6-0cd4d7a4ac94","Type":"ContainerDied","Data":"132d6f212760973682ca74ace381f063d7e7f5a62402e8c5257357be0b161bd7"} Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.600538 4978 scope.go:117] "RemoveContainer" containerID="0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.613889 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.643537 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.647872 4978 scope.go:117] "RemoveContainer" containerID="635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.648019 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:08:56 crc kubenswrapper[4978]: E0225 07:08:56.648544 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" containerName="nova-metadata-log" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.648571 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" containerName="nova-metadata-log" Feb 25 07:08:56 crc kubenswrapper[4978]: E0225 07:08:56.648606 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" containerName="nova-metadata-metadata" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.648616 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" containerName="nova-metadata-metadata" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.648881 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" containerName="nova-metadata-log" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.648936 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" containerName="nova-metadata-metadata" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.650181 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: E0225 07:08:56.651285 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67\": container with ID starting with 635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67 not found: ID does not exist" containerID="635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.651318 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67"} err="failed to get container status \"635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67\": rpc error: code = NotFound desc = could not find container \"635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67\": container with ID starting with 635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67 not found: ID does not exist" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.651345 4978 scope.go:117] "RemoveContainer" containerID="0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823" Feb 25 07:08:56 crc kubenswrapper[4978]: E0225 07:08:56.653398 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823\": container with ID starting with 0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823 not found: ID does not exist" containerID="0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.653500 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823"} err="failed to get container status \"0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823\": rpc error: code = NotFound desc = could not find container \"0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823\": container with ID starting with 0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823 not found: ID does not exist" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.653557 4978 scope.go:117] "RemoveContainer" containerID="635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.654030 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67"} err="failed to get container status \"635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67\": rpc error: code = NotFound desc = could not find container \"635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67\": container with ID starting with 635b3bad8cb1d3371794b9ac7f8ed993b6156d97125a676e2021c5dcd6fb7f67 not found: ID does not exist" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.654080 4978 scope.go:117] "RemoveContainer" containerID="0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.654458 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823"} err="failed to get container status \"0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823\": rpc error: code = NotFound desc = could not find container \"0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823\": container with ID starting with 0eb4cb3711bd8826e9c408f1802b7b22e8931cb8ddec7492602f48daf58ec823 not found: ID does not exist" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.654769 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.659841 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.660129 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.678605 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lz5v\" (UniqueName: \"kubernetes.io/projected/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-kube-api-access-2lz5v\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.678715 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-config-data\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.678761 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.678792 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-logs\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.678864 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.780433 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-config-data\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.780488 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.780513 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-logs\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.780565 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.780614 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lz5v\" (UniqueName: \"kubernetes.io/projected/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-kube-api-access-2lz5v\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.781735 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-logs\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.784551 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-config-data\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.785195 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.787402 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:56 crc kubenswrapper[4978]: I0225 07:08:56.797860 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lz5v\" (UniqueName: \"kubernetes.io/projected/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-kube-api-access-2lz5v\") pod \"nova-metadata-0\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " pod="openstack/nova-metadata-0" Feb 25 07:08:57 crc kubenswrapper[4978]: I0225 07:08:57.030526 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:08:57 crc kubenswrapper[4978]: I0225 07:08:57.349836 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02f6ec24-b512-4176-8ba6-0cd4d7a4ac94" path="/var/lib/kubelet/pods/02f6ec24-b512-4176-8ba6-0cd4d7a4ac94/volumes" Feb 25 07:08:57 crc kubenswrapper[4978]: I0225 07:08:57.552893 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:08:57 crc kubenswrapper[4978]: W0225 07:08:57.566251 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd06b54a6_6f8d_4ed0_9b15_d09a5a187d18.slice/crio-86649b3c8b209a0d322cf92d0885ad7214c7f87eb681117f3feec77a8ffcb191 WatchSource:0}: Error finding container 86649b3c8b209a0d322cf92d0885ad7214c7f87eb681117f3feec77a8ffcb191: Status 404 returned error can't find the container with id 86649b3c8b209a0d322cf92d0885ad7214c7f87eb681117f3feec77a8ffcb191 Feb 25 07:08:58 crc kubenswrapper[4978]: I0225 07:08:58.586149 4978 generic.go:334] "Generic (PLEG): container finished" podID="a2cdb535-d30e-445f-97f9-259185cb4223" containerID="348413bef14e6c205395d2c485c730cf829af44894b23680e9874614a44661ea" exitCode=0 Feb 25 07:08:58 crc kubenswrapper[4978]: I0225 07:08:58.586250 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kr7xw" event={"ID":"a2cdb535-d30e-445f-97f9-259185cb4223","Type":"ContainerDied","Data":"348413bef14e6c205395d2c485c730cf829af44894b23680e9874614a44661ea"} Feb 25 07:08:58 crc kubenswrapper[4978]: I0225 07:08:58.590167 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18","Type":"ContainerStarted","Data":"ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f"} Feb 25 07:08:58 crc kubenswrapper[4978]: I0225 07:08:58.590232 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18","Type":"ContainerStarted","Data":"417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb"} Feb 25 07:08:58 crc kubenswrapper[4978]: I0225 07:08:58.590253 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18","Type":"ContainerStarted","Data":"86649b3c8b209a0d322cf92d0885ad7214c7f87eb681117f3feec77a8ffcb191"} Feb 25 07:08:58 crc kubenswrapper[4978]: I0225 07:08:58.670766 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.670741554 podStartE2EDuration="2.670741554s" podCreationTimestamp="2026-02-25 07:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:08:58.645696995 +0000 UTC m=+1432.084953534" watchObservedRunningTime="2026-02-25 07:08:58.670741554 +0000 UTC m=+1432.109998033" Feb 25 07:08:59 crc kubenswrapper[4978]: I0225 07:08:59.606818 4978 generic.go:334] "Generic (PLEG): container finished" podID="12c4ce63-060f-4f53-8e1a-a70b6ba97255" containerID="e36db710c1ab7f0d8cb9355b7504726128ca6de40de4ab2bc21a32b4e767925d" exitCode=0 Feb 25 07:08:59 crc kubenswrapper[4978]: I0225 07:08:59.606935 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-b7zzg" event={"ID":"12c4ce63-060f-4f53-8e1a-a70b6ba97255","Type":"ContainerDied","Data":"e36db710c1ab7f0d8cb9355b7504726128ca6de40de4ab2bc21a32b4e767925d"} Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.129897 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.258628 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-config-data\") pod \"a2cdb535-d30e-445f-97f9-259185cb4223\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.258823 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-combined-ca-bundle\") pod \"a2cdb535-d30e-445f-97f9-259185cb4223\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.258945 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-scripts\") pod \"a2cdb535-d30e-445f-97f9-259185cb4223\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.259010 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwcfb\" (UniqueName: \"kubernetes.io/projected/a2cdb535-d30e-445f-97f9-259185cb4223-kube-api-access-xwcfb\") pod \"a2cdb535-d30e-445f-97f9-259185cb4223\" (UID: \"a2cdb535-d30e-445f-97f9-259185cb4223\") " Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.265125 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-scripts" (OuterVolumeSpecName: "scripts") pod "a2cdb535-d30e-445f-97f9-259185cb4223" (UID: "a2cdb535-d30e-445f-97f9-259185cb4223"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.267004 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2cdb535-d30e-445f-97f9-259185cb4223-kube-api-access-xwcfb" (OuterVolumeSpecName: "kube-api-access-xwcfb") pod "a2cdb535-d30e-445f-97f9-259185cb4223" (UID: "a2cdb535-d30e-445f-97f9-259185cb4223"). InnerVolumeSpecName "kube-api-access-xwcfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.306921 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2cdb535-d30e-445f-97f9-259185cb4223" (UID: "a2cdb535-d30e-445f-97f9-259185cb4223"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.309195 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-config-data" (OuterVolumeSpecName: "config-data") pod "a2cdb535-d30e-445f-97f9-259185cb4223" (UID: "a2cdb535-d30e-445f-97f9-259185cb4223"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.361716 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.361769 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwcfb\" (UniqueName: \"kubernetes.io/projected/a2cdb535-d30e-445f-97f9-259185cb4223-kube-api-access-xwcfb\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.361855 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.361875 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2cdb535-d30e-445f-97f9-259185cb4223-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.625997 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-kr7xw" event={"ID":"a2cdb535-d30e-445f-97f9-259185cb4223","Type":"ContainerDied","Data":"0fd8b31242343b99e59754ece96eaecddb138090b572edd5a8729ae7338b8fd2"} Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.626603 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fd8b31242343b99e59754ece96eaecddb138090b572edd5a8729ae7338b8fd2" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.626095 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-kr7xw" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.671974 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.672050 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.792540 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.820443 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.838569 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.839426 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.853124 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.853329 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" containerName="nova-metadata-log" containerID="cri-o://417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb" gracePeriod=30 Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.853498 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" containerName="nova-metadata-metadata" containerID="cri-o://ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f" gracePeriod=30 Feb 25 07:09:00 crc kubenswrapper[4978]: I0225 07:09:00.955813 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.080833 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plmp2\" (UniqueName: \"kubernetes.io/projected/12c4ce63-060f-4f53-8e1a-a70b6ba97255-kube-api-access-plmp2\") pod \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.081186 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-config-data\") pod \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.081285 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-scripts\") pod \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.081325 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-combined-ca-bundle\") pod \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\" (UID: \"12c4ce63-060f-4f53-8e1a-a70b6ba97255\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.087576 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c4ce63-060f-4f53-8e1a-a70b6ba97255-kube-api-access-plmp2" (OuterVolumeSpecName: "kube-api-access-plmp2") pod "12c4ce63-060f-4f53-8e1a-a70b6ba97255" (UID: "12c4ce63-060f-4f53-8e1a-a70b6ba97255"). InnerVolumeSpecName "kube-api-access-plmp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.088459 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-scripts" (OuterVolumeSpecName: "scripts") pod "12c4ce63-060f-4f53-8e1a-a70b6ba97255" (UID: "12c4ce63-060f-4f53-8e1a-a70b6ba97255"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.106398 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12c4ce63-060f-4f53-8e1a-a70b6ba97255" (UID: "12c4ce63-060f-4f53-8e1a-a70b6ba97255"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.106897 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-config-data" (OuterVolumeSpecName: "config-data") pod "12c4ce63-060f-4f53-8e1a-a70b6ba97255" (UID: "12c4ce63-060f-4f53-8e1a-a70b6ba97255"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.113342 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.183114 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.183147 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.183155 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c4ce63-060f-4f53-8e1a-a70b6ba97255-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.183166 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plmp2\" (UniqueName: \"kubernetes.io/projected/12c4ce63-060f-4f53-8e1a-a70b6ba97255-kube-api-access-plmp2\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.187752 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d4ffdd57-jtdkj"] Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.194495 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" podUID="32b6e3c5-3010-4286-b0aa-a2bd20517933" containerName="dnsmasq-dns" containerID="cri-o://8b4f5ceb88e5882edab1442efa5555b2f11dddb6c403904c3906cd14a276a6f4" gracePeriod=10 Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.475556 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.589235 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-nova-metadata-tls-certs\") pod \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.589346 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-combined-ca-bundle\") pod \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.589418 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-config-data\") pod \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.589437 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-logs\") pod \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.589492 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lz5v\" (UniqueName: \"kubernetes.io/projected/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-kube-api-access-2lz5v\") pod \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\" (UID: \"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.594742 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-logs" (OuterVolumeSpecName: "logs") pod "d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" (UID: "d06b54a6-6f8d-4ed0-9b15-d09a5a187d18"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.595111 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-kube-api-access-2lz5v" (OuterVolumeSpecName: "kube-api-access-2lz5v") pod "d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" (UID: "d06b54a6-6f8d-4ed0-9b15-d09a5a187d18"). InnerVolumeSpecName "kube-api-access-2lz5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.626669 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-config-data" (OuterVolumeSpecName: "config-data") pod "d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" (UID: "d06b54a6-6f8d-4ed0-9b15-d09a5a187d18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.642461 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" (UID: "d06b54a6-6f8d-4ed0-9b15-d09a5a187d18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.643591 4978 generic.go:334] "Generic (PLEG): container finished" podID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" containerID="ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f" exitCode=0 Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.643619 4978 generic.go:334] "Generic (PLEG): container finished" podID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" containerID="417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb" exitCode=143 Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.643685 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.643721 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18","Type":"ContainerDied","Data":"ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f"} Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.643754 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18","Type":"ContainerDied","Data":"417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb"} Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.643765 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d06b54a6-6f8d-4ed0-9b15-d09a5a187d18","Type":"ContainerDied","Data":"86649b3c8b209a0d322cf92d0885ad7214c7f87eb681117f3feec77a8ffcb191"} Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.643780 4978 scope.go:117] "RemoveContainer" containerID="ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.652290 4978 generic.go:334] "Generic (PLEG): container finished" podID="32b6e3c5-3010-4286-b0aa-a2bd20517933" containerID="8b4f5ceb88e5882edab1442efa5555b2f11dddb6c403904c3906cd14a276a6f4" exitCode=0 Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.652344 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" event={"ID":"32b6e3c5-3010-4286-b0aa-a2bd20517933","Type":"ContainerDied","Data":"8b4f5ceb88e5882edab1442efa5555b2f11dddb6c403904c3906cd14a276a6f4"} Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.653675 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerName="nova-api-log" containerID="cri-o://99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7" gracePeriod=30 Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.653781 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-b7zzg" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.654394 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-b7zzg" event={"ID":"12c4ce63-060f-4f53-8e1a-a70b6ba97255","Type":"ContainerDied","Data":"fbc6761becc794725b0ccbf16c9c1bf328e772f99cc5f19e9130378489671099"} Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.654407 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbc6761becc794725b0ccbf16c9c1bf328e772f99cc5f19e9130378489671099" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.654484 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="1e877097-9faf-4497-9328-a46e828f79eb" containerName="nova-scheduler-scheduler" containerID="cri-o://90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" gracePeriod=30 Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.655040 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerName="nova-api-api" containerID="cri-o://79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5" gracePeriod=30 Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.657554 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" (UID: "d06b54a6-6f8d-4ed0-9b15-d09a5a187d18"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: E0225 07:09:01.664134 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.665872 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": EOF" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.669759 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.191:8774/\": EOF" Feb 25 07:09:01 crc kubenswrapper[4978]: E0225 07:09:01.688102 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.691518 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lz5v\" (UniqueName: \"kubernetes.io/projected/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-kube-api-access-2lz5v\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.691552 4978 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.691567 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.691579 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.691591 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: E0225 07:09:01.695545 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:09:01 crc kubenswrapper[4978]: E0225 07:09:01.695611 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1e877097-9faf-4497-9328-a46e828f79eb" containerName="nova-scheduler-scheduler" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.696722 4978 scope.go:117] "RemoveContainer" containerID="417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.719231 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 07:09:01 crc kubenswrapper[4978]: E0225 07:09:01.721504 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2cdb535-d30e-445f-97f9-259185cb4223" containerName="nova-manage" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.721543 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2cdb535-d30e-445f-97f9-259185cb4223" containerName="nova-manage" Feb 25 07:09:01 crc kubenswrapper[4978]: E0225 07:09:01.721558 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" containerName="nova-metadata-metadata" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.721567 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" containerName="nova-metadata-metadata" Feb 25 07:09:01 crc kubenswrapper[4978]: E0225 07:09:01.721623 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c4ce63-060f-4f53-8e1a-a70b6ba97255" containerName="nova-cell1-conductor-db-sync" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.721631 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c4ce63-060f-4f53-8e1a-a70b6ba97255" containerName="nova-cell1-conductor-db-sync" Feb 25 07:09:01 crc kubenswrapper[4978]: E0225 07:09:01.721650 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" containerName="nova-metadata-log" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.721658 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" containerName="nova-metadata-log" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.721866 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c4ce63-060f-4f53-8e1a-a70b6ba97255" containerName="nova-cell1-conductor-db-sync" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.721886 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" containerName="nova-metadata-metadata" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.721902 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2cdb535-d30e-445f-97f9-259185cb4223" containerName="nova-manage" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.721924 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" containerName="nova-metadata-log" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.722640 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.725591 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.728633 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.729826 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.741138 4978 scope.go:117] "RemoveContainer" containerID="ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f" Feb 25 07:09:01 crc kubenswrapper[4978]: E0225 07:09:01.741543 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f\": container with ID starting with ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f not found: ID does not exist" containerID="ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.741582 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f"} err="failed to get container status \"ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f\": rpc error: code = NotFound desc = could not find container \"ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f\": container with ID starting with ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f not found: ID does not exist" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.741612 4978 scope.go:117] "RemoveContainer" containerID="417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb" Feb 25 07:09:01 crc kubenswrapper[4978]: E0225 07:09:01.741852 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb\": container with ID starting with 417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb not found: ID does not exist" containerID="417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.741879 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb"} err="failed to get container status \"417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb\": rpc error: code = NotFound desc = could not find container \"417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb\": container with ID starting with 417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb not found: ID does not exist" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.741895 4978 scope.go:117] "RemoveContainer" containerID="ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.742082 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f"} err="failed to get container status \"ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f\": rpc error: code = NotFound desc = could not find container \"ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f\": container with ID starting with ec7756f8f063a40458f0c38a9d70b6820a51889e62ab3b41c15e8049108f2e5f not found: ID does not exist" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.742099 4978 scope.go:117] "RemoveContainer" containerID="417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.744086 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb"} err="failed to get container status \"417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb\": rpc error: code = NotFound desc = could not find container \"417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb\": container with ID starting with 417cc79210de3172ab8e2658a4f5e3b8915dc837258ddbfa14c0a82cc8b20eeb not found: ID does not exist" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.894909 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-config\") pod \"32b6e3c5-3010-4286-b0aa-a2bd20517933\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.894982 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-swift-storage-0\") pod \"32b6e3c5-3010-4286-b0aa-a2bd20517933\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.895168 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-svc\") pod \"32b6e3c5-3010-4286-b0aa-a2bd20517933\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.895230 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-nb\") pod \"32b6e3c5-3010-4286-b0aa-a2bd20517933\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.895252 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-sb\") pod \"32b6e3c5-3010-4286-b0aa-a2bd20517933\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.895280 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckxj2\" (UniqueName: \"kubernetes.io/projected/32b6e3c5-3010-4286-b0aa-a2bd20517933-kube-api-access-ckxj2\") pod \"32b6e3c5-3010-4286-b0aa-a2bd20517933\" (UID: \"32b6e3c5-3010-4286-b0aa-a2bd20517933\") " Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.895533 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krlk2\" (UniqueName: \"kubernetes.io/projected/8a6b84e2-353d-47ef-82be-c7e182533a57-kube-api-access-krlk2\") pod \"nova-cell1-conductor-0\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.895714 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.895834 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.899453 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32b6e3c5-3010-4286-b0aa-a2bd20517933-kube-api-access-ckxj2" (OuterVolumeSpecName: "kube-api-access-ckxj2") pod "32b6e3c5-3010-4286-b0aa-a2bd20517933" (UID: "32b6e3c5-3010-4286-b0aa-a2bd20517933"). InnerVolumeSpecName "kube-api-access-ckxj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.936856 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "32b6e3c5-3010-4286-b0aa-a2bd20517933" (UID: "32b6e3c5-3010-4286-b0aa-a2bd20517933"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.939562 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "32b6e3c5-3010-4286-b0aa-a2bd20517933" (UID: "32b6e3c5-3010-4286-b0aa-a2bd20517933"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.940618 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "32b6e3c5-3010-4286-b0aa-a2bd20517933" (UID: "32b6e3c5-3010-4286-b0aa-a2bd20517933"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.941421 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "32b6e3c5-3010-4286-b0aa-a2bd20517933" (UID: "32b6e3c5-3010-4286-b0aa-a2bd20517933"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.970061 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-config" (OuterVolumeSpecName: "config") pod "32b6e3c5-3010-4286-b0aa-a2bd20517933" (UID: "32b6e3c5-3010-4286-b0aa-a2bd20517933"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.989497 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.997777 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.997906 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krlk2\" (UniqueName: \"kubernetes.io/projected/8a6b84e2-353d-47ef-82be-c7e182533a57-kube-api-access-krlk2\") pod \"nova-cell1-conductor-0\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.997957 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.998007 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.998018 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.998027 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.998036 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckxj2\" (UniqueName: \"kubernetes.io/projected/32b6e3c5-3010-4286-b0aa-a2bd20517933-kube-api-access-ckxj2\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.998044 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.998056 4978 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/32b6e3c5-3010-4286-b0aa-a2bd20517933-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:01 crc kubenswrapper[4978]: I0225 07:09:01.999416 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.002597 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.026443 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:02 crc kubenswrapper[4978]: E0225 07:09:02.027233 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b6e3c5-3010-4286-b0aa-a2bd20517933" containerName="init" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.027252 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b6e3c5-3010-4286-b0aa-a2bd20517933" containerName="init" Feb 25 07:09:02 crc kubenswrapper[4978]: E0225 07:09:02.027280 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b6e3c5-3010-4286-b0aa-a2bd20517933" containerName="dnsmasq-dns" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.027287 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b6e3c5-3010-4286-b0aa-a2bd20517933" containerName="dnsmasq-dns" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.027666 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="32b6e3c5-3010-4286-b0aa-a2bd20517933" containerName="dnsmasq-dns" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.029241 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.032078 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.032896 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.036986 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.042188 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.044091 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krlk2\" (UniqueName: \"kubernetes.io/projected/8a6b84e2-353d-47ef-82be-c7e182533a57-kube-api-access-krlk2\") pod \"nova-cell1-conductor-0\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.201284 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8876c6c-57b4-421d-bc34-a70080f86c63-logs\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.201413 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-config-data\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.201473 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.201545 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trlbn\" (UniqueName: \"kubernetes.io/projected/b8876c6c-57b4-421d-bc34-a70080f86c63-kube-api-access-trlbn\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.201722 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.303195 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8876c6c-57b4-421d-bc34-a70080f86c63-logs\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.303537 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-config-data\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.303570 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.303621 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trlbn\" (UniqueName: \"kubernetes.io/projected/b8876c6c-57b4-421d-bc34-a70080f86c63-kube-api-access-trlbn\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.303730 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.304030 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8876c6c-57b4-421d-bc34-a70080f86c63-logs\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.307901 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-config-data\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.308231 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.310499 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.321469 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trlbn\" (UniqueName: \"kubernetes.io/projected/b8876c6c-57b4-421d-bc34-a70080f86c63-kube-api-access-trlbn\") pod \"nova-metadata-0\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.337249 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.355097 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.663934 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" event={"ID":"32b6e3c5-3010-4286-b0aa-a2bd20517933","Type":"ContainerDied","Data":"b385d710d3f2187ec84337f9dd4741031041c92d7cb4ca69bdee370613eec7db"} Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.664193 4978 scope.go:117] "RemoveContainer" containerID="8b4f5ceb88e5882edab1442efa5555b2f11dddb6c403904c3906cd14a276a6f4" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.664021 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d4ffdd57-jtdkj" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.666037 4978 generic.go:334] "Generic (PLEG): container finished" podID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerID="99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7" exitCode=143 Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.666083 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e93f8648-b297-4a22-93fa-61d3ffe59bff","Type":"ContainerDied","Data":"99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7"} Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.688735 4978 scope.go:117] "RemoveContainer" containerID="cbffce9c96e1a8b231f5183e40f0fa24451e08a5a2711947712f77aa834fbc2d" Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.761778 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d4ffdd57-jtdkj"] Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.769145 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d4ffdd57-jtdkj"] Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.821068 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 07:09:02 crc kubenswrapper[4978]: W0225 07:09:02.823364 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a6b84e2_353d_47ef_82be_c7e182533a57.slice/crio-e9b14ecbd2a5d502ac04dc0d81c1735e4284d9d51f459cf8d68919a3d771888c WatchSource:0}: Error finding container e9b14ecbd2a5d502ac04dc0d81c1735e4284d9d51f459cf8d68919a3d771888c: Status 404 returned error can't find the container with id e9b14ecbd2a5d502ac04dc0d81c1735e4284d9d51f459cf8d68919a3d771888c Feb 25 07:09:02 crc kubenswrapper[4978]: I0225 07:09:02.959098 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:03 crc kubenswrapper[4978]: I0225 07:09:03.348425 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32b6e3c5-3010-4286-b0aa-a2bd20517933" path="/var/lib/kubelet/pods/32b6e3c5-3010-4286-b0aa-a2bd20517933/volumes" Feb 25 07:09:03 crc kubenswrapper[4978]: I0225 07:09:03.349734 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d06b54a6-6f8d-4ed0-9b15-d09a5a187d18" path="/var/lib/kubelet/pods/d06b54a6-6f8d-4ed0-9b15-d09a5a187d18/volumes" Feb 25 07:09:03 crc kubenswrapper[4978]: I0225 07:09:03.691082 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b8876c6c-57b4-421d-bc34-a70080f86c63","Type":"ContainerStarted","Data":"ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f"} Feb 25 07:09:03 crc kubenswrapper[4978]: I0225 07:09:03.691392 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b8876c6c-57b4-421d-bc34-a70080f86c63","Type":"ContainerStarted","Data":"2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b"} Feb 25 07:09:03 crc kubenswrapper[4978]: I0225 07:09:03.691419 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b8876c6c-57b4-421d-bc34-a70080f86c63","Type":"ContainerStarted","Data":"a03a801508afc147f9267b051b4db4877175d115e1bd7121f944f0859392e0f7"} Feb 25 07:09:03 crc kubenswrapper[4978]: I0225 07:09:03.694334 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8a6b84e2-353d-47ef-82be-c7e182533a57","Type":"ContainerStarted","Data":"5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc"} Feb 25 07:09:03 crc kubenswrapper[4978]: I0225 07:09:03.694381 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8a6b84e2-353d-47ef-82be-c7e182533a57","Type":"ContainerStarted","Data":"e9b14ecbd2a5d502ac04dc0d81c1735e4284d9d51f459cf8d68919a3d771888c"} Feb 25 07:09:03 crc kubenswrapper[4978]: I0225 07:09:03.694759 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:03 crc kubenswrapper[4978]: I0225 07:09:03.717355 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7173395339999997 podStartE2EDuration="2.717339534s" podCreationTimestamp="2026-02-25 07:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:03.709028192 +0000 UTC m=+1437.148284661" watchObservedRunningTime="2026-02-25 07:09:03.717339534 +0000 UTC m=+1437.156595993" Feb 25 07:09:03 crc kubenswrapper[4978]: I0225 07:09:03.732838 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.732823041 podStartE2EDuration="2.732823041s" podCreationTimestamp="2026-02-25 07:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:03.729281939 +0000 UTC m=+1437.168538418" watchObservedRunningTime="2026-02-25 07:09:03.732823041 +0000 UTC m=+1437.172079500" Feb 25 07:09:05 crc kubenswrapper[4978]: E0225 07:09:05.792033 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:09:05 crc kubenswrapper[4978]: E0225 07:09:05.794192 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:09:05 crc kubenswrapper[4978]: E0225 07:09:05.799513 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:09:05 crc kubenswrapper[4978]: E0225 07:09:05.799646 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="1e877097-9faf-4497-9328-a46e828f79eb" containerName="nova-scheduler-scheduler" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.506532 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.607936 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-combined-ca-bundle\") pod \"1e877097-9faf-4497-9328-a46e828f79eb\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.608110 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxd9t\" (UniqueName: \"kubernetes.io/projected/1e877097-9faf-4497-9328-a46e828f79eb-kube-api-access-rxd9t\") pod \"1e877097-9faf-4497-9328-a46e828f79eb\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.608178 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-config-data\") pod \"1e877097-9faf-4497-9328-a46e828f79eb\" (UID: \"1e877097-9faf-4497-9328-a46e828f79eb\") " Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.614749 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e877097-9faf-4497-9328-a46e828f79eb-kube-api-access-rxd9t" (OuterVolumeSpecName: "kube-api-access-rxd9t") pod "1e877097-9faf-4497-9328-a46e828f79eb" (UID: "1e877097-9faf-4497-9328-a46e828f79eb"). InnerVolumeSpecName "kube-api-access-rxd9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.637176 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-config-data" (OuterVolumeSpecName: "config-data") pod "1e877097-9faf-4497-9328-a46e828f79eb" (UID: "1e877097-9faf-4497-9328-a46e828f79eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.653436 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e877097-9faf-4497-9328-a46e828f79eb" (UID: "1e877097-9faf-4497-9328-a46e828f79eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.711651 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxd9t\" (UniqueName: \"kubernetes.io/projected/1e877097-9faf-4497-9328-a46e828f79eb-kube-api-access-rxd9t\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.712203 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.712221 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e877097-9faf-4497-9328-a46e828f79eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.747292 4978 generic.go:334] "Generic (PLEG): container finished" podID="1e877097-9faf-4497-9328-a46e828f79eb" containerID="90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" exitCode=0 Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.747330 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e877097-9faf-4497-9328-a46e828f79eb","Type":"ContainerDied","Data":"90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268"} Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.747343 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.747352 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1e877097-9faf-4497-9328-a46e828f79eb","Type":"ContainerDied","Data":"d2014a38192231d7483d5ab7cbfa79ed27b6f085d9127dbed18a9ba37ffce361"} Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.747381 4978 scope.go:117] "RemoveContainer" containerID="90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.771082 4978 scope.go:117] "RemoveContainer" containerID="90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" Feb 25 07:09:06 crc kubenswrapper[4978]: E0225 07:09:06.771450 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268\": container with ID starting with 90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268 not found: ID does not exist" containerID="90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.771484 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268"} err="failed to get container status \"90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268\": rpc error: code = NotFound desc = could not find container \"90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268\": container with ID starting with 90127d6f35e602204e02c2b4a90e845a4e83cce777bc979cfc8d65af004cd268 not found: ID does not exist" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.780836 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.804332 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.815109 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:06 crc kubenswrapper[4978]: E0225 07:09:06.815577 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e877097-9faf-4497-9328-a46e828f79eb" containerName="nova-scheduler-scheduler" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.815595 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e877097-9faf-4497-9328-a46e828f79eb" containerName="nova-scheduler-scheduler" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.815827 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e877097-9faf-4497-9328-a46e828f79eb" containerName="nova-scheduler-scheduler" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.816616 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.819596 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.825598 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.915532 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-config-data\") pod \"nova-scheduler-0\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.915667 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:06 crc kubenswrapper[4978]: I0225 07:09:06.915746 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqgdb\" (UniqueName: \"kubernetes.io/projected/ef038d4c-269b-407d-95db-43e90be12da7-kube-api-access-xqgdb\") pod \"nova-scheduler-0\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.017506 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-config-data\") pod \"nova-scheduler-0\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.017587 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.017631 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqgdb\" (UniqueName: \"kubernetes.io/projected/ef038d4c-269b-407d-95db-43e90be12da7-kube-api-access-xqgdb\") pod \"nova-scheduler-0\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.023064 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.024300 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-config-data\") pod \"nova-scheduler-0\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.035358 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqgdb\" (UniqueName: \"kubernetes.io/projected/ef038d4c-269b-407d-95db-43e90be12da7-kube-api-access-xqgdb\") pod \"nova-scheduler-0\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.133065 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.345177 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e877097-9faf-4497-9328-a46e828f79eb" path="/var/lib/kubelet/pods/1e877097-9faf-4497-9328-a46e828f79eb/volumes" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.355753 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.357064 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.512442 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.635033 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e93f8648-b297-4a22-93fa-61d3ffe59bff-logs\") pod \"e93f8648-b297-4a22-93fa-61d3ffe59bff\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.635277 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws8f4\" (UniqueName: \"kubernetes.io/projected/e93f8648-b297-4a22-93fa-61d3ffe59bff-kube-api-access-ws8f4\") pod \"e93f8648-b297-4a22-93fa-61d3ffe59bff\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.635337 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-combined-ca-bundle\") pod \"e93f8648-b297-4a22-93fa-61d3ffe59bff\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.635560 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-config-data\") pod \"e93f8648-b297-4a22-93fa-61d3ffe59bff\" (UID: \"e93f8648-b297-4a22-93fa-61d3ffe59bff\") " Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.635700 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e93f8648-b297-4a22-93fa-61d3ffe59bff-logs" (OuterVolumeSpecName: "logs") pod "e93f8648-b297-4a22-93fa-61d3ffe59bff" (UID: "e93f8648-b297-4a22-93fa-61d3ffe59bff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.637199 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e93f8648-b297-4a22-93fa-61d3ffe59bff-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.643701 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e93f8648-b297-4a22-93fa-61d3ffe59bff-kube-api-access-ws8f4" (OuterVolumeSpecName: "kube-api-access-ws8f4") pod "e93f8648-b297-4a22-93fa-61d3ffe59bff" (UID: "e93f8648-b297-4a22-93fa-61d3ffe59bff"). InnerVolumeSpecName "kube-api-access-ws8f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.665866 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.687592 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e93f8648-b297-4a22-93fa-61d3ffe59bff" (UID: "e93f8648-b297-4a22-93fa-61d3ffe59bff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.696460 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-config-data" (OuterVolumeSpecName: "config-data") pod "e93f8648-b297-4a22-93fa-61d3ffe59bff" (UID: "e93f8648-b297-4a22-93fa-61d3ffe59bff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.739448 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.739511 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws8f4\" (UniqueName: \"kubernetes.io/projected/e93f8648-b297-4a22-93fa-61d3ffe59bff-kube-api-access-ws8f4\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.739536 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e93f8648-b297-4a22-93fa-61d3ffe59bff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.765506 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef038d4c-269b-407d-95db-43e90be12da7","Type":"ContainerStarted","Data":"660fca13a8a30130884e543bdc7c942ab8f2443e00b844c995cb12021ced081c"} Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.770634 4978 generic.go:334] "Generic (PLEG): container finished" podID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerID="79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5" exitCode=0 Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.770705 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e93f8648-b297-4a22-93fa-61d3ffe59bff","Type":"ContainerDied","Data":"79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5"} Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.770770 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e93f8648-b297-4a22-93fa-61d3ffe59bff","Type":"ContainerDied","Data":"ff97ad5acf565fd7d1506bf40a98318eeb3a3d22d04d1f88a3f1e1a79c33b0e9"} Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.770778 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.770853 4978 scope.go:117] "RemoveContainer" containerID="79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.801875 4978 scope.go:117] "RemoveContainer" containerID="99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.833425 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.844240 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.853733 4978 scope.go:117] "RemoveContainer" containerID="79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.853810 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:07 crc kubenswrapper[4978]: E0225 07:09:07.857800 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5\": container with ID starting with 79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5 not found: ID does not exist" containerID="79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.857901 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5"} err="failed to get container status \"79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5\": rpc error: code = NotFound desc = could not find container \"79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5\": container with ID starting with 79274ebd89f92d8cfe21c506753aff77ac834ac49639edac6830a984ba3a85c5 not found: ID does not exist" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.857940 4978 scope.go:117] "RemoveContainer" containerID="99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7" Feb 25 07:09:07 crc kubenswrapper[4978]: E0225 07:09:07.869569 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7\": container with ID starting with 99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7 not found: ID does not exist" containerID="99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.869622 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7"} err="failed to get container status \"99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7\": rpc error: code = NotFound desc = could not find container \"99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7\": container with ID starting with 99bf5c179eaf3a6668fb98cddf31cc6ebb9d9f1d1f4860134dc0bb68aac70fc7 not found: ID does not exist" Feb 25 07:09:07 crc kubenswrapper[4978]: E0225 07:09:07.903989 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerName="nova-api-log" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.904033 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerName="nova-api-log" Feb 25 07:09:07 crc kubenswrapper[4978]: E0225 07:09:07.904061 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerName="nova-api-api" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.904071 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerName="nova-api-api" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.904885 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerName="nova-api-api" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.904906 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" containerName="nova-api-log" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.906623 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.906731 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:07 crc kubenswrapper[4978]: I0225 07:09:07.908848 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.045530 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.045608 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-config-data\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.045843 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr2tw\" (UniqueName: \"kubernetes.io/projected/50b68f0d-5478-49ba-9671-0017c6edc37b-kube-api-access-hr2tw\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.045968 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b68f0d-5478-49ba-9671-0017c6edc37b-logs\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.147774 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr2tw\" (UniqueName: \"kubernetes.io/projected/50b68f0d-5478-49ba-9671-0017c6edc37b-kube-api-access-hr2tw\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.147868 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b68f0d-5478-49ba-9671-0017c6edc37b-logs\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.148013 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.148083 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-config-data\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.148358 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b68f0d-5478-49ba-9671-0017c6edc37b-logs\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.165677 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.165884 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-config-data\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.170840 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr2tw\" (UniqueName: \"kubernetes.io/projected/50b68f0d-5478-49ba-9671-0017c6edc37b-kube-api-access-hr2tw\") pod \"nova-api-0\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.231901 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.756116 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.786460 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b68f0d-5478-49ba-9671-0017c6edc37b","Type":"ContainerStarted","Data":"37bccae3093bfd6dc38a73bf86adba09b88c393016dd44405c2549bb3eac11b5"} Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.790549 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef038d4c-269b-407d-95db-43e90be12da7","Type":"ContainerStarted","Data":"53f477948bc0c813efe4de0ab156f490c6b1adba3139158be8a034b57455194b"} Feb 25 07:09:08 crc kubenswrapper[4978]: I0225 07:09:08.809826 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.809808067 podStartE2EDuration="2.809808067s" podCreationTimestamp="2026-02-25 07:09:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:08.805840572 +0000 UTC m=+1442.245097031" watchObservedRunningTime="2026-02-25 07:09:08.809808067 +0000 UTC m=+1442.249064526" Feb 25 07:09:09 crc kubenswrapper[4978]: I0225 07:09:09.341738 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e93f8648-b297-4a22-93fa-61d3ffe59bff" path="/var/lib/kubelet/pods/e93f8648-b297-4a22-93fa-61d3ffe59bff/volumes" Feb 25 07:09:09 crc kubenswrapper[4978]: I0225 07:09:09.806727 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b68f0d-5478-49ba-9671-0017c6edc37b","Type":"ContainerStarted","Data":"8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2"} Feb 25 07:09:09 crc kubenswrapper[4978]: I0225 07:09:09.806793 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b68f0d-5478-49ba-9671-0017c6edc37b","Type":"ContainerStarted","Data":"3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148"} Feb 25 07:09:12 crc kubenswrapper[4978]: I0225 07:09:12.133341 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 25 07:09:12 crc kubenswrapper[4978]: I0225 07:09:12.355577 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 07:09:12 crc kubenswrapper[4978]: I0225 07:09:12.355639 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 07:09:12 crc kubenswrapper[4978]: I0225 07:09:12.390900 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 25 07:09:12 crc kubenswrapper[4978]: I0225 07:09:12.429321 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=5.429291616 podStartE2EDuration="5.429291616s" podCreationTimestamp="2026-02-25 07:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:09.842485412 +0000 UTC m=+1443.281741901" watchObservedRunningTime="2026-02-25 07:09:12.429291616 +0000 UTC m=+1445.868548105" Feb 25 07:09:13 crc kubenswrapper[4978]: I0225 07:09:13.372579 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 07:09:13 crc kubenswrapper[4978]: I0225 07:09:13.372678 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 07:09:15 crc kubenswrapper[4978]: I0225 07:09:15.016174 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 25 07:09:17 crc kubenswrapper[4978]: I0225 07:09:17.134085 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 25 07:09:17 crc kubenswrapper[4978]: I0225 07:09:17.188248 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 25 07:09:18 crc kubenswrapper[4978]: I0225 07:09:18.173160 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 25 07:09:18 crc kubenswrapper[4978]: I0225 07:09:18.233665 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 07:09:18 crc kubenswrapper[4978]: I0225 07:09:18.233710 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 07:09:18 crc kubenswrapper[4978]: I0225 07:09:18.988356 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:09:18 crc kubenswrapper[4978]: I0225 07:09:18.988792 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="fbf80328-4523-49ff-b390-0e8b3dd9cc02" containerName="kube-state-metrics" containerID="cri-o://037e82b5ff3e20c28f0050cfb9295e40e3a11fad5596b116d220432860b85e2d" gracePeriod=30 Feb 25 07:09:19 crc kubenswrapper[4978]: I0225 07:09:19.159119 4978 generic.go:334] "Generic (PLEG): container finished" podID="fbf80328-4523-49ff-b390-0e8b3dd9cc02" containerID="037e82b5ff3e20c28f0050cfb9295e40e3a11fad5596b116d220432860b85e2d" exitCode=2 Feb 25 07:09:19 crc kubenswrapper[4978]: I0225 07:09:19.159223 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fbf80328-4523-49ff-b390-0e8b3dd9cc02","Type":"ContainerDied","Data":"037e82b5ff3e20c28f0050cfb9295e40e3a11fad5596b116d220432860b85e2d"} Feb 25 07:09:19 crc kubenswrapper[4978]: I0225 07:09:19.321313 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 07:09:19 crc kubenswrapper[4978]: I0225 07:09:19.321314 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 07:09:19 crc kubenswrapper[4978]: I0225 07:09:19.513817 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 07:09:19 crc kubenswrapper[4978]: I0225 07:09:19.609101 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjnfp\" (UniqueName: \"kubernetes.io/projected/fbf80328-4523-49ff-b390-0e8b3dd9cc02-kube-api-access-bjnfp\") pod \"fbf80328-4523-49ff-b390-0e8b3dd9cc02\" (UID: \"fbf80328-4523-49ff-b390-0e8b3dd9cc02\") " Feb 25 07:09:19 crc kubenswrapper[4978]: I0225 07:09:19.615084 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbf80328-4523-49ff-b390-0e8b3dd9cc02-kube-api-access-bjnfp" (OuterVolumeSpecName: "kube-api-access-bjnfp") pod "fbf80328-4523-49ff-b390-0e8b3dd9cc02" (UID: "fbf80328-4523-49ff-b390-0e8b3dd9cc02"). InnerVolumeSpecName "kube-api-access-bjnfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:19 crc kubenswrapper[4978]: I0225 07:09:19.711750 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjnfp\" (UniqueName: \"kubernetes.io/projected/fbf80328-4523-49ff-b390-0e8b3dd9cc02-kube-api-access-bjnfp\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.229087 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fbf80328-4523-49ff-b390-0e8b3dd9cc02","Type":"ContainerDied","Data":"80c3b9c65701b46a40069afdb55e90b6da48c0e1e407e3ce8974dfc61a35413c"} Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.229138 4978 scope.go:117] "RemoveContainer" containerID="037e82b5ff3e20c28f0050cfb9295e40e3a11fad5596b116d220432860b85e2d" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.229266 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.275061 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.287051 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.304112 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:09:20 crc kubenswrapper[4978]: E0225 07:09:20.305007 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbf80328-4523-49ff-b390-0e8b3dd9cc02" containerName="kube-state-metrics" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.305038 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbf80328-4523-49ff-b390-0e8b3dd9cc02" containerName="kube-state-metrics" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.305397 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbf80328-4523-49ff-b390-0e8b3dd9cc02" containerName="kube-state-metrics" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.306431 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.309642 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.309814 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.311329 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.437376 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.438119 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jh5v\" (UniqueName: \"kubernetes.io/projected/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-api-access-9jh5v\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.438284 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.438395 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.539834 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.540146 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jh5v\" (UniqueName: \"kubernetes.io/projected/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-api-access-9jh5v\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.540272 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.540362 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.545509 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.545725 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.546609 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.564584 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jh5v\" (UniqueName: \"kubernetes.io/projected/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-api-access-9jh5v\") pod \"kube-state-metrics-0\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.629161 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.704630 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.704918 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="ceilometer-central-agent" containerID="cri-o://0fab927ded83dc237f13d657ba5e00c5469912aaca90ef0539d5006a4a702098" gracePeriod=30 Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.704939 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="proxy-httpd" containerID="cri-o://00360f4efb86fb57db32d1190a034b7291ca7ec862281fdebb829d0d236c5560" gracePeriod=30 Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.705041 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="ceilometer-notification-agent" containerID="cri-o://f077e044a528cd13bb7a6f089c9729303595fa1f4b7b0316534c5443b54448c1" gracePeriod=30 Feb 25 07:09:20 crc kubenswrapper[4978]: I0225 07:09:20.705052 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="sg-core" containerID="cri-o://7e58ad9997f17eb86b9ef45aba593e1784829f72024c363155d853f453ed3c18" gracePeriod=30 Feb 25 07:09:21 crc kubenswrapper[4978]: I0225 07:09:21.116656 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:09:21 crc kubenswrapper[4978]: I0225 07:09:21.241907 4978 generic.go:334] "Generic (PLEG): container finished" podID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerID="00360f4efb86fb57db32d1190a034b7291ca7ec862281fdebb829d0d236c5560" exitCode=0 Feb 25 07:09:21 crc kubenswrapper[4978]: I0225 07:09:21.241962 4978 generic.go:334] "Generic (PLEG): container finished" podID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerID="7e58ad9997f17eb86b9ef45aba593e1784829f72024c363155d853f453ed3c18" exitCode=2 Feb 25 07:09:21 crc kubenswrapper[4978]: I0225 07:09:21.241980 4978 generic.go:334] "Generic (PLEG): container finished" podID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerID="0fab927ded83dc237f13d657ba5e00c5469912aaca90ef0539d5006a4a702098" exitCode=0 Feb 25 07:09:21 crc kubenswrapper[4978]: I0225 07:09:21.241987 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07069098-bc61-49f2-bdab-55e8e72f4faa","Type":"ContainerDied","Data":"00360f4efb86fb57db32d1190a034b7291ca7ec862281fdebb829d0d236c5560"} Feb 25 07:09:21 crc kubenswrapper[4978]: I0225 07:09:21.242048 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07069098-bc61-49f2-bdab-55e8e72f4faa","Type":"ContainerDied","Data":"7e58ad9997f17eb86b9ef45aba593e1784829f72024c363155d853f453ed3c18"} Feb 25 07:09:21 crc kubenswrapper[4978]: I0225 07:09:21.242068 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07069098-bc61-49f2-bdab-55e8e72f4faa","Type":"ContainerDied","Data":"0fab927ded83dc237f13d657ba5e00c5469912aaca90ef0539d5006a4a702098"} Feb 25 07:09:21 crc kubenswrapper[4978]: I0225 07:09:21.243358 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3be5007a-36e1-4b55-a8d8-5fd560efb8fe","Type":"ContainerStarted","Data":"9a69be620040b4f2389b269e72f36725d52d5fe065d63127c807f79176956fa0"} Feb 25 07:09:21 crc kubenswrapper[4978]: I0225 07:09:21.343817 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fbf80328-4523-49ff-b390-0e8b3dd9cc02" path="/var/lib/kubelet/pods/fbf80328-4523-49ff-b390-0e8b3dd9cc02/volumes" Feb 25 07:09:22 crc kubenswrapper[4978]: I0225 07:09:22.261995 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3be5007a-36e1-4b55-a8d8-5fd560efb8fe","Type":"ContainerStarted","Data":"78414f7a4e1d623e9a11a84d9c8ca28b022bcdb53fed8a932a715712c16abf12"} Feb 25 07:09:22 crc kubenswrapper[4978]: I0225 07:09:22.262459 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 25 07:09:22 crc kubenswrapper[4978]: I0225 07:09:22.287453 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.903991366 podStartE2EDuration="2.287430275s" podCreationTimestamp="2026-02-25 07:09:20 +0000 UTC" firstStartedPulling="2026-02-25 07:09:21.116363465 +0000 UTC m=+1454.555619964" lastFinishedPulling="2026-02-25 07:09:21.499802404 +0000 UTC m=+1454.939058873" observedRunningTime="2026-02-25 07:09:22.284757431 +0000 UTC m=+1455.724013920" watchObservedRunningTime="2026-02-25 07:09:22.287430275 +0000 UTC m=+1455.726686774" Feb 25 07:09:22 crc kubenswrapper[4978]: I0225 07:09:22.363785 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 07:09:22 crc kubenswrapper[4978]: I0225 07:09:22.368975 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 07:09:22 crc kubenswrapper[4978]: I0225 07:09:22.372178 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.273893 4978 generic.go:334] "Generic (PLEG): container finished" podID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerID="f077e044a528cd13bb7a6f089c9729303595fa1f4b7b0316534c5443b54448c1" exitCode=0 Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.273962 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07069098-bc61-49f2-bdab-55e8e72f4faa","Type":"ContainerDied","Data":"f077e044a528cd13bb7a6f089c9729303595fa1f4b7b0316534c5443b54448c1"} Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.279728 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.473724 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.602079 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-combined-ca-bundle\") pod \"07069098-bc61-49f2-bdab-55e8e72f4faa\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.602125 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-log-httpd\") pod \"07069098-bc61-49f2-bdab-55e8e72f4faa\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.602200 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfbvh\" (UniqueName: \"kubernetes.io/projected/07069098-bc61-49f2-bdab-55e8e72f4faa-kube-api-access-qfbvh\") pod \"07069098-bc61-49f2-bdab-55e8e72f4faa\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.602227 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-run-httpd\") pod \"07069098-bc61-49f2-bdab-55e8e72f4faa\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.602257 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-config-data\") pod \"07069098-bc61-49f2-bdab-55e8e72f4faa\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.602340 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-scripts\") pod \"07069098-bc61-49f2-bdab-55e8e72f4faa\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.602451 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-sg-core-conf-yaml\") pod \"07069098-bc61-49f2-bdab-55e8e72f4faa\" (UID: \"07069098-bc61-49f2-bdab-55e8e72f4faa\") " Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.602606 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "07069098-bc61-49f2-bdab-55e8e72f4faa" (UID: "07069098-bc61-49f2-bdab-55e8e72f4faa"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.603155 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "07069098-bc61-49f2-bdab-55e8e72f4faa" (UID: "07069098-bc61-49f2-bdab-55e8e72f4faa"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.603215 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.625872 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-scripts" (OuterVolumeSpecName: "scripts") pod "07069098-bc61-49f2-bdab-55e8e72f4faa" (UID: "07069098-bc61-49f2-bdab-55e8e72f4faa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.638882 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07069098-bc61-49f2-bdab-55e8e72f4faa-kube-api-access-qfbvh" (OuterVolumeSpecName: "kube-api-access-qfbvh") pod "07069098-bc61-49f2-bdab-55e8e72f4faa" (UID: "07069098-bc61-49f2-bdab-55e8e72f4faa"). InnerVolumeSpecName "kube-api-access-qfbvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.653523 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "07069098-bc61-49f2-bdab-55e8e72f4faa" (UID: "07069098-bc61-49f2-bdab-55e8e72f4faa"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.700809 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07069098-bc61-49f2-bdab-55e8e72f4faa" (UID: "07069098-bc61-49f2-bdab-55e8e72f4faa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.705609 4978 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.705651 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.705665 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfbvh\" (UniqueName: \"kubernetes.io/projected/07069098-bc61-49f2-bdab-55e8e72f4faa-kube-api-access-qfbvh\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.705681 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07069098-bc61-49f2-bdab-55e8e72f4faa-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.705692 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.717962 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-config-data" (OuterVolumeSpecName: "config-data") pod "07069098-bc61-49f2-bdab-55e8e72f4faa" (UID: "07069098-bc61-49f2-bdab-55e8e72f4faa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:23 crc kubenswrapper[4978]: I0225 07:09:23.807678 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07069098-bc61-49f2-bdab-55e8e72f4faa-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.290456 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07069098-bc61-49f2-bdab-55e8e72f4faa","Type":"ContainerDied","Data":"9bb5e9afa4e86fd8fe93d6ef9285750c8cb1729a3a67348dc857b7a484792a13"} Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.290520 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.290902 4978 scope.go:117] "RemoveContainer" containerID="00360f4efb86fb57db32d1190a034b7291ca7ec862281fdebb829d0d236c5560" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.326888 4978 scope.go:117] "RemoveContainer" containerID="7e58ad9997f17eb86b9ef45aba593e1784829f72024c363155d853f453ed3c18" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.351765 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.358384 4978 scope.go:117] "RemoveContainer" containerID="f077e044a528cd13bb7a6f089c9729303595fa1f4b7b0316534c5443b54448c1" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.364671 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.384245 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:24 crc kubenswrapper[4978]: E0225 07:09:24.384782 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="proxy-httpd" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.384801 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="proxy-httpd" Feb 25 07:09:24 crc kubenswrapper[4978]: E0225 07:09:24.384837 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="ceilometer-central-agent" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.384846 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="ceilometer-central-agent" Feb 25 07:09:24 crc kubenswrapper[4978]: E0225 07:09:24.384867 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="ceilometer-notification-agent" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.384875 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="ceilometer-notification-agent" Feb 25 07:09:24 crc kubenswrapper[4978]: E0225 07:09:24.384898 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="sg-core" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.384905 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="sg-core" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.385112 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="ceilometer-central-agent" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.385131 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="sg-core" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.385145 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="ceilometer-notification-agent" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.385161 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" containerName="proxy-httpd" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.387039 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.391497 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.391619 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.391907 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.393537 4978 scope.go:117] "RemoveContainer" containerID="0fab927ded83dc237f13d657ba5e00c5469912aaca90ef0539d5006a4a702098" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.396247 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.423476 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.423557 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwfhh\" (UniqueName: \"kubernetes.io/projected/aa5bc4db-5446-4147-9321-23746eb83646-kube-api-access-hwfhh\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.423653 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.423695 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-run-httpd\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.423720 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.423857 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-config-data\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.424186 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-log-httpd\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.424259 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-scripts\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.525839 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-scripts\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.525948 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.526026 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwfhh\" (UniqueName: \"kubernetes.io/projected/aa5bc4db-5446-4147-9321-23746eb83646-kube-api-access-hwfhh\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.526312 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.526423 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-run-httpd\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.526500 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.526848 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-config-data\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.526910 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-log-httpd\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.529178 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-run-httpd\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.530428 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-scripts\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.530958 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-log-httpd\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.531852 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.533135 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.535128 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-config-data\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.536241 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.557549 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwfhh\" (UniqueName: \"kubernetes.io/projected/aa5bc4db-5446-4147-9321-23746eb83646-kube-api-access-hwfhh\") pod \"ceilometer-0\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " pod="openstack/ceilometer-0" Feb 25 07:09:24 crc kubenswrapper[4978]: I0225 07:09:24.709612 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:09:25 crc kubenswrapper[4978]: I0225 07:09:25.172259 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:25 crc kubenswrapper[4978]: I0225 07:09:25.300984 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa5bc4db-5446-4147-9321-23746eb83646","Type":"ContainerStarted","Data":"883c779c5d06fdeb909aec2baa0440c2f97a8d9ec96c923407229d46a3da9d41"} Feb 25 07:09:25 crc kubenswrapper[4978]: I0225 07:09:25.339031 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07069098-bc61-49f2-bdab-55e8e72f4faa" path="/var/lib/kubelet/pods/07069098-bc61-49f2-bdab-55e8e72f4faa/volumes" Feb 25 07:09:25 crc kubenswrapper[4978]: I0225 07:09:25.911960 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:25 crc kubenswrapper[4978]: I0225 07:09:25.952918 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-config-data\") pod \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " Feb 25 07:09:25 crc kubenswrapper[4978]: I0225 07:09:25.953084 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-combined-ca-bundle\") pod \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " Feb 25 07:09:25 crc kubenswrapper[4978]: I0225 07:09:25.953182 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmzjp\" (UniqueName: \"kubernetes.io/projected/4ef26333-cc4a-4e32-bc18-72ffa6735f71-kube-api-access-hmzjp\") pod \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\" (UID: \"4ef26333-cc4a-4e32-bc18-72ffa6735f71\") " Feb 25 07:09:25 crc kubenswrapper[4978]: I0225 07:09:25.960198 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ef26333-cc4a-4e32-bc18-72ffa6735f71-kube-api-access-hmzjp" (OuterVolumeSpecName: "kube-api-access-hmzjp") pod "4ef26333-cc4a-4e32-bc18-72ffa6735f71" (UID: "4ef26333-cc4a-4e32-bc18-72ffa6735f71"). InnerVolumeSpecName "kube-api-access-hmzjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:25 crc kubenswrapper[4978]: I0225 07:09:25.988573 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-config-data" (OuterVolumeSpecName: "config-data") pod "4ef26333-cc4a-4e32-bc18-72ffa6735f71" (UID: "4ef26333-cc4a-4e32-bc18-72ffa6735f71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:25 crc kubenswrapper[4978]: I0225 07:09:25.997775 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ef26333-cc4a-4e32-bc18-72ffa6735f71" (UID: "4ef26333-cc4a-4e32-bc18-72ffa6735f71"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.054561 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.054589 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ef26333-cc4a-4e32-bc18-72ffa6735f71-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.054600 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmzjp\" (UniqueName: \"kubernetes.io/projected/4ef26333-cc4a-4e32-bc18-72ffa6735f71-kube-api-access-hmzjp\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.312768 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa5bc4db-5446-4147-9321-23746eb83646","Type":"ContainerStarted","Data":"ed8197914042cd9ab3b1211846b5e04fb861c6311c0f0947427559fdff9d6da6"} Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.316077 4978 generic.go:334] "Generic (PLEG): container finished" podID="4ef26333-cc4a-4e32-bc18-72ffa6735f71" containerID="9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a" exitCode=137 Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.316137 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ef26333-cc4a-4e32-bc18-72ffa6735f71","Type":"ContainerDied","Data":"9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a"} Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.316182 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4ef26333-cc4a-4e32-bc18-72ffa6735f71","Type":"ContainerDied","Data":"af5a64bd80bb02a120422a57d3c91be4516d2b6a5f0e184a9daaeefe93e4f073"} Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.316206 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.316219 4978 scope.go:117] "RemoveContainer" containerID="9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.341375 4978 scope.go:117] "RemoveContainer" containerID="9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a" Feb 25 07:09:26 crc kubenswrapper[4978]: E0225 07:09:26.342103 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a\": container with ID starting with 9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a not found: ID does not exist" containerID="9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.342144 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a"} err="failed to get container status \"9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a\": rpc error: code = NotFound desc = could not find container \"9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a\": container with ID starting with 9e01793bb7593f4558198f5c27a9fd79b9e5f78ae8feff47b7660c40a4dec07a not found: ID does not exist" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.378669 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.408090 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.418940 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:09:26 crc kubenswrapper[4978]: E0225 07:09:26.419554 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ef26333-cc4a-4e32-bc18-72ffa6735f71" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.419581 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ef26333-cc4a-4e32-bc18-72ffa6735f71" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.419950 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ef26333-cc4a-4e32-bc18-72ffa6735f71" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.421151 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.423997 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.424614 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.424749 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.429311 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.562123 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.562331 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.562820 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.562951 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.562996 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfs5p\" (UniqueName: \"kubernetes.io/projected/185c1474-2c9a-408d-a3b1-8eb4820f6109-kube-api-access-zfs5p\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.666278 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.666955 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfs5p\" (UniqueName: \"kubernetes.io/projected/185c1474-2c9a-408d-a3b1-8eb4820f6109-kube-api-access-zfs5p\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.669230 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.671141 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.680152 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.680219 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.682822 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.684746 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.685723 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.696690 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfs5p\" (UniqueName: \"kubernetes.io/projected/185c1474-2c9a-408d-a3b1-8eb4820f6109-kube-api-access-zfs5p\") pod \"nova-cell1-novncproxy-0\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:26 crc kubenswrapper[4978]: I0225 07:09:26.740921 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:27 crc kubenswrapper[4978]: W0225 07:09:27.197955 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod185c1474_2c9a_408d_a3b1_8eb4820f6109.slice/crio-fddc28cae6f8e09ad937d7d26c7a1b34f83080dc03593ec85877b1767e65396d WatchSource:0}: Error finding container fddc28cae6f8e09ad937d7d26c7a1b34f83080dc03593ec85877b1767e65396d: Status 404 returned error can't find the container with id fddc28cae6f8e09ad937d7d26c7a1b34f83080dc03593ec85877b1767e65396d Feb 25 07:09:27 crc kubenswrapper[4978]: I0225 07:09:27.199321 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:09:27 crc kubenswrapper[4978]: I0225 07:09:27.341140 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ef26333-cc4a-4e32-bc18-72ffa6735f71" path="/var/lib/kubelet/pods/4ef26333-cc4a-4e32-bc18-72ffa6735f71/volumes" Feb 25 07:09:27 crc kubenswrapper[4978]: I0225 07:09:27.342060 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"185c1474-2c9a-408d-a3b1-8eb4820f6109","Type":"ContainerStarted","Data":"fddc28cae6f8e09ad937d7d26c7a1b34f83080dc03593ec85877b1767e65396d"} Feb 25 07:09:27 crc kubenswrapper[4978]: I0225 07:09:27.342088 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa5bc4db-5446-4147-9321-23746eb83646","Type":"ContainerStarted","Data":"76500bbe7caf0eced1dcb68ce3f787a778951b2789599f97b351ac6ddceb932a"} Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.238629 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.239648 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.240850 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.242568 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.348616 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"185c1474-2c9a-408d-a3b1-8eb4820f6109","Type":"ContainerStarted","Data":"70fcc3b61d5c76d88b1ebc2b70685558693748a8e9f0734e5a69556982ffa4cb"} Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.352261 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa5bc4db-5446-4147-9321-23746eb83646","Type":"ContainerStarted","Data":"0ba293ebd4f5e9425e034a3da3325f3fc7b26879385036b59a27861ad49c44ee"} Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.352624 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.362996 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.380811 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.380787214 podStartE2EDuration="2.380787214s" podCreationTimestamp="2026-02-25 07:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:28.365174713 +0000 UTC m=+1461.804431192" watchObservedRunningTime="2026-02-25 07:09:28.380787214 +0000 UTC m=+1461.820043673" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.567348 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c64474bcc-4rmct"] Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.587236 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.601950 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c64474bcc-4rmct"] Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.627463 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-sb\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.627587 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-swift-storage-0\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.627695 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-nb\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.627827 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-config\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.627940 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-svc\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.628174 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k26lx\" (UniqueName: \"kubernetes.io/projected/e78e5151-267c-4b33-9473-6b249b01f12f-kube-api-access-k26lx\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.729359 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-config\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.729444 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-svc\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.729525 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k26lx\" (UniqueName: \"kubernetes.io/projected/e78e5151-267c-4b33-9473-6b249b01f12f-kube-api-access-k26lx\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.729552 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-sb\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.729580 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-swift-storage-0\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.729605 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-nb\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.730312 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-config\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.730453 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-nb\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.730614 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-sb\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.730660 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-svc\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.730783 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-swift-storage-0\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.756072 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k26lx\" (UniqueName: \"kubernetes.io/projected/e78e5151-267c-4b33-9473-6b249b01f12f-kube-api-access-k26lx\") pod \"dnsmasq-dns-5c64474bcc-4rmct\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:28 crc kubenswrapper[4978]: I0225 07:09:28.920510 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:29 crc kubenswrapper[4978]: I0225 07:09:29.366060 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa5bc4db-5446-4147-9321-23746eb83646","Type":"ContainerStarted","Data":"25986ea6640170fa0c25c24271dcb7d998bdeec77a0e361224c83ddf87791ddc"} Feb 25 07:09:29 crc kubenswrapper[4978]: I0225 07:09:29.368079 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 07:09:29 crc kubenswrapper[4978]: I0225 07:09:29.404407 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.57761715 podStartE2EDuration="5.404388883s" podCreationTimestamp="2026-02-25 07:09:24 +0000 UTC" firstStartedPulling="2026-02-25 07:09:25.178968502 +0000 UTC m=+1458.618224971" lastFinishedPulling="2026-02-25 07:09:29.005740235 +0000 UTC m=+1462.444996704" observedRunningTime="2026-02-25 07:09:29.389545976 +0000 UTC m=+1462.828802445" watchObservedRunningTime="2026-02-25 07:09:29.404388883 +0000 UTC m=+1462.843645342" Feb 25 07:09:29 crc kubenswrapper[4978]: W0225 07:09:29.487238 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode78e5151_267c_4b33_9473_6b249b01f12f.slice/crio-ad1c1aa647da09a023e8ffbe67cb46a2adb7b22fab5ea2529869b0f4806759e0 WatchSource:0}: Error finding container ad1c1aa647da09a023e8ffbe67cb46a2adb7b22fab5ea2529869b0f4806759e0: Status 404 returned error can't find the container with id ad1c1aa647da09a023e8ffbe67cb46a2adb7b22fab5ea2529869b0f4806759e0 Feb 25 07:09:29 crc kubenswrapper[4978]: I0225 07:09:29.489924 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c64474bcc-4rmct"] Feb 25 07:09:30 crc kubenswrapper[4978]: I0225 07:09:30.376956 4978 generic.go:334] "Generic (PLEG): container finished" podID="e78e5151-267c-4b33-9473-6b249b01f12f" containerID="772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263" exitCode=0 Feb 25 07:09:30 crc kubenswrapper[4978]: I0225 07:09:30.377055 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" event={"ID":"e78e5151-267c-4b33-9473-6b249b01f12f","Type":"ContainerDied","Data":"772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263"} Feb 25 07:09:30 crc kubenswrapper[4978]: I0225 07:09:30.377422 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" event={"ID":"e78e5151-267c-4b33-9473-6b249b01f12f","Type":"ContainerStarted","Data":"ad1c1aa647da09a023e8ffbe67cb46a2adb7b22fab5ea2529869b0f4806759e0"} Feb 25 07:09:30 crc kubenswrapper[4978]: I0225 07:09:30.639204 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 25 07:09:30 crc kubenswrapper[4978]: I0225 07:09:30.814935 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:31 crc kubenswrapper[4978]: I0225 07:09:31.074467 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:31 crc kubenswrapper[4978]: I0225 07:09:31.388021 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" event={"ID":"e78e5151-267c-4b33-9473-6b249b01f12f","Type":"ContainerStarted","Data":"4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706"} Feb 25 07:09:31 crc kubenswrapper[4978]: I0225 07:09:31.388176 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerName="nova-api-log" containerID="cri-o://3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148" gracePeriod=30 Feb 25 07:09:31 crc kubenswrapper[4978]: I0225 07:09:31.388236 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerName="nova-api-api" containerID="cri-o://8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2" gracePeriod=30 Feb 25 07:09:31 crc kubenswrapper[4978]: I0225 07:09:31.388485 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="ceilometer-central-agent" containerID="cri-o://ed8197914042cd9ab3b1211846b5e04fb861c6311c0f0947427559fdff9d6da6" gracePeriod=30 Feb 25 07:09:31 crc kubenswrapper[4978]: I0225 07:09:31.388506 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="ceilometer-notification-agent" containerID="cri-o://76500bbe7caf0eced1dcb68ce3f787a778951b2789599f97b351ac6ddceb932a" gracePeriod=30 Feb 25 07:09:31 crc kubenswrapper[4978]: I0225 07:09:31.388531 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="proxy-httpd" containerID="cri-o://25986ea6640170fa0c25c24271dcb7d998bdeec77a0e361224c83ddf87791ddc" gracePeriod=30 Feb 25 07:09:31 crc kubenswrapper[4978]: I0225 07:09:31.388507 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="sg-core" containerID="cri-o://0ba293ebd4f5e9425e034a3da3325f3fc7b26879385036b59a27861ad49c44ee" gracePeriod=30 Feb 25 07:09:31 crc kubenswrapper[4978]: I0225 07:09:31.418324 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" podStartSLOduration=3.418299434 podStartE2EDuration="3.418299434s" podCreationTimestamp="2026-02-25 07:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:31.412196242 +0000 UTC m=+1464.851452701" watchObservedRunningTime="2026-02-25 07:09:31.418299434 +0000 UTC m=+1464.857555913" Feb 25 07:09:31 crc kubenswrapper[4978]: I0225 07:09:31.743428 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:32 crc kubenswrapper[4978]: I0225 07:09:32.406779 4978 generic.go:334] "Generic (PLEG): container finished" podID="aa5bc4db-5446-4147-9321-23746eb83646" containerID="25986ea6640170fa0c25c24271dcb7d998bdeec77a0e361224c83ddf87791ddc" exitCode=0 Feb 25 07:09:32 crc kubenswrapper[4978]: I0225 07:09:32.406884 4978 generic.go:334] "Generic (PLEG): container finished" podID="aa5bc4db-5446-4147-9321-23746eb83646" containerID="0ba293ebd4f5e9425e034a3da3325f3fc7b26879385036b59a27861ad49c44ee" exitCode=2 Feb 25 07:09:32 crc kubenswrapper[4978]: I0225 07:09:32.406931 4978 generic.go:334] "Generic (PLEG): container finished" podID="aa5bc4db-5446-4147-9321-23746eb83646" containerID="76500bbe7caf0eced1dcb68ce3f787a778951b2789599f97b351ac6ddceb932a" exitCode=0 Feb 25 07:09:32 crc kubenswrapper[4978]: I0225 07:09:32.406931 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa5bc4db-5446-4147-9321-23746eb83646","Type":"ContainerDied","Data":"25986ea6640170fa0c25c24271dcb7d998bdeec77a0e361224c83ddf87791ddc"} Feb 25 07:09:32 crc kubenswrapper[4978]: I0225 07:09:32.406963 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa5bc4db-5446-4147-9321-23746eb83646","Type":"ContainerDied","Data":"0ba293ebd4f5e9425e034a3da3325f3fc7b26879385036b59a27861ad49c44ee"} Feb 25 07:09:32 crc kubenswrapper[4978]: I0225 07:09:32.406972 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa5bc4db-5446-4147-9321-23746eb83646","Type":"ContainerDied","Data":"76500bbe7caf0eced1dcb68ce3f787a778951b2789599f97b351ac6ddceb932a"} Feb 25 07:09:32 crc kubenswrapper[4978]: I0225 07:09:32.409781 4978 generic.go:334] "Generic (PLEG): container finished" podID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerID="3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148" exitCode=143 Feb 25 07:09:32 crc kubenswrapper[4978]: I0225 07:09:32.411091 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b68f0d-5478-49ba-9671-0017c6edc37b","Type":"ContainerDied","Data":"3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148"} Feb 25 07:09:32 crc kubenswrapper[4978]: I0225 07:09:32.411117 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.420487 4978 generic.go:334] "Generic (PLEG): container finished" podID="aa5bc4db-5446-4147-9321-23746eb83646" containerID="ed8197914042cd9ab3b1211846b5e04fb861c6311c0f0947427559fdff9d6da6" exitCode=0 Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.421526 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa5bc4db-5446-4147-9321-23746eb83646","Type":"ContainerDied","Data":"ed8197914042cd9ab3b1211846b5e04fb861c6311c0f0947427559fdff9d6da6"} Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.722210 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.857016 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-log-httpd\") pod \"aa5bc4db-5446-4147-9321-23746eb83646\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.857146 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-run-httpd\") pod \"aa5bc4db-5446-4147-9321-23746eb83646\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.857181 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-ceilometer-tls-certs\") pod \"aa5bc4db-5446-4147-9321-23746eb83646\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.857261 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-config-data\") pod \"aa5bc4db-5446-4147-9321-23746eb83646\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.857341 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-sg-core-conf-yaml\") pod \"aa5bc4db-5446-4147-9321-23746eb83646\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.857413 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwfhh\" (UniqueName: \"kubernetes.io/projected/aa5bc4db-5446-4147-9321-23746eb83646-kube-api-access-hwfhh\") pod \"aa5bc4db-5446-4147-9321-23746eb83646\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.857463 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-scripts\") pod \"aa5bc4db-5446-4147-9321-23746eb83646\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.857487 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-combined-ca-bundle\") pod \"aa5bc4db-5446-4147-9321-23746eb83646\" (UID: \"aa5bc4db-5446-4147-9321-23746eb83646\") " Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.859491 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aa5bc4db-5446-4147-9321-23746eb83646" (UID: "aa5bc4db-5446-4147-9321-23746eb83646"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.859672 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aa5bc4db-5446-4147-9321-23746eb83646" (UID: "aa5bc4db-5446-4147-9321-23746eb83646"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.864949 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-scripts" (OuterVolumeSpecName: "scripts") pod "aa5bc4db-5446-4147-9321-23746eb83646" (UID: "aa5bc4db-5446-4147-9321-23746eb83646"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.866828 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa5bc4db-5446-4147-9321-23746eb83646-kube-api-access-hwfhh" (OuterVolumeSpecName: "kube-api-access-hwfhh") pod "aa5bc4db-5446-4147-9321-23746eb83646" (UID: "aa5bc4db-5446-4147-9321-23746eb83646"). InnerVolumeSpecName "kube-api-access-hwfhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.896724 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aa5bc4db-5446-4147-9321-23746eb83646" (UID: "aa5bc4db-5446-4147-9321-23746eb83646"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.934600 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aa5bc4db-5446-4147-9321-23746eb83646" (UID: "aa5bc4db-5446-4147-9321-23746eb83646"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.960212 4978 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.960254 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwfhh\" (UniqueName: \"kubernetes.io/projected/aa5bc4db-5446-4147-9321-23746eb83646-kube-api-access-hwfhh\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.960275 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.960290 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.960305 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa5bc4db-5446-4147-9321-23746eb83646-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.960319 4978 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:33 crc kubenswrapper[4978]: I0225 07:09:33.976552 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa5bc4db-5446-4147-9321-23746eb83646" (UID: "aa5bc4db-5446-4147-9321-23746eb83646"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.000584 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-config-data" (OuterVolumeSpecName: "config-data") pod "aa5bc4db-5446-4147-9321-23746eb83646" (UID: "aa5bc4db-5446-4147-9321-23746eb83646"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.061736 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.061893 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5bc4db-5446-4147-9321-23746eb83646-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.436418 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa5bc4db-5446-4147-9321-23746eb83646","Type":"ContainerDied","Data":"883c779c5d06fdeb909aec2baa0440c2f97a8d9ec96c923407229d46a3da9d41"} Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.436498 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.436504 4978 scope.go:117] "RemoveContainer" containerID="25986ea6640170fa0c25c24271dcb7d998bdeec77a0e361224c83ddf87791ddc" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.473175 4978 scope.go:117] "RemoveContainer" containerID="0ba293ebd4f5e9425e034a3da3325f3fc7b26879385036b59a27861ad49c44ee" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.490161 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.507477 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.511363 4978 scope.go:117] "RemoveContainer" containerID="76500bbe7caf0eced1dcb68ce3f787a778951b2789599f97b351ac6ddceb932a" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.516545 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:34 crc kubenswrapper[4978]: E0225 07:09:34.516902 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="proxy-httpd" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.516918 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="proxy-httpd" Feb 25 07:09:34 crc kubenswrapper[4978]: E0225 07:09:34.516941 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="ceilometer-central-agent" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.516947 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="ceilometer-central-agent" Feb 25 07:09:34 crc kubenswrapper[4978]: E0225 07:09:34.516957 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="sg-core" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.516964 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="sg-core" Feb 25 07:09:34 crc kubenswrapper[4978]: E0225 07:09:34.516985 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="ceilometer-notification-agent" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.516991 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="ceilometer-notification-agent" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.517146 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="ceilometer-notification-agent" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.517165 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="ceilometer-central-agent" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.517176 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="proxy-httpd" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.517191 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5bc4db-5446-4147-9321-23746eb83646" containerName="sg-core" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.520953 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.523620 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.523647 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.523804 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.543855 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.550906 4978 scope.go:117] "RemoveContainer" containerID="ed8197914042cd9ab3b1211846b5e04fb861c6311c0f0947427559fdff9d6da6" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.577203 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75l5j\" (UniqueName: \"kubernetes.io/projected/1ddb624b-1c38-44b8-8f02-bd177937ef4d-kube-api-access-75l5j\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.577249 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-run-httpd\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.577574 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-config-data\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.577618 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-scripts\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.577718 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.577939 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-log-httpd\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.577997 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.578027 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.680076 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.680201 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-log-httpd\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.680235 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.680258 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.680386 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-run-httpd\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.680415 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75l5j\" (UniqueName: \"kubernetes.io/projected/1ddb624b-1c38-44b8-8f02-bd177937ef4d-kube-api-access-75l5j\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.680474 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-config-data\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.680514 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-scripts\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.682477 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-run-httpd\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.682853 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-log-httpd\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.685420 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-scripts\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.687204 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.687718 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.688449 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.689407 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-config-data\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.701284 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75l5j\" (UniqueName: \"kubernetes.io/projected/1ddb624b-1c38-44b8-8f02-bd177937ef4d-kube-api-access-75l5j\") pod \"ceilometer-0\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " pod="openstack/ceilometer-0" Feb 25 07:09:34 crc kubenswrapper[4978]: I0225 07:09:34.837892 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.002432 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.087208 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr2tw\" (UniqueName: \"kubernetes.io/projected/50b68f0d-5478-49ba-9671-0017c6edc37b-kube-api-access-hr2tw\") pod \"50b68f0d-5478-49ba-9671-0017c6edc37b\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.087495 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-combined-ca-bundle\") pod \"50b68f0d-5478-49ba-9671-0017c6edc37b\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.088322 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-config-data\") pod \"50b68f0d-5478-49ba-9671-0017c6edc37b\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.088389 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b68f0d-5478-49ba-9671-0017c6edc37b-logs\") pod \"50b68f0d-5478-49ba-9671-0017c6edc37b\" (UID: \"50b68f0d-5478-49ba-9671-0017c6edc37b\") " Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.089809 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50b68f0d-5478-49ba-9671-0017c6edc37b-logs" (OuterVolumeSpecName: "logs") pod "50b68f0d-5478-49ba-9671-0017c6edc37b" (UID: "50b68f0d-5478-49ba-9671-0017c6edc37b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.099571 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b68f0d-5478-49ba-9671-0017c6edc37b-kube-api-access-hr2tw" (OuterVolumeSpecName: "kube-api-access-hr2tw") pod "50b68f0d-5478-49ba-9671-0017c6edc37b" (UID: "50b68f0d-5478-49ba-9671-0017c6edc37b"). InnerVolumeSpecName "kube-api-access-hr2tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.118726 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-config-data" (OuterVolumeSpecName: "config-data") pod "50b68f0d-5478-49ba-9671-0017c6edc37b" (UID: "50b68f0d-5478-49ba-9671-0017c6edc37b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.120597 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50b68f0d-5478-49ba-9671-0017c6edc37b" (UID: "50b68f0d-5478-49ba-9671-0017c6edc37b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.190934 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.190967 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b68f0d-5478-49ba-9671-0017c6edc37b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.190980 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b68f0d-5478-49ba-9671-0017c6edc37b-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.190995 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr2tw\" (UniqueName: \"kubernetes.io/projected/50b68f0d-5478-49ba-9671-0017c6edc37b-kube-api-access-hr2tw\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.340969 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa5bc4db-5446-4147-9321-23746eb83646" path="/var/lib/kubelet/pods/aa5bc4db-5446-4147-9321-23746eb83646/volumes" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.448691 4978 generic.go:334] "Generic (PLEG): container finished" podID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerID="8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2" exitCode=0 Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.448733 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.448738 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b68f0d-5478-49ba-9671-0017c6edc37b","Type":"ContainerDied","Data":"8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2"} Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.448863 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b68f0d-5478-49ba-9671-0017c6edc37b","Type":"ContainerDied","Data":"37bccae3093bfd6dc38a73bf86adba09b88c393016dd44405c2549bb3eac11b5"} Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.448891 4978 scope.go:117] "RemoveContainer" containerID="8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.481954 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.492221 4978 scope.go:117] "RemoveContainer" containerID="3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.504576 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.523122 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:35 crc kubenswrapper[4978]: E0225 07:09:35.523660 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerName="nova-api-api" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.523679 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerName="nova-api-api" Feb 25 07:09:35 crc kubenswrapper[4978]: E0225 07:09:35.523705 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerName="nova-api-log" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.523712 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerName="nova-api-log" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.523939 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerName="nova-api-log" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.523976 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" containerName="nova-api-api" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.525438 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.528439 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.528495 4978 scope.go:117] "RemoveContainer" containerID="8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.528614 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.528693 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 07:09:35 crc kubenswrapper[4978]: E0225 07:09:35.529568 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2\": container with ID starting with 8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2 not found: ID does not exist" containerID="8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.529593 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2"} err="failed to get container status \"8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2\": rpc error: code = NotFound desc = could not find container \"8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2\": container with ID starting with 8a2a4b9076248aa175906f0c9c9c248d51b73213151659d50d99cf3d72fc7fd2 not found: ID does not exist" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.529619 4978 scope.go:117] "RemoveContainer" containerID="3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148" Feb 25 07:09:35 crc kubenswrapper[4978]: E0225 07:09:35.534501 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148\": container with ID starting with 3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148 not found: ID does not exist" containerID="3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.534530 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148"} err="failed to get container status \"3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148\": rpc error: code = NotFound desc = could not find container \"3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148\": container with ID starting with 3b7b91f82191fd28cb58626422ecd9787785e9dd001037be04c8356e26205148 not found: ID does not exist" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.535609 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.602463 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.602797 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-public-tls-certs\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.602858 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae57db5c-1b89-40c2-935c-696dec4d4944-logs\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.602898 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.603188 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-config-data\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.603212 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wf8k\" (UniqueName: \"kubernetes.io/projected/ae57db5c-1b89-40c2-935c-696dec4d4944-kube-api-access-9wf8k\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.704653 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-config-data\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.704699 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wf8k\" (UniqueName: \"kubernetes.io/projected/ae57db5c-1b89-40c2-935c-696dec4d4944-kube-api-access-9wf8k\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.704755 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.704782 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-public-tls-certs\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.704805 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae57db5c-1b89-40c2-935c-696dec4d4944-logs\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.704823 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.705738 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae57db5c-1b89-40c2-935c-696dec4d4944-logs\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.708259 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-config-data\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.713437 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.721233 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.721385 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-public-tls-certs\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.722298 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wf8k\" (UniqueName: \"kubernetes.io/projected/ae57db5c-1b89-40c2-935c-696dec4d4944-kube-api-access-9wf8k\") pod \"nova-api-0\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " pod="openstack/nova-api-0" Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.760813 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.760884 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 07:09:35 crc kubenswrapper[4978]: I0225 07:09:35.842414 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:36 crc kubenswrapper[4978]: I0225 07:09:36.311873 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:36 crc kubenswrapper[4978]: W0225 07:09:36.319207 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae57db5c_1b89_40c2_935c_696dec4d4944.slice/crio-1c02d72f91e314b04473bd8594d0a4b91f5a86f3817af5c85b0cb165285ab737 WatchSource:0}: Error finding container 1c02d72f91e314b04473bd8594d0a4b91f5a86f3817af5c85b0cb165285ab737: Status 404 returned error can't find the container with id 1c02d72f91e314b04473bd8594d0a4b91f5a86f3817af5c85b0cb165285ab737 Feb 25 07:09:36 crc kubenswrapper[4978]: I0225 07:09:36.458601 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ddb624b-1c38-44b8-8f02-bd177937ef4d","Type":"ContainerStarted","Data":"a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a"} Feb 25 07:09:36 crc kubenswrapper[4978]: I0225 07:09:36.458645 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ddb624b-1c38-44b8-8f02-bd177937ef4d","Type":"ContainerStarted","Data":"48b11d3570e4f26f90f450bca7435c16a6f6a7b6b849ce2e5353a33ace7745e9"} Feb 25 07:09:36 crc kubenswrapper[4978]: I0225 07:09:36.463541 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae57db5c-1b89-40c2-935c-696dec4d4944","Type":"ContainerStarted","Data":"1c02d72f91e314b04473bd8594d0a4b91f5a86f3817af5c85b0cb165285ab737"} Feb 25 07:09:36 crc kubenswrapper[4978]: I0225 07:09:36.741127 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:36 crc kubenswrapper[4978]: I0225 07:09:36.777274 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.344569 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50b68f0d-5478-49ba-9671-0017c6edc37b" path="/var/lib/kubelet/pods/50b68f0d-5478-49ba-9671-0017c6edc37b/volumes" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.476625 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ddb624b-1c38-44b8-8f02-bd177937ef4d","Type":"ContainerStarted","Data":"48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae"} Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.479513 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae57db5c-1b89-40c2-935c-696dec4d4944","Type":"ContainerStarted","Data":"7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3"} Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.479585 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae57db5c-1b89-40c2-935c-696dec4d4944","Type":"ContainerStarted","Data":"d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4"} Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.511929 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.518505 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.518483447 podStartE2EDuration="2.518483447s" podCreationTimestamp="2026-02-25 07:09:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:37.511088264 +0000 UTC m=+1470.950344743" watchObservedRunningTime="2026-02-25 07:09:37.518483447 +0000 UTC m=+1470.957739906" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.651346 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-2dsvp"] Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.653404 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.660949 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.661488 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.662295 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2dsvp"] Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.743538 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-config-data\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.743602 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.743805 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-scripts\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.743834 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzfhg\" (UniqueName: \"kubernetes.io/projected/e0764b55-360c-46e0-a702-28ddb93b475d-kube-api-access-pzfhg\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.846008 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-scripts\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.846054 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzfhg\" (UniqueName: \"kubernetes.io/projected/e0764b55-360c-46e0-a702-28ddb93b475d-kube-api-access-pzfhg\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.846110 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-config-data\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.846125 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.851084 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-config-data\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.851111 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-scripts\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.852918 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.871799 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzfhg\" (UniqueName: \"kubernetes.io/projected/e0764b55-360c-46e0-a702-28ddb93b475d-kube-api-access-pzfhg\") pod \"nova-cell1-cell-mapping-2dsvp\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:37 crc kubenswrapper[4978]: I0225 07:09:37.988764 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:38 crc kubenswrapper[4978]: I0225 07:09:38.513241 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ddb624b-1c38-44b8-8f02-bd177937ef4d","Type":"ContainerStarted","Data":"887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d"} Feb 25 07:09:38 crc kubenswrapper[4978]: I0225 07:09:38.532292 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2dsvp"] Feb 25 07:09:38 crc kubenswrapper[4978]: I0225 07:09:38.922662 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.043155 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-689695c99c-nmcjm"] Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.043394 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" podUID="617c5921-2132-4867-b0ef-5f10a7c1c938" containerName="dnsmasq-dns" containerID="cri-o://e227b0a33aecace45163696f40de57e5900930437a82fb19dd2b00654d1ceff6" gracePeriod=10 Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.525086 4978 generic.go:334] "Generic (PLEG): container finished" podID="617c5921-2132-4867-b0ef-5f10a7c1c938" containerID="e227b0a33aecace45163696f40de57e5900930437a82fb19dd2b00654d1ceff6" exitCode=0 Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.525153 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" event={"ID":"617c5921-2132-4867-b0ef-5f10a7c1c938","Type":"ContainerDied","Data":"e227b0a33aecace45163696f40de57e5900930437a82fb19dd2b00654d1ceff6"} Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.525481 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" event={"ID":"617c5921-2132-4867-b0ef-5f10a7c1c938","Type":"ContainerDied","Data":"f8e2b1a2a9d0f0c7eb7b2873583675d29458e5d253699766e9daa9b96e3c0855"} Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.525508 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8e2b1a2a9d0f0c7eb7b2873583675d29458e5d253699766e9daa9b96e3c0855" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.529865 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ddb624b-1c38-44b8-8f02-bd177937ef4d","Type":"ContainerStarted","Data":"f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b"} Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.530027 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.533626 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2dsvp" event={"ID":"e0764b55-360c-46e0-a702-28ddb93b475d","Type":"ContainerStarted","Data":"c61141b61f21cb127aa4ec4c820436879baedf614cff727d7cb229d414bd57ad"} Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.533674 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2dsvp" event={"ID":"e0764b55-360c-46e0-a702-28ddb93b475d","Type":"ContainerStarted","Data":"6cd50fee6d962184a40fe1e767072e2d53c68d215967391ecaa1977c8e33fa1d"} Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.553044 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.560839 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.07466171 podStartE2EDuration="5.560815482s" podCreationTimestamp="2026-02-25 07:09:34 +0000 UTC" firstStartedPulling="2026-02-25 07:09:35.760673297 +0000 UTC m=+1469.199929756" lastFinishedPulling="2026-02-25 07:09:39.246827069 +0000 UTC m=+1472.686083528" observedRunningTime="2026-02-25 07:09:39.55184449 +0000 UTC m=+1472.991100959" watchObservedRunningTime="2026-02-25 07:09:39.560815482 +0000 UTC m=+1473.000071931" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.575123 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-2dsvp" podStartSLOduration=2.575101822 podStartE2EDuration="2.575101822s" podCreationTimestamp="2026-02-25 07:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:39.566590634 +0000 UTC m=+1473.005847093" watchObservedRunningTime="2026-02-25 07:09:39.575101822 +0000 UTC m=+1473.014358281" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.581472 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-swift-storage-0\") pod \"617c5921-2132-4867-b0ef-5f10a7c1c938\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.581572 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-sb\") pod \"617c5921-2132-4867-b0ef-5f10a7c1c938\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.581664 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-nb\") pod \"617c5921-2132-4867-b0ef-5f10a7c1c938\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.581900 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-svc\") pod \"617c5921-2132-4867-b0ef-5f10a7c1c938\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.581974 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-config\") pod \"617c5921-2132-4867-b0ef-5f10a7c1c938\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.582035 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgbc9\" (UniqueName: \"kubernetes.io/projected/617c5921-2132-4867-b0ef-5f10a7c1c938-kube-api-access-dgbc9\") pod \"617c5921-2132-4867-b0ef-5f10a7c1c938\" (UID: \"617c5921-2132-4867-b0ef-5f10a7c1c938\") " Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.599570 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/617c5921-2132-4867-b0ef-5f10a7c1c938-kube-api-access-dgbc9" (OuterVolumeSpecName: "kube-api-access-dgbc9") pod "617c5921-2132-4867-b0ef-5f10a7c1c938" (UID: "617c5921-2132-4867-b0ef-5f10a7c1c938"). InnerVolumeSpecName "kube-api-access-dgbc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.635071 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "617c5921-2132-4867-b0ef-5f10a7c1c938" (UID: "617c5921-2132-4867-b0ef-5f10a7c1c938"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.639125 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "617c5921-2132-4867-b0ef-5f10a7c1c938" (UID: "617c5921-2132-4867-b0ef-5f10a7c1c938"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.666641 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "617c5921-2132-4867-b0ef-5f10a7c1c938" (UID: "617c5921-2132-4867-b0ef-5f10a7c1c938"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.666644 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "617c5921-2132-4867-b0ef-5f10a7c1c938" (UID: "617c5921-2132-4867-b0ef-5f10a7c1c938"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.671319 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-config" (OuterVolumeSpecName: "config") pod "617c5921-2132-4867-b0ef-5f10a7c1c938" (UID: "617c5921-2132-4867-b0ef-5f10a7c1c938"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.684520 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.684542 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.684551 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgbc9\" (UniqueName: \"kubernetes.io/projected/617c5921-2132-4867-b0ef-5f10a7c1c938-kube-api-access-dgbc9\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.684563 4978 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.684572 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:39 crc kubenswrapper[4978]: I0225 07:09:39.684580 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/617c5921-2132-4867-b0ef-5f10a7c1c938-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:40 crc kubenswrapper[4978]: I0225 07:09:40.543633 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-689695c99c-nmcjm" Feb 25 07:09:40 crc kubenswrapper[4978]: I0225 07:09:40.591595 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-689695c99c-nmcjm"] Feb 25 07:09:40 crc kubenswrapper[4978]: I0225 07:09:40.602856 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-689695c99c-nmcjm"] Feb 25 07:09:41 crc kubenswrapper[4978]: I0225 07:09:41.353735 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="617c5921-2132-4867-b0ef-5f10a7c1c938" path="/var/lib/kubelet/pods/617c5921-2132-4867-b0ef-5f10a7c1c938/volumes" Feb 25 07:09:43 crc kubenswrapper[4978]: I0225 07:09:43.583278 4978 generic.go:334] "Generic (PLEG): container finished" podID="e0764b55-360c-46e0-a702-28ddb93b475d" containerID="c61141b61f21cb127aa4ec4c820436879baedf614cff727d7cb229d414bd57ad" exitCode=0 Feb 25 07:09:43 crc kubenswrapper[4978]: I0225 07:09:43.583393 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2dsvp" event={"ID":"e0764b55-360c-46e0-a702-28ddb93b475d","Type":"ContainerDied","Data":"c61141b61f21cb127aa4ec4c820436879baedf614cff727d7cb229d414bd57ad"} Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.081621 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.096606 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzfhg\" (UniqueName: \"kubernetes.io/projected/e0764b55-360c-46e0-a702-28ddb93b475d-kube-api-access-pzfhg\") pod \"e0764b55-360c-46e0-a702-28ddb93b475d\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.097242 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-scripts\") pod \"e0764b55-360c-46e0-a702-28ddb93b475d\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.097332 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-config-data\") pod \"e0764b55-360c-46e0-a702-28ddb93b475d\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.097518 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-combined-ca-bundle\") pod \"e0764b55-360c-46e0-a702-28ddb93b475d\" (UID: \"e0764b55-360c-46e0-a702-28ddb93b475d\") " Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.102410 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0764b55-360c-46e0-a702-28ddb93b475d-kube-api-access-pzfhg" (OuterVolumeSpecName: "kube-api-access-pzfhg") pod "e0764b55-360c-46e0-a702-28ddb93b475d" (UID: "e0764b55-360c-46e0-a702-28ddb93b475d"). InnerVolumeSpecName "kube-api-access-pzfhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.102633 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-scripts" (OuterVolumeSpecName: "scripts") pod "e0764b55-360c-46e0-a702-28ddb93b475d" (UID: "e0764b55-360c-46e0-a702-28ddb93b475d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.125214 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-config-data" (OuterVolumeSpecName: "config-data") pod "e0764b55-360c-46e0-a702-28ddb93b475d" (UID: "e0764b55-360c-46e0-a702-28ddb93b475d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.133708 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0764b55-360c-46e0-a702-28ddb93b475d" (UID: "e0764b55-360c-46e0-a702-28ddb93b475d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.199812 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.199845 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzfhg\" (UniqueName: \"kubernetes.io/projected/e0764b55-360c-46e0-a702-28ddb93b475d-kube-api-access-pzfhg\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.199860 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.199873 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0764b55-360c-46e0-a702-28ddb93b475d-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.626583 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2dsvp" event={"ID":"e0764b55-360c-46e0-a702-28ddb93b475d","Type":"ContainerDied","Data":"6cd50fee6d962184a40fe1e767072e2d53c68d215967391ecaa1977c8e33fa1d"} Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.626629 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cd50fee6d962184a40fe1e767072e2d53c68d215967391ecaa1977c8e33fa1d" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.626801 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2dsvp" Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.826224 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.826635 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ae57db5c-1b89-40c2-935c-696dec4d4944" containerName="nova-api-log" containerID="cri-o://d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4" gracePeriod=30 Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.826748 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ae57db5c-1b89-40c2-935c-696dec4d4944" containerName="nova-api-api" containerID="cri-o://7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3" gracePeriod=30 Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.850150 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.850544 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ef038d4c-269b-407d-95db-43e90be12da7" containerName="nova-scheduler-scheduler" containerID="cri-o://53f477948bc0c813efe4de0ab156f490c6b1adba3139158be8a034b57455194b" gracePeriod=30 Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.869313 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.869758 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-metadata" containerID="cri-o://ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f" gracePeriod=30 Feb 25 07:09:45 crc kubenswrapper[4978]: I0225 07:09:45.869623 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-log" containerID="cri-o://2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b" gracePeriod=30 Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.378180 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.420071 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae57db5c-1b89-40c2-935c-696dec4d4944-logs\") pod \"ae57db5c-1b89-40c2-935c-696dec4d4944\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.420132 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-config-data\") pod \"ae57db5c-1b89-40c2-935c-696dec4d4944\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.420221 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-internal-tls-certs\") pod \"ae57db5c-1b89-40c2-935c-696dec4d4944\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.420256 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-combined-ca-bundle\") pod \"ae57db5c-1b89-40c2-935c-696dec4d4944\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.420336 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-public-tls-certs\") pod \"ae57db5c-1b89-40c2-935c-696dec4d4944\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.420388 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wf8k\" (UniqueName: \"kubernetes.io/projected/ae57db5c-1b89-40c2-935c-696dec4d4944-kube-api-access-9wf8k\") pod \"ae57db5c-1b89-40c2-935c-696dec4d4944\" (UID: \"ae57db5c-1b89-40c2-935c-696dec4d4944\") " Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.420721 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae57db5c-1b89-40c2-935c-696dec4d4944-logs" (OuterVolumeSpecName: "logs") pod "ae57db5c-1b89-40c2-935c-696dec4d4944" (UID: "ae57db5c-1b89-40c2-935c-696dec4d4944"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.421111 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae57db5c-1b89-40c2-935c-696dec4d4944-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.425071 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae57db5c-1b89-40c2-935c-696dec4d4944-kube-api-access-9wf8k" (OuterVolumeSpecName: "kube-api-access-9wf8k") pod "ae57db5c-1b89-40c2-935c-696dec4d4944" (UID: "ae57db5c-1b89-40c2-935c-696dec4d4944"). InnerVolumeSpecName "kube-api-access-9wf8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.459761 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae57db5c-1b89-40c2-935c-696dec4d4944" (UID: "ae57db5c-1b89-40c2-935c-696dec4d4944"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.470780 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-config-data" (OuterVolumeSpecName: "config-data") pod "ae57db5c-1b89-40c2-935c-696dec4d4944" (UID: "ae57db5c-1b89-40c2-935c-696dec4d4944"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.476397 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ae57db5c-1b89-40c2-935c-696dec4d4944" (UID: "ae57db5c-1b89-40c2-935c-696dec4d4944"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.480841 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ae57db5c-1b89-40c2-935c-696dec4d4944" (UID: "ae57db5c-1b89-40c2-935c-696dec4d4944"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.522705 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.522742 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.522755 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wf8k\" (UniqueName: \"kubernetes.io/projected/ae57db5c-1b89-40c2-935c-696dec4d4944-kube-api-access-9wf8k\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.522766 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.522775 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae57db5c-1b89-40c2-935c-696dec4d4944-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.639101 4978 generic.go:334] "Generic (PLEG): container finished" podID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerID="2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b" exitCode=143 Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.639169 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b8876c6c-57b4-421d-bc34-a70080f86c63","Type":"ContainerDied","Data":"2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b"} Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.647007 4978 generic.go:334] "Generic (PLEG): container finished" podID="ef038d4c-269b-407d-95db-43e90be12da7" containerID="53f477948bc0c813efe4de0ab156f490c6b1adba3139158be8a034b57455194b" exitCode=0 Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.647089 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef038d4c-269b-407d-95db-43e90be12da7","Type":"ContainerDied","Data":"53f477948bc0c813efe4de0ab156f490c6b1adba3139158be8a034b57455194b"} Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.649125 4978 generic.go:334] "Generic (PLEG): container finished" podID="ae57db5c-1b89-40c2-935c-696dec4d4944" containerID="7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3" exitCode=0 Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.649164 4978 generic.go:334] "Generic (PLEG): container finished" podID="ae57db5c-1b89-40c2-935c-696dec4d4944" containerID="d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4" exitCode=143 Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.649191 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae57db5c-1b89-40c2-935c-696dec4d4944","Type":"ContainerDied","Data":"7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3"} Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.649230 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae57db5c-1b89-40c2-935c-696dec4d4944","Type":"ContainerDied","Data":"d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4"} Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.649244 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ae57db5c-1b89-40c2-935c-696dec4d4944","Type":"ContainerDied","Data":"1c02d72f91e314b04473bd8594d0a4b91f5a86f3817af5c85b0cb165285ab737"} Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.649236 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.649306 4978 scope.go:117] "RemoveContainer" containerID="7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.700814 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.701634 4978 scope.go:117] "RemoveContainer" containerID="d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.719558 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.739128 4978 scope.go:117] "RemoveContainer" containerID="7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.740578 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:46 crc kubenswrapper[4978]: E0225 07:09:46.740599 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3\": container with ID starting with 7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3 not found: ID does not exist" containerID="7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.740651 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3"} err="failed to get container status \"7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3\": rpc error: code = NotFound desc = could not find container \"7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3\": container with ID starting with 7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3 not found: ID does not exist" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.740687 4978 scope.go:117] "RemoveContainer" containerID="d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4" Feb 25 07:09:46 crc kubenswrapper[4978]: E0225 07:09:46.741004 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="617c5921-2132-4867-b0ef-5f10a7c1c938" containerName="dnsmasq-dns" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.741019 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="617c5921-2132-4867-b0ef-5f10a7c1c938" containerName="dnsmasq-dns" Feb 25 07:09:46 crc kubenswrapper[4978]: E0225 07:09:46.741042 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae57db5c-1b89-40c2-935c-696dec4d4944" containerName="nova-api-api" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.741049 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae57db5c-1b89-40c2-935c-696dec4d4944" containerName="nova-api-api" Feb 25 07:09:46 crc kubenswrapper[4978]: E0225 07:09:46.741066 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0764b55-360c-46e0-a702-28ddb93b475d" containerName="nova-manage" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.741072 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0764b55-360c-46e0-a702-28ddb93b475d" containerName="nova-manage" Feb 25 07:09:46 crc kubenswrapper[4978]: E0225 07:09:46.741086 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae57db5c-1b89-40c2-935c-696dec4d4944" containerName="nova-api-log" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.741092 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae57db5c-1b89-40c2-935c-696dec4d4944" containerName="nova-api-log" Feb 25 07:09:46 crc kubenswrapper[4978]: E0225 07:09:46.741103 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="617c5921-2132-4867-b0ef-5f10a7c1c938" containerName="init" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.741109 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="617c5921-2132-4867-b0ef-5f10a7c1c938" containerName="init" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.741295 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae57db5c-1b89-40c2-935c-696dec4d4944" containerName="nova-api-log" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.741315 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae57db5c-1b89-40c2-935c-696dec4d4944" containerName="nova-api-api" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.741332 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="617c5921-2132-4867-b0ef-5f10a7c1c938" containerName="dnsmasq-dns" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.741341 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0764b55-360c-46e0-a702-28ddb93b475d" containerName="nova-manage" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.743958 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: E0225 07:09:46.747688 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4\": container with ID starting with d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4 not found: ID does not exist" containerID="d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.747731 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4"} err="failed to get container status \"d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4\": rpc error: code = NotFound desc = could not find container \"d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4\": container with ID starting with d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4 not found: ID does not exist" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.747757 4978 scope.go:117] "RemoveContainer" containerID="7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.748139 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.748314 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.748507 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.750880 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3"} err="failed to get container status \"7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3\": rpc error: code = NotFound desc = could not find container \"7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3\": container with ID starting with 7dd7dd2c416f9d90319048efca10549fd39fe75f428b608b71c1c4b5a933a2b3 not found: ID does not exist" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.750911 4978 scope.go:117] "RemoveContainer" containerID="d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.760153 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4"} err="failed to get container status \"d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4\": rpc error: code = NotFound desc = could not find container \"d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4\": container with ID starting with d2e6f5ef11f1d7630704f7e12ccca50f6b8b2a69d3d2f9a1dd0f83d19c7ebcb4 not found: ID does not exist" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.770654 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.828618 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e39fd28-0520-44bc-9486-a118cec688c0-logs\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.828827 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9fp8\" (UniqueName: \"kubernetes.io/projected/2e39fd28-0520-44bc-9486-a118cec688c0-kube-api-access-s9fp8\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.828923 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.829024 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.829099 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-config-data\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.829170 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-public-tls-certs\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.929792 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-config-data\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.929843 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-public-tls-certs\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.929910 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e39fd28-0520-44bc-9486-a118cec688c0-logs\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.929940 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9fp8\" (UniqueName: \"kubernetes.io/projected/2e39fd28-0520-44bc-9486-a118cec688c0-kube-api-access-s9fp8\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.929979 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.930025 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.933922 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e39fd28-0520-44bc-9486-a118cec688c0-logs\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.934849 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-public-tls-certs\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.935578 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.937643 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-config-data\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.938584 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.953053 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9fp8\" (UniqueName: \"kubernetes.io/projected/2e39fd28-0520-44bc-9486-a118cec688c0-kube-api-access-s9fp8\") pod \"nova-api-0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " pod="openstack/nova-api-0" Feb 25 07:09:46 crc kubenswrapper[4978]: I0225 07:09:46.969221 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.076529 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.133330 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-combined-ca-bundle\") pod \"ef038d4c-269b-407d-95db-43e90be12da7\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.133451 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqgdb\" (UniqueName: \"kubernetes.io/projected/ef038d4c-269b-407d-95db-43e90be12da7-kube-api-access-xqgdb\") pod \"ef038d4c-269b-407d-95db-43e90be12da7\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.133568 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-config-data\") pod \"ef038d4c-269b-407d-95db-43e90be12da7\" (UID: \"ef038d4c-269b-407d-95db-43e90be12da7\") " Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.139851 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef038d4c-269b-407d-95db-43e90be12da7-kube-api-access-xqgdb" (OuterVolumeSpecName: "kube-api-access-xqgdb") pod "ef038d4c-269b-407d-95db-43e90be12da7" (UID: "ef038d4c-269b-407d-95db-43e90be12da7"). InnerVolumeSpecName "kube-api-access-xqgdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.167848 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef038d4c-269b-407d-95db-43e90be12da7" (UID: "ef038d4c-269b-407d-95db-43e90be12da7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.181336 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-config-data" (OuterVolumeSpecName: "config-data") pod "ef038d4c-269b-407d-95db-43e90be12da7" (UID: "ef038d4c-269b-407d-95db-43e90be12da7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.235806 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.235839 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqgdb\" (UniqueName: \"kubernetes.io/projected/ef038d4c-269b-407d-95db-43e90be12da7-kube-api-access-xqgdb\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.235850 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef038d4c-269b-407d-95db-43e90be12da7-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.344491 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae57db5c-1b89-40c2-935c-696dec4d4944" path="/var/lib/kubelet/pods/ae57db5c-1b89-40c2-935c-696dec4d4944/volumes" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.574087 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:09:47 crc kubenswrapper[4978]: W0225 07:09:47.588041 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e39fd28_0520_44bc_9486_a118cec688c0.slice/crio-580286c6c360cd5f8d7a3305e3d16a1600058344e7c9c9a4bcfa71811e035a60 WatchSource:0}: Error finding container 580286c6c360cd5f8d7a3305e3d16a1600058344e7c9c9a4bcfa71811e035a60: Status 404 returned error can't find the container with id 580286c6c360cd5f8d7a3305e3d16a1600058344e7c9c9a4bcfa71811e035a60 Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.664391 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.664400 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef038d4c-269b-407d-95db-43e90be12da7","Type":"ContainerDied","Data":"660fca13a8a30130884e543bdc7c942ab8f2443e00b844c995cb12021ced081c"} Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.664470 4978 scope.go:117] "RemoveContainer" containerID="53f477948bc0c813efe4de0ab156f490c6b1adba3139158be8a034b57455194b" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.676978 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e39fd28-0520-44bc-9486-a118cec688c0","Type":"ContainerStarted","Data":"580286c6c360cd5f8d7a3305e3d16a1600058344e7c9c9a4bcfa71811e035a60"} Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.698228 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.725942 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.738138 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:47 crc kubenswrapper[4978]: E0225 07:09:47.738896 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef038d4c-269b-407d-95db-43e90be12da7" containerName="nova-scheduler-scheduler" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.738932 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef038d4c-269b-407d-95db-43e90be12da7" containerName="nova-scheduler-scheduler" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.739302 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef038d4c-269b-407d-95db-43e90be12da7" containerName="nova-scheduler-scheduler" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.740482 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.743932 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.746880 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.746974 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-config-data\") pod \"nova-scheduler-0\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.747151 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvbkl\" (UniqueName: \"kubernetes.io/projected/7456f008-69db-4a20-86c4-2ab52eb93aa7-kube-api-access-lvbkl\") pod \"nova-scheduler-0\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.748513 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.848719 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.848777 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-config-data\") pod \"nova-scheduler-0\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.848851 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvbkl\" (UniqueName: \"kubernetes.io/projected/7456f008-69db-4a20-86c4-2ab52eb93aa7-kube-api-access-lvbkl\") pod \"nova-scheduler-0\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.852280 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-config-data\") pod \"nova-scheduler-0\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.855258 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:47 crc kubenswrapper[4978]: I0225 07:09:47.865710 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvbkl\" (UniqueName: \"kubernetes.io/projected/7456f008-69db-4a20-86c4-2ab52eb93aa7-kube-api-access-lvbkl\") pod \"nova-scheduler-0\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " pod="openstack/nova-scheduler-0" Feb 25 07:09:48 crc kubenswrapper[4978]: I0225 07:09:48.091654 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:09:48 crc kubenswrapper[4978]: I0225 07:09:48.565142 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:09:48 crc kubenswrapper[4978]: W0225 07:09:48.566452 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7456f008_69db_4a20_86c4_2ab52eb93aa7.slice/crio-21b8107a73d909dcb7058b275e04afcd4a8f5bcbbdbed6e5e7901b5a9d1bdfc9 WatchSource:0}: Error finding container 21b8107a73d909dcb7058b275e04afcd4a8f5bcbbdbed6e5e7901b5a9d1bdfc9: Status 404 returned error can't find the container with id 21b8107a73d909dcb7058b275e04afcd4a8f5bcbbdbed6e5e7901b5a9d1bdfc9 Feb 25 07:09:48 crc kubenswrapper[4978]: I0225 07:09:48.708177 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e39fd28-0520-44bc-9486-a118cec688c0","Type":"ContainerStarted","Data":"ed748fe64533cd208cfaa954b8175f2c179485a3b36021fb250466d83a45c91e"} Feb 25 07:09:48 crc kubenswrapper[4978]: I0225 07:09:48.708229 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e39fd28-0520-44bc-9486-a118cec688c0","Type":"ContainerStarted","Data":"d48b870ae7f0075341f3f007177eb69704874efaa7faac91545002adf5db1de3"} Feb 25 07:09:48 crc kubenswrapper[4978]: I0225 07:09:48.712642 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7456f008-69db-4a20-86c4-2ab52eb93aa7","Type":"ContainerStarted","Data":"21b8107a73d909dcb7058b275e04afcd4a8f5bcbbdbed6e5e7901b5a9d1bdfc9"} Feb 25 07:09:48 crc kubenswrapper[4978]: I0225 07:09:48.743056 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.742936484 podStartE2EDuration="2.742936484s" podCreationTimestamp="2026-02-25 07:09:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:48.733087604 +0000 UTC m=+1482.172344063" watchObservedRunningTime="2026-02-25 07:09:48.742936484 +0000 UTC m=+1482.182192943" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.016553 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:57540->10.217.0.199:8775: read: connection reset by peer" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.016659 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.199:8775/\": read tcp 10.217.0.2:57542->10.217.0.199:8775: read: connection reset by peer" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.344158 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef038d4c-269b-407d-95db-43e90be12da7" path="/var/lib/kubelet/pods/ef038d4c-269b-407d-95db-43e90be12da7/volumes" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.465568 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.594751 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-combined-ca-bundle\") pod \"b8876c6c-57b4-421d-bc34-a70080f86c63\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.595132 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-nova-metadata-tls-certs\") pod \"b8876c6c-57b4-421d-bc34-a70080f86c63\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.595226 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trlbn\" (UniqueName: \"kubernetes.io/projected/b8876c6c-57b4-421d-bc34-a70080f86c63-kube-api-access-trlbn\") pod \"b8876c6c-57b4-421d-bc34-a70080f86c63\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.595264 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8876c6c-57b4-421d-bc34-a70080f86c63-logs\") pod \"b8876c6c-57b4-421d-bc34-a70080f86c63\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.595343 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-config-data\") pod \"b8876c6c-57b4-421d-bc34-a70080f86c63\" (UID: \"b8876c6c-57b4-421d-bc34-a70080f86c63\") " Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.595728 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8876c6c-57b4-421d-bc34-a70080f86c63-logs" (OuterVolumeSpecName: "logs") pod "b8876c6c-57b4-421d-bc34-a70080f86c63" (UID: "b8876c6c-57b4-421d-bc34-a70080f86c63"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.595850 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8876c6c-57b4-421d-bc34-a70080f86c63-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.605706 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8876c6c-57b4-421d-bc34-a70080f86c63-kube-api-access-trlbn" (OuterVolumeSpecName: "kube-api-access-trlbn") pod "b8876c6c-57b4-421d-bc34-a70080f86c63" (UID: "b8876c6c-57b4-421d-bc34-a70080f86c63"). InnerVolumeSpecName "kube-api-access-trlbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.624819 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-config-data" (OuterVolumeSpecName: "config-data") pod "b8876c6c-57b4-421d-bc34-a70080f86c63" (UID: "b8876c6c-57b4-421d-bc34-a70080f86c63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.634991 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8876c6c-57b4-421d-bc34-a70080f86c63" (UID: "b8876c6c-57b4-421d-bc34-a70080f86c63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.671299 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b8876c6c-57b4-421d-bc34-a70080f86c63" (UID: "b8876c6c-57b4-421d-bc34-a70080f86c63"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.697911 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.697972 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.697983 4978 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8876c6c-57b4-421d-bc34-a70080f86c63-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.697993 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trlbn\" (UniqueName: \"kubernetes.io/projected/b8876c6c-57b4-421d-bc34-a70080f86c63-kube-api-access-trlbn\") on node \"crc\" DevicePath \"\"" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.737391 4978 generic.go:334] "Generic (PLEG): container finished" podID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerID="ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f" exitCode=0 Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.737490 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.737591 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b8876c6c-57b4-421d-bc34-a70080f86c63","Type":"ContainerDied","Data":"ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f"} Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.737636 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b8876c6c-57b4-421d-bc34-a70080f86c63","Type":"ContainerDied","Data":"a03a801508afc147f9267b051b4db4877175d115e1bd7121f944f0859392e0f7"} Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.737662 4978 scope.go:117] "RemoveContainer" containerID="ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.741300 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7456f008-69db-4a20-86c4-2ab52eb93aa7","Type":"ContainerStarted","Data":"b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3"} Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.773667 4978 scope.go:117] "RemoveContainer" containerID="2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.774079 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.77406031 podStartE2EDuration="2.77406031s" podCreationTimestamp="2026-02-25 07:09:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:49.764114027 +0000 UTC m=+1483.203370496" watchObservedRunningTime="2026-02-25 07:09:49.77406031 +0000 UTC m=+1483.213316769" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.790434 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.817675 4978 scope.go:117] "RemoveContainer" containerID="ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f" Feb 25 07:09:49 crc kubenswrapper[4978]: E0225 07:09:49.819612 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f\": container with ID starting with ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f not found: ID does not exist" containerID="ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.819650 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f"} err="failed to get container status \"ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f\": rpc error: code = NotFound desc = could not find container \"ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f\": container with ID starting with ecf5c673b22504b37a1c0b6afa641cce82a6b0c75df3340877eb4945b548485f not found: ID does not exist" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.819675 4978 scope.go:117] "RemoveContainer" containerID="2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b" Feb 25 07:09:49 crc kubenswrapper[4978]: E0225 07:09:49.821740 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b\": container with ID starting with 2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b not found: ID does not exist" containerID="2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.821768 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b"} err="failed to get container status \"2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b\": rpc error: code = NotFound desc = could not find container \"2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b\": container with ID starting with 2b3328d05af175a3efe5972a9da3d2ab7a08b5c21283cf3100d4a1a1655c923b not found: ID does not exist" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.825837 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.834895 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:49 crc kubenswrapper[4978]: E0225 07:09:49.835325 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-metadata" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.835341 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-metadata" Feb 25 07:09:49 crc kubenswrapper[4978]: E0225 07:09:49.835391 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-log" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.835398 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-log" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.835571 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-metadata" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.835605 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" containerName="nova-metadata-log" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.836554 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.841406 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.842323 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:49 crc kubenswrapper[4978]: I0225 07:09:49.846829 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.004230 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.004266 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fgrc\" (UniqueName: \"kubernetes.io/projected/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-kube-api-access-6fgrc\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.004337 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.004416 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-logs\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.004434 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-config-data\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.107909 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.108104 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-logs\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.108149 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-config-data\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.108267 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.108313 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fgrc\" (UniqueName: \"kubernetes.io/projected/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-kube-api-access-6fgrc\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.108591 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-logs\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.114819 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.115007 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-config-data\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.122115 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.130897 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fgrc\" (UniqueName: \"kubernetes.io/projected/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-kube-api-access-6fgrc\") pod \"nova-metadata-0\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.151178 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.599612 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:09:50 crc kubenswrapper[4978]: I0225 07:09:50.755670 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8","Type":"ContainerStarted","Data":"4f341182a2e529d3fb3a7105185d2c313bf7306e97ceb3b01d58a2bf81fcb51a"} Feb 25 07:09:51 crc kubenswrapper[4978]: I0225 07:09:51.344213 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8876c6c-57b4-421d-bc34-a70080f86c63" path="/var/lib/kubelet/pods/b8876c6c-57b4-421d-bc34-a70080f86c63/volumes" Feb 25 07:09:51 crc kubenswrapper[4978]: I0225 07:09:51.773007 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8","Type":"ContainerStarted","Data":"33f35943890b2d404b619750739644e060d0152191e4ffd0d1bb79cdebfb8e65"} Feb 25 07:09:51 crc kubenswrapper[4978]: I0225 07:09:51.773072 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8","Type":"ContainerStarted","Data":"2881499eed920adb23f199015c996a12e05fc38be2beb2a1bf0cc87bbbb6b34a"} Feb 25 07:09:51 crc kubenswrapper[4978]: I0225 07:09:51.816120 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.816066816 podStartE2EDuration="2.816066816s" podCreationTimestamp="2026-02-25 07:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 07:09:51.799909797 +0000 UTC m=+1485.239166296" watchObservedRunningTime="2026-02-25 07:09:51.816066816 +0000 UTC m=+1485.255323305" Feb 25 07:09:53 crc kubenswrapper[4978]: I0225 07:09:53.092748 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 25 07:09:55 crc kubenswrapper[4978]: I0225 07:09:55.152005 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 07:09:55 crc kubenswrapper[4978]: I0225 07:09:55.152775 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 07:09:57 crc kubenswrapper[4978]: I0225 07:09:57.076879 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 07:09:57 crc kubenswrapper[4978]: I0225 07:09:57.077212 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 07:09:58 crc kubenswrapper[4978]: I0225 07:09:58.092325 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 25 07:09:58 crc kubenswrapper[4978]: I0225 07:09:58.126669 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 07:09:58 crc kubenswrapper[4978]: I0225 07:09:58.126720 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.209:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 07:09:58 crc kubenswrapper[4978]: I0225 07:09:58.152264 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 25 07:09:58 crc kubenswrapper[4978]: I0225 07:09:58.910481 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.153680 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.155283 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.191346 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533390-2vswx"] Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.193763 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533390-2vswx" Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.197203 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.197410 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.197786 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.222577 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533390-2vswx"] Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.340028 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcpm5\" (UniqueName: \"kubernetes.io/projected/0c029995-b152-4e0a-84f3-0e9c617ee812-kube-api-access-kcpm5\") pod \"auto-csr-approver-29533390-2vswx\" (UID: \"0c029995-b152-4e0a-84f3-0e9c617ee812\") " pod="openshift-infra/auto-csr-approver-29533390-2vswx" Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.442339 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcpm5\" (UniqueName: \"kubernetes.io/projected/0c029995-b152-4e0a-84f3-0e9c617ee812-kube-api-access-kcpm5\") pod \"auto-csr-approver-29533390-2vswx\" (UID: \"0c029995-b152-4e0a-84f3-0e9c617ee812\") " pod="openshift-infra/auto-csr-approver-29533390-2vswx" Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.475509 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcpm5\" (UniqueName: \"kubernetes.io/projected/0c029995-b152-4e0a-84f3-0e9c617ee812-kube-api-access-kcpm5\") pod \"auto-csr-approver-29533390-2vswx\" (UID: \"0c029995-b152-4e0a-84f3-0e9c617ee812\") " pod="openshift-infra/auto-csr-approver-29533390-2vswx" Feb 25 07:10:00 crc kubenswrapper[4978]: I0225 07:10:00.523616 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533390-2vswx" Feb 25 07:10:01 crc kubenswrapper[4978]: I0225 07:10:01.014390 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533390-2vswx"] Feb 25 07:10:01 crc kubenswrapper[4978]: W0225 07:10:01.019756 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c029995_b152_4e0a_84f3_0e9c617ee812.slice/crio-04e3582d7723c5f4da5200e9fd65bda46bc27fa637da80e8fbd34974169a1d77 WatchSource:0}: Error finding container 04e3582d7723c5f4da5200e9fd65bda46bc27fa637da80e8fbd34974169a1d77: Status 404 returned error can't find the container with id 04e3582d7723c5f4da5200e9fd65bda46bc27fa637da80e8fbd34974169a1d77 Feb 25 07:10:01 crc kubenswrapper[4978]: I0225 07:10:01.177607 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 07:10:01 crc kubenswrapper[4978]: I0225 07:10:01.177757 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 07:10:01 crc kubenswrapper[4978]: I0225 07:10:01.885104 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533390-2vswx" event={"ID":"0c029995-b152-4e0a-84f3-0e9c617ee812","Type":"ContainerStarted","Data":"04e3582d7723c5f4da5200e9fd65bda46bc27fa637da80e8fbd34974169a1d77"} Feb 25 07:10:03 crc kubenswrapper[4978]: I0225 07:10:03.911903 4978 generic.go:334] "Generic (PLEG): container finished" podID="0c029995-b152-4e0a-84f3-0e9c617ee812" containerID="09b0228a0ba6058441efd4b4d0555b7a4cdbaf500b54897828dd5c3c56a204ef" exitCode=0 Feb 25 07:10:03 crc kubenswrapper[4978]: I0225 07:10:03.912011 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533390-2vswx" event={"ID":"0c029995-b152-4e0a-84f3-0e9c617ee812","Type":"ContainerDied","Data":"09b0228a0ba6058441efd4b4d0555b7a4cdbaf500b54897828dd5c3c56a204ef"} Feb 25 07:10:04 crc kubenswrapper[4978]: I0225 07:10:04.857667 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 25 07:10:05 crc kubenswrapper[4978]: I0225 07:10:05.325865 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533390-2vswx" Feb 25 07:10:05 crc kubenswrapper[4978]: I0225 07:10:05.456745 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcpm5\" (UniqueName: \"kubernetes.io/projected/0c029995-b152-4e0a-84f3-0e9c617ee812-kube-api-access-kcpm5\") pod \"0c029995-b152-4e0a-84f3-0e9c617ee812\" (UID: \"0c029995-b152-4e0a-84f3-0e9c617ee812\") " Feb 25 07:10:05 crc kubenswrapper[4978]: I0225 07:10:05.468569 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c029995-b152-4e0a-84f3-0e9c617ee812-kube-api-access-kcpm5" (OuterVolumeSpecName: "kube-api-access-kcpm5") pod "0c029995-b152-4e0a-84f3-0e9c617ee812" (UID: "0c029995-b152-4e0a-84f3-0e9c617ee812"). InnerVolumeSpecName "kube-api-access-kcpm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:05 crc kubenswrapper[4978]: I0225 07:10:05.559721 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcpm5\" (UniqueName: \"kubernetes.io/projected/0c029995-b152-4e0a-84f3-0e9c617ee812-kube-api-access-kcpm5\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:05 crc kubenswrapper[4978]: I0225 07:10:05.940970 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533390-2vswx" event={"ID":"0c029995-b152-4e0a-84f3-0e9c617ee812","Type":"ContainerDied","Data":"04e3582d7723c5f4da5200e9fd65bda46bc27fa637da80e8fbd34974169a1d77"} Feb 25 07:10:05 crc kubenswrapper[4978]: I0225 07:10:05.941435 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04e3582d7723c5f4da5200e9fd65bda46bc27fa637da80e8fbd34974169a1d77" Feb 25 07:10:05 crc kubenswrapper[4978]: I0225 07:10:05.941031 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533390-2vswx" Feb 25 07:10:06 crc kubenswrapper[4978]: I0225 07:10:06.466127 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533384-9q4l8"] Feb 25 07:10:06 crc kubenswrapper[4978]: I0225 07:10:06.480788 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533384-9q4l8"] Feb 25 07:10:07 crc kubenswrapper[4978]: I0225 07:10:07.086236 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 07:10:07 crc kubenswrapper[4978]: I0225 07:10:07.086773 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 07:10:07 crc kubenswrapper[4978]: I0225 07:10:07.087273 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 07:10:07 crc kubenswrapper[4978]: I0225 07:10:07.087324 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 07:10:07 crc kubenswrapper[4978]: I0225 07:10:07.094160 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 07:10:07 crc kubenswrapper[4978]: I0225 07:10:07.102610 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 07:10:07 crc kubenswrapper[4978]: I0225 07:10:07.347288 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28930518-6de3-48af-94ff-63e0a1608fc4" path="/var/lib/kubelet/pods/28930518-6de3-48af-94ff-63e0a1608fc4/volumes" Feb 25 07:10:10 crc kubenswrapper[4978]: I0225 07:10:10.158997 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 07:10:10 crc kubenswrapper[4978]: I0225 07:10:10.163797 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 07:10:10 crc kubenswrapper[4978]: I0225 07:10:10.172280 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 07:10:11 crc kubenswrapper[4978]: I0225 07:10:11.012593 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 07:10:13 crc kubenswrapper[4978]: I0225 07:10:13.834492 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bwjnj"] Feb 25 07:10:13 crc kubenswrapper[4978]: E0225 07:10:13.835805 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c029995-b152-4e0a-84f3-0e9c617ee812" containerName="oc" Feb 25 07:10:13 crc kubenswrapper[4978]: I0225 07:10:13.835842 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c029995-b152-4e0a-84f3-0e9c617ee812" containerName="oc" Feb 25 07:10:13 crc kubenswrapper[4978]: I0225 07:10:13.836342 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c029995-b152-4e0a-84f3-0e9c617ee812" containerName="oc" Feb 25 07:10:13 crc kubenswrapper[4978]: I0225 07:10:13.839483 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:13 crc kubenswrapper[4978]: I0225 07:10:13.856810 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bwjnj"] Feb 25 07:10:13 crc kubenswrapper[4978]: I0225 07:10:13.947708 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-utilities\") pod \"redhat-operators-bwjnj\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:13 crc kubenswrapper[4978]: I0225 07:10:13.948164 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-catalog-content\") pod \"redhat-operators-bwjnj\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:13 crc kubenswrapper[4978]: I0225 07:10:13.948196 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smsvj\" (UniqueName: \"kubernetes.io/projected/b052d38e-931f-4eb1-85ea-142ef3da9881-kube-api-access-smsvj\") pod \"redhat-operators-bwjnj\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:14 crc kubenswrapper[4978]: I0225 07:10:14.049917 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-catalog-content\") pod \"redhat-operators-bwjnj\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:14 crc kubenswrapper[4978]: I0225 07:10:14.049968 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smsvj\" (UniqueName: \"kubernetes.io/projected/b052d38e-931f-4eb1-85ea-142ef3da9881-kube-api-access-smsvj\") pod \"redhat-operators-bwjnj\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:14 crc kubenswrapper[4978]: I0225 07:10:14.050040 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-utilities\") pod \"redhat-operators-bwjnj\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:14 crc kubenswrapper[4978]: I0225 07:10:14.050508 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-utilities\") pod \"redhat-operators-bwjnj\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:14 crc kubenswrapper[4978]: I0225 07:10:14.050560 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-catalog-content\") pod \"redhat-operators-bwjnj\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:14 crc kubenswrapper[4978]: I0225 07:10:14.073664 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smsvj\" (UniqueName: \"kubernetes.io/projected/b052d38e-931f-4eb1-85ea-142ef3da9881-kube-api-access-smsvj\") pod \"redhat-operators-bwjnj\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:14 crc kubenswrapper[4978]: I0225 07:10:14.185149 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:14 crc kubenswrapper[4978]: I0225 07:10:14.671115 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bwjnj"] Feb 25 07:10:15 crc kubenswrapper[4978]: I0225 07:10:15.045940 4978 generic.go:334] "Generic (PLEG): container finished" podID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerID="b4bc7b9cb20fc0680c1a54d545eae4f31248c2cdb972faf3ab5ce0ae43327b10" exitCode=0 Feb 25 07:10:15 crc kubenswrapper[4978]: I0225 07:10:15.046094 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwjnj" event={"ID":"b052d38e-931f-4eb1-85ea-142ef3da9881","Type":"ContainerDied","Data":"b4bc7b9cb20fc0680c1a54d545eae4f31248c2cdb972faf3ab5ce0ae43327b10"} Feb 25 07:10:15 crc kubenswrapper[4978]: I0225 07:10:15.046187 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwjnj" event={"ID":"b052d38e-931f-4eb1-85ea-142ef3da9881","Type":"ContainerStarted","Data":"c68878c14403d9b5655354b7c8052649095c8f102429053cec66c5675279fbdb"} Feb 25 07:10:16 crc kubenswrapper[4978]: I0225 07:10:16.540910 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:10:16 crc kubenswrapper[4978]: I0225 07:10:16.541324 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:10:17 crc kubenswrapper[4978]: I0225 07:10:17.514800 4978 scope.go:117] "RemoveContainer" containerID="52e71a0bce55e4c42a5a657ce506d03913c188711c74ade67560c87430666754" Feb 25 07:10:23 crc kubenswrapper[4978]: I0225 07:10:23.126856 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwjnj" event={"ID":"b052d38e-931f-4eb1-85ea-142ef3da9881","Type":"ContainerStarted","Data":"d443f24582cf5b7de4d32d1b3624d1fa54799a46d0e10d2aca168fcfdc53db2d"} Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.056095 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwmn"] Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.059019 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.071439 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwmn"] Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.187885 4978 generic.go:334] "Generic (PLEG): container finished" podID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerID="d443f24582cf5b7de4d32d1b3624d1fa54799a46d0e10d2aca168fcfdc53db2d" exitCode=0 Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.187929 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwjnj" event={"ID":"b052d38e-931f-4eb1-85ea-142ef3da9881","Type":"ContainerDied","Data":"d443f24582cf5b7de4d32d1b3624d1fa54799a46d0e10d2aca168fcfdc53db2d"} Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.202857 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-utilities\") pod \"redhat-marketplace-nxwmn\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.202890 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-catalog-content\") pod \"redhat-marketplace-nxwmn\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.202912 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj5qw\" (UniqueName: \"kubernetes.io/projected/34752acd-4270-408b-bc82-af9fc515da2b-kube-api-access-sj5qw\") pod \"redhat-marketplace-nxwmn\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.304679 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj5qw\" (UniqueName: \"kubernetes.io/projected/34752acd-4270-408b-bc82-af9fc515da2b-kube-api-access-sj5qw\") pod \"redhat-marketplace-nxwmn\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.305313 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-utilities\") pod \"redhat-marketplace-nxwmn\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.305363 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-catalog-content\") pod \"redhat-marketplace-nxwmn\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.305948 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-utilities\") pod \"redhat-marketplace-nxwmn\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.306108 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-catalog-content\") pod \"redhat-marketplace-nxwmn\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.325820 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj5qw\" (UniqueName: \"kubernetes.io/projected/34752acd-4270-408b-bc82-af9fc515da2b-kube-api-access-sj5qw\") pod \"redhat-marketplace-nxwmn\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.377734 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.684263 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.703638 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="ecf946d2-e6a8-4df6-ab76-b53d03db5d27" containerName="openstackclient" containerID="cri-o://b043db8fffc821daf3cf627f0556f152620f9dbbc236549845734482a005f759" gracePeriod=2 Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.763455 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.809440 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-lhjvp"] Feb 25 07:10:28 crc kubenswrapper[4978]: E0225 07:10:28.809969 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecf946d2-e6a8-4df6-ab76-b53d03db5d27" containerName="openstackclient" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.809983 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecf946d2-e6a8-4df6-ab76-b53d03db5d27" containerName="openstackclient" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.810178 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecf946d2-e6a8-4df6-ab76-b53d03db5d27" containerName="openstackclient" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.810862 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lhjvp" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.847069 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.849258 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-lhjvp"] Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.915153 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97e7baca-befe-4eba-9652-5474e0e88899-operator-scripts\") pod \"root-account-create-update-lhjvp\" (UID: \"97e7baca-befe-4eba-9652-5474e0e88899\") " pod="openstack/root-account-create-update-lhjvp" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.915260 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9vv7\" (UniqueName: \"kubernetes.io/projected/97e7baca-befe-4eba-9652-5474e0e88899-kube-api-access-b9vv7\") pod \"root-account-create-update-lhjvp\" (UID: \"97e7baca-befe-4eba-9652-5474e0e88899\") " pod="openstack/root-account-create-update-lhjvp" Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.920675 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-f7574"] Feb 25 07:10:28 crc kubenswrapper[4978]: I0225 07:10:28.955013 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-f7574"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.000487 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-cb34-account-create-update-2blwx"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.002505 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cb34-account-create-update-2blwx" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.005295 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.021128 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97e7baca-befe-4eba-9652-5474e0e88899-operator-scripts\") pod \"root-account-create-update-lhjvp\" (UID: \"97e7baca-befe-4eba-9652-5474e0e88899\") " pod="openstack/root-account-create-update-lhjvp" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.027732 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97e7baca-befe-4eba-9652-5474e0e88899-operator-scripts\") pod \"root-account-create-update-lhjvp\" (UID: \"97e7baca-befe-4eba-9652-5474e0e88899\") " pod="openstack/root-account-create-update-lhjvp" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.027991 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9vv7\" (UniqueName: \"kubernetes.io/projected/97e7baca-befe-4eba-9652-5474e0e88899-kube-api-access-b9vv7\") pod \"root-account-create-update-lhjvp\" (UID: \"97e7baca-befe-4eba-9652-5474e0e88899\") " pod="openstack/root-account-create-update-lhjvp" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.050318 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-1b7b-account-create-update-qlxzw"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.051454 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1b7b-account-create-update-qlxzw" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.058160 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.108326 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-cb34-account-create-update-2blwx"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.109335 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9vv7\" (UniqueName: \"kubernetes.io/projected/97e7baca-befe-4eba-9652-5474e0e88899-kube-api-access-b9vv7\") pod \"root-account-create-update-lhjvp\" (UID: \"97e7baca-befe-4eba-9652-5474e0e88899\") " pod="openstack/root-account-create-update-lhjvp" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.129400 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s8d7\" (UniqueName: \"kubernetes.io/projected/7cf4ce11-507f-4be4-9652-05035315d26d-kube-api-access-9s8d7\") pod \"placement-cb34-account-create-update-2blwx\" (UID: \"7cf4ce11-507f-4be4-9652-05035315d26d\") " pod="openstack/placement-cb34-account-create-update-2blwx" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.129648 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cf4ce11-507f-4be4-9652-05035315d26d-operator-scripts\") pod \"placement-cb34-account-create-update-2blwx\" (UID: \"7cf4ce11-507f-4be4-9652-05035315d26d\") " pod="openstack/placement-cb34-account-create-update-2blwx" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.129762 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkzsx\" (UniqueName: \"kubernetes.io/projected/bc147cb6-a5d8-47f7-8ce6-168455f18512-kube-api-access-zkzsx\") pod \"glance-1b7b-account-create-update-qlxzw\" (UID: \"bc147cb6-a5d8-47f7-8ce6-168455f18512\") " pod="openstack/glance-1b7b-account-create-update-qlxzw" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.129918 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc147cb6-a5d8-47f7-8ce6-168455f18512-operator-scripts\") pod \"glance-1b7b-account-create-update-qlxzw\" (UID: \"bc147cb6-a5d8-47f7-8ce6-168455f18512\") " pod="openstack/glance-1b7b-account-create-update-qlxzw" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.142864 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8a8c-account-create-update-d7wm6"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.144297 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-d7wm6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.152327 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.154427 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-cb34-account-create-update-k6dk6"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.197125 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lhjvp" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.226465 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-cb34-account-create-update-k6dk6"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.241496 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59csl\" (UniqueName: \"kubernetes.io/projected/83bd2445-af69-432a-bc55-d5cee299162d-kube-api-access-59csl\") pod \"barbican-8a8c-account-create-update-d7wm6\" (UID: \"83bd2445-af69-432a-bc55-d5cee299162d\") " pod="openstack/barbican-8a8c-account-create-update-d7wm6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.241559 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s8d7\" (UniqueName: \"kubernetes.io/projected/7cf4ce11-507f-4be4-9652-05035315d26d-kube-api-access-9s8d7\") pod \"placement-cb34-account-create-update-2blwx\" (UID: \"7cf4ce11-507f-4be4-9652-05035315d26d\") " pod="openstack/placement-cb34-account-create-update-2blwx" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.241583 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cf4ce11-507f-4be4-9652-05035315d26d-operator-scripts\") pod \"placement-cb34-account-create-update-2blwx\" (UID: \"7cf4ce11-507f-4be4-9652-05035315d26d\") " pod="openstack/placement-cb34-account-create-update-2blwx" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.241624 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkzsx\" (UniqueName: \"kubernetes.io/projected/bc147cb6-a5d8-47f7-8ce6-168455f18512-kube-api-access-zkzsx\") pod \"glance-1b7b-account-create-update-qlxzw\" (UID: \"bc147cb6-a5d8-47f7-8ce6-168455f18512\") " pod="openstack/glance-1b7b-account-create-update-qlxzw" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.241676 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bd2445-af69-432a-bc55-d5cee299162d-operator-scripts\") pod \"barbican-8a8c-account-create-update-d7wm6\" (UID: \"83bd2445-af69-432a-bc55-d5cee299162d\") " pod="openstack/barbican-8a8c-account-create-update-d7wm6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.241742 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc147cb6-a5d8-47f7-8ce6-168455f18512-operator-scripts\") pod \"glance-1b7b-account-create-update-qlxzw\" (UID: \"bc147cb6-a5d8-47f7-8ce6-168455f18512\") " pod="openstack/glance-1b7b-account-create-update-qlxzw" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.242408 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc147cb6-a5d8-47f7-8ce6-168455f18512-operator-scripts\") pod \"glance-1b7b-account-create-update-qlxzw\" (UID: \"bc147cb6-a5d8-47f7-8ce6-168455f18512\") " pod="openstack/glance-1b7b-account-create-update-qlxzw" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.243333 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cf4ce11-507f-4be4-9652-05035315d26d-operator-scripts\") pod \"placement-cb34-account-create-update-2blwx\" (UID: \"7cf4ce11-507f-4be4-9652-05035315d26d\") " pod="openstack/placement-cb34-account-create-update-2blwx" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.277725 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-1b7b-account-create-update-qlxzw"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.323112 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s8d7\" (UniqueName: \"kubernetes.io/projected/7cf4ce11-507f-4be4-9652-05035315d26d-kube-api-access-9s8d7\") pod \"placement-cb34-account-create-update-2blwx\" (UID: \"7cf4ce11-507f-4be4-9652-05035315d26d\") " pod="openstack/placement-cb34-account-create-update-2blwx" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.323181 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8a8c-account-create-update-d7wm6"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.355977 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkzsx\" (UniqueName: \"kubernetes.io/projected/bc147cb6-a5d8-47f7-8ce6-168455f18512-kube-api-access-zkzsx\") pod \"glance-1b7b-account-create-update-qlxzw\" (UID: \"bc147cb6-a5d8-47f7-8ce6-168455f18512\") " pod="openstack/glance-1b7b-account-create-update-qlxzw" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.358565 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59csl\" (UniqueName: \"kubernetes.io/projected/83bd2445-af69-432a-bc55-d5cee299162d-kube-api-access-59csl\") pod \"barbican-8a8c-account-create-update-d7wm6\" (UID: \"83bd2445-af69-432a-bc55-d5cee299162d\") " pod="openstack/barbican-8a8c-account-create-update-d7wm6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.358724 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bd2445-af69-432a-bc55-d5cee299162d-operator-scripts\") pod \"barbican-8a8c-account-create-update-d7wm6\" (UID: \"83bd2445-af69-432a-bc55-d5cee299162d\") " pod="openstack/barbican-8a8c-account-create-update-d7wm6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.365670 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cb34-account-create-update-2blwx" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.368840 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bd2445-af69-432a-bc55-d5cee299162d-operator-scripts\") pod \"barbican-8a8c-account-create-update-d7wm6\" (UID: \"83bd2445-af69-432a-bc55-d5cee299162d\") " pod="openstack/barbican-8a8c-account-create-update-d7wm6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.437508 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ce72ec5-eb9b-46f3-ae30-b8380615465c" path="/var/lib/kubelet/pods/1ce72ec5-eb9b-46f3-ae30-b8380615465c/volumes" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.438065 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59csl\" (UniqueName: \"kubernetes.io/projected/83bd2445-af69-432a-bc55-d5cee299162d-kube-api-access-59csl\") pod \"barbican-8a8c-account-create-update-d7wm6\" (UID: \"83bd2445-af69-432a-bc55-d5cee299162d\") " pod="openstack/barbican-8a8c-account-create-update-d7wm6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.441210 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="608f63c7-1251-42ea-996d-c7629a992f4a" path="/var/lib/kubelet/pods/608f63c7-1251-42ea-996d-c7629a992f4a/volumes" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.457911 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-1b7b-account-create-update-z4plm"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.464487 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1b7b-account-create-update-qlxzw" Feb 25 07:10:29 crc kubenswrapper[4978]: E0225 07:10:29.491697 4978 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Feb 25 07:10:29 crc kubenswrapper[4978]: E0225 07:10:29.491947 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data podName:a762aa77-9eb1-42eb-98e5-016cc3acac10 nodeName:}" failed. No retries permitted until 2026-02-25 07:10:29.991924229 +0000 UTC m=+1523.431180688 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data") pod "barbican-worker-7696768c55-mnqh9" (UID: "a762aa77-9eb1-42eb-98e5-016cc3acac10") : secret "barbican-config-data" not found Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.528388 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-d7wm6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.539930 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-1b7b-account-create-update-z4plm"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.602520 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-njndh"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.620500 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-4z765"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.621074 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-4z765" podUID="a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" containerName="openstack-network-exporter" containerID="cri-o://986d5b46fb6a179bd4f7544f78f8396ed0327bcfeea74164711b40b9816145da" gracePeriod=30 Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.641483 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.641773 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" containerName="cinder-scheduler" containerID="cri-o://abc550e17c1b943bad86597853b93ec464de90bcf8090229b1e9cb5d9aeca0d7" gracePeriod=30 Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.642162 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" containerName="probe" containerID="cri-o://8a5993b793c50295987a7b0055fc564f776c502b1b808c0832ceeaaacca5e945" gracePeriod=30 Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.669655 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6ec4-account-create-update-649r6"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.670922 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ec4-account-create-update-649r6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.673572 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.703637 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q7kb\" (UniqueName: \"kubernetes.io/projected/84e32f1f-5ad1-40a8-8f61-21025d84d003-kube-api-access-8q7kb\") pod \"neutron-6ec4-account-create-update-649r6\" (UID: \"84e32f1f-5ad1-40a8-8f61-21025d84d003\") " pod="openstack/neutron-6ec4-account-create-update-649r6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.703719 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84e32f1f-5ad1-40a8-8f61-21025d84d003-operator-scripts\") pod \"neutron-6ec4-account-create-update-649r6\" (UID: \"84e32f1f-5ad1-40a8-8f61-21025d84d003\") " pod="openstack/neutron-6ec4-account-create-update-649r6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.712220 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-r45vg"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.733392 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6ec4-account-create-update-649r6"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.780431 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8a8c-account-create-update-sffwp"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.805198 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q7kb\" (UniqueName: \"kubernetes.io/projected/84e32f1f-5ad1-40a8-8f61-21025d84d003-kube-api-access-8q7kb\") pod \"neutron-6ec4-account-create-update-649r6\" (UID: \"84e32f1f-5ad1-40a8-8f61-21025d84d003\") " pod="openstack/neutron-6ec4-account-create-update-649r6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.805313 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84e32f1f-5ad1-40a8-8f61-21025d84d003-operator-scripts\") pod \"neutron-6ec4-account-create-update-649r6\" (UID: \"84e32f1f-5ad1-40a8-8f61-21025d84d003\") " pod="openstack/neutron-6ec4-account-create-update-649r6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.806704 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84e32f1f-5ad1-40a8-8f61-21025d84d003-operator-scripts\") pod \"neutron-6ec4-account-create-update-649r6\" (UID: \"84e32f1f-5ad1-40a8-8f61-21025d84d003\") " pod="openstack/neutron-6ec4-account-create-update-649r6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.823879 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8a8c-account-create-update-sffwp"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.840921 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q7kb\" (UniqueName: \"kubernetes.io/projected/84e32f1f-5ad1-40a8-8f61-21025d84d003-kube-api-access-8q7kb\") pod \"neutron-6ec4-account-create-update-649r6\" (UID: \"84e32f1f-5ad1-40a8-8f61-21025d84d003\") " pod="openstack/neutron-6ec4-account-create-update-649r6" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.848666 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.849154 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="807eb4fa-5af0-450e-be50-f30c96731b46" containerName="cinder-api-log" containerID="cri-o://05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc" gracePeriod=30 Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.849528 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="807eb4fa-5af0-450e-be50-f30c96731b46" containerName="cinder-api" containerID="cri-o://e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644" gracePeriod=30 Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.868427 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.868715 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerName="ovn-northd" containerID="cri-o://6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98" gracePeriod=30 Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.869145 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerName="openstack-network-exporter" containerID="cri-o://c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d" gracePeriod=30 Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.881310 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-9eb7-account-create-update-f2s6t"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.887999 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9eb7-account-create-update-f2s6t" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.895140 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.896644 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9eb7-account-create-update-f2s6t"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.931720 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6ec4-account-create-update-8mwd5"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.948443 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6ec4-account-create-update-8mwd5"] Feb 25 07:10:29 crc kubenswrapper[4978]: I0225 07:10:29.971450 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwmn"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.010255 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9eb7-account-create-update-zszr6"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.013387 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-operator-scripts\") pod \"cinder-9eb7-account-create-update-f2s6t\" (UID: \"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce\") " pod="openstack/cinder-9eb7-account-create-update-f2s6t" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.013485 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzlkq\" (UniqueName: \"kubernetes.io/projected/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-kube-api-access-mzlkq\") pod \"cinder-9eb7-account-create-update-f2s6t\" (UID: \"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce\") " pod="openstack/cinder-9eb7-account-create-update-f2s6t" Feb 25 07:10:30 crc kubenswrapper[4978]: E0225 07:10:30.013685 4978 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Feb 25 07:10:30 crc kubenswrapper[4978]: E0225 07:10:30.013735 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data podName:a762aa77-9eb1-42eb-98e5-016cc3acac10 nodeName:}" failed. No retries permitted until 2026-02-25 07:10:31.013719236 +0000 UTC m=+1524.452975695 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data") pod "barbican-worker-7696768c55-mnqh9" (UID: "a762aa77-9eb1-42eb-98e5-016cc3acac10") : secret "barbican-config-data" not found Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.034165 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ec4-account-create-update-649r6" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.048612 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-9eb7-account-create-update-zszr6"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.072560 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-7ts6t"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.093609 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-7ts6t"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.119489 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzlkq\" (UniqueName: \"kubernetes.io/projected/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-kube-api-access-mzlkq\") pod \"cinder-9eb7-account-create-update-f2s6t\" (UID: \"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce\") " pod="openstack/cinder-9eb7-account-create-update-f2s6t" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.119632 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-operator-scripts\") pod \"cinder-9eb7-account-create-update-f2s6t\" (UID: \"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce\") " pod="openstack/cinder-9eb7-account-create-update-f2s6t" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.120475 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-operator-scripts\") pod \"cinder-9eb7-account-create-update-f2s6t\" (UID: \"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce\") " pod="openstack/cinder-9eb7-account-create-update-f2s6t" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.127028 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.143175 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-9373-account-create-update-c5bk6"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.158044 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzlkq\" (UniqueName: \"kubernetes.io/projected/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-kube-api-access-mzlkq\") pod \"cinder-9eb7-account-create-update-f2s6t\" (UID: \"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce\") " pod="openstack/cinder-9eb7-account-create-update-f2s6t" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.159725 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-flx4x"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.181446 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-9373-account-create-update-c5bk6"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.212173 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-flx4x"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.220331 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c64474bcc-4rmct"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.220579 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" podUID="e78e5151-267c-4b33-9473-6b249b01f12f" containerName="dnsmasq-dns" containerID="cri-o://4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706" gracePeriod=10 Feb 25 07:10:30 crc kubenswrapper[4978]: E0225 07:10:30.221152 4978 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Feb 25 07:10:30 crc kubenswrapper[4978]: E0225 07:10:30.221195 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data podName:09112202-a3ef-4742-bdbe-b46c1c9a55ba nodeName:}" failed. No retries permitted until 2026-02-25 07:10:30.721181583 +0000 UTC m=+1524.160438042 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data") pod "rabbitmq-server-0" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba") : configmap "rabbitmq-config-data" not found Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.228419 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-9373-account-create-update-lqszb"] Feb 25 07:10:30 crc kubenswrapper[4978]: E0225 07:10:30.230207 4978 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-r45vg" message="Exiting ovn-controller (1) " Feb 25 07:10:30 crc kubenswrapper[4978]: E0225 07:10:30.230241 4978 kuberuntime_container.go:691] "PreStop hook failed" err="command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: " pod="openstack/ovn-controller-r45vg" podUID="9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" containerName="ovn-controller" containerID="cri-o://aefb1dba292711435eac1e73e9e529ab2e03a7b4863f808c55d3071e985aa85e" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.230270 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-r45vg" podUID="9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" containerName="ovn-controller" containerID="cri-o://aefb1dba292711435eac1e73e9e529ab2e03a7b4863f808c55d3071e985aa85e" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.243448 4978 generic.go:334] "Generic (PLEG): container finished" podID="807eb4fa-5af0-450e-be50-f30c96731b46" containerID="05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc" exitCode=143 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.244180 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"807eb4fa-5af0-450e-be50-f30c96731b46","Type":"ContainerDied","Data":"05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc"} Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.244346 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9373-account-create-update-lqszb" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.245736 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4z765_a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682/openstack-network-exporter/0.log" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.245762 4978 generic.go:334] "Generic (PLEG): container finished" podID="a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" containerID="986d5b46fb6a179bd4f7544f78f8396ed0327bcfeea74164711b40b9816145da" exitCode=2 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.245804 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4z765" event={"ID":"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682","Type":"ContainerDied","Data":"986d5b46fb6a179bd4f7544f78f8396ed0327bcfeea74164711b40b9816145da"} Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.249060 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.249293 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9373-account-create-update-lqszb"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.264643 4978 generic.go:334] "Generic (PLEG): container finished" podID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerID="c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d" exitCode=2 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.264728 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"27864863-0a60-43ec-ac3e-f86d9f68c7d2","Type":"ContainerDied","Data":"c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d"} Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.275481 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwmn" event={"ID":"34752acd-4270-408b-bc82-af9fc515da2b","Type":"ContainerStarted","Data":"840671ede6def3b430ea5faed42cf1a9b295e209141efd979450b48941cc4fe0"} Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.351377 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-tc5ks"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.354585 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whfrj\" (UniqueName: \"kubernetes.io/projected/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-kube-api-access-whfrj\") pod \"nova-api-9373-account-create-update-lqszb\" (UID: \"cbba37d2-4c5c-467c-9e1c-90c2c3aea110\") " pod="openstack/nova-api-9373-account-create-update-lqszb" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.354788 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-operator-scripts\") pod \"nova-api-9373-account-create-update-lqszb\" (UID: \"cbba37d2-4c5c-467c-9e1c-90c2c3aea110\") " pod="openstack/nova-api-9373-account-create-update-lqszb" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.377856 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-tc5ks"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.412586 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-b71c-account-create-update-jjh4g"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.432550 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-b71c-account-create-update-jjh4g"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.450044 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.450956 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerName="openstack-network-exporter" containerID="cri-o://c227278c28e8d88bb7c9ddfa3eaa4e8de5003b5ef2d93630c76d010d8ca314e0" gracePeriod=300 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.463565 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2sqb8"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.468376 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whfrj\" (UniqueName: \"kubernetes.io/projected/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-kube-api-access-whfrj\") pod \"nova-api-9373-account-create-update-lqszb\" (UID: \"cbba37d2-4c5c-467c-9e1c-90c2c3aea110\") " pod="openstack/nova-api-9373-account-create-update-lqszb" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.468450 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-operator-scripts\") pod \"nova-api-9373-account-create-update-lqszb\" (UID: \"cbba37d2-4c5c-467c-9e1c-90c2c3aea110\") " pod="openstack/nova-api-9373-account-create-update-lqszb" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.472048 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-operator-scripts\") pod \"nova-api-9373-account-create-update-lqszb\" (UID: \"cbba37d2-4c5c-467c-9e1c-90c2c3aea110\") " pod="openstack/nova-api-9373-account-create-update-lqszb" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.484760 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2sqb8"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.493943 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whfrj\" (UniqueName: \"kubernetes.io/projected/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-kube-api-access-whfrj\") pod \"nova-api-9373-account-create-update-lqszb\" (UID: \"cbba37d2-4c5c-467c-9e1c-90c2c3aea110\") " pod="openstack/nova-api-9373-account-create-update-lqszb" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.500274 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-xpwds"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.515222 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-xpwds"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.534404 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.535021 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="ed9859d2-701a-4011-b9d6-d8cf6291265a" containerName="openstack-network-exporter" containerID="cri-o://a15f02424a1013ac6691c287f99c51cc3e6aff1c7146b986bb694436609fc027" gracePeriod=300 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.556450 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.556719 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" containerName="glance-log" containerID="cri-o://251a79da9fba47a65d4db2f16a1ee10da3ad7f02ffcf859eb40da544a9209781" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.557129 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" containerName="glance-httpd" containerID="cri-o://231167ca97b4be8ade22723cde0ed17906f6163a3cedd000bd07b8e2dddaaa83" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.608536 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-zszqp"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.637570 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-69794fc6d6-zqnw6"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.637798 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-69794fc6d6-zqnw6" podUID="c6a16d36-c629-483e-9b15-fbd7b8913883" containerName="placement-log" containerID="cri-o://9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.637918 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-69794fc6d6-zqnw6" podUID="c6a16d36-c629-483e-9b15-fbd7b8913883" containerName="placement-api" containerID="cri-o://03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.687619 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-zszqp"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.699691 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-kr7xw"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.714567 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-kr7xw"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.721545 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerName="ovsdbserver-sb" containerID="cri-o://c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7" gracePeriod=300 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.724935 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.725171 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" containerName="glance-log" containerID="cri-o://23b908a2d2c8c5b19239c4d15a74e513b855d7ec6155c2aab43b358150b36fe5" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.725573 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" containerName="glance-httpd" containerID="cri-o://4b497e57dce20bdb93e7a82daed8baa0e9acd2b6048c9e18954a4341399c8f5b" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.748348 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ac49-account-create-update-2rrpr"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.763154 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="ed9859d2-701a-4011-b9d6-d8cf6291265a" containerName="ovsdbserver-nb" containerID="cri-o://8baa6e062d410fe1594d058fc3924f5a1ed85894951c8f77a4918acff9d1e1bd" gracePeriod=300 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.771044 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ac49-account-create-update-2rrpr"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.786875 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-z5fgm"] Feb 25 07:10:30 crc kubenswrapper[4978]: E0225 07:10:30.789869 4978 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Feb 25 07:10:30 crc kubenswrapper[4978]: E0225 07:10:30.789931 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data podName:09112202-a3ef-4742-bdbe-b46c1c9a55ba nodeName:}" failed. No retries permitted until 2026-02-25 07:10:31.789916673 +0000 UTC m=+1525.229173132 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data") pod "rabbitmq-server-0" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba") : configmap "rabbitmq-config-data" not found Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.802103 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-z5fgm"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.818060 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-1b7b-account-create-update-qlxzw"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.827417 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.827930 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-server" containerID="cri-o://fe36f33cfa027ca4f84b465a94aa8c1bbd4436f1982627dab087ee9bdbe66214" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828307 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="swift-recon-cron" containerID="cri-o://ce64ee6e29ee6ea1ff2d376f27c1117c98cf4060372883033763a60e602660f9" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828355 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="rsync" containerID="cri-o://d908490f57d2f2a43ad75a1a8576681b6638c3b6cc2217e47bfbe190b3f31f3b" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828405 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-expirer" containerID="cri-o://4dfed42feb8ac7a42412c46dcae347795ab03fb0af477b3ffbb5372f2f1e92c3" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828437 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-updater" containerID="cri-o://c5c66777c2f36490e61ea071a8cde68af4d8aa4113f9db174c104aae89ea3baf" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828474 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-server" containerID="cri-o://2c1a63624536475bb718a04a658404433f192ece6f6682b1e5ca98142ac9eb79" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828504 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-reaper" containerID="cri-o://c196bca824e71c59dc1d9939514fc9f5d009c35f1cb8b40a2d2b99b8d722f60b" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828499 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-replicator" containerID="cri-o://36f71fa276b25e88398462e3bbc6d6b87e2bee8c2aade1e2f5f0ffdc61440473" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828544 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-auditor" containerID="cri-o://25a3891e332bc2d905bcc5238aea824519156ba640e57709ba7a03f1986ab36b" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828573 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-replicator" containerID="cri-o://bb9960184024cbe585b528d58a3dd1a8b9fdeda2c6566bd7d76615473e70311e" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828612 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-auditor" containerID="cri-o://1ec7901b94aa973e76af694ac0b6f07b5c9788edfe7b888e4ae6f0ec6806b20b" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828648 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-replicator" containerID="cri-o://6fefdf96f9ea31d2462085c35bca5e00e4f192d43b6b9dbe2f57601d1afde09c" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828684 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-server" containerID="cri-o://cfdc985f7b5bb332d961a90b34a12545a7880322a9b3eaa992362b444ccba073" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828719 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-updater" containerID="cri-o://f892ef9c825b59066a0c023cd3b507cce63da08b00077af4f81ced0830c89f1c" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.828752 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-auditor" containerID="cri-o://c247f762435a5bf5a26de2224e27277501e932169aad0cc00bd5b31c4c134647" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.831643 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-2dsvp"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.869103 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-2dsvp"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.899085 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.925686 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-d4b94"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.948921 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-d4b94"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.985095 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5cdb588b97-7jkd8"] Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.985868 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5cdb588b97-7jkd8" podUID="d1ba0ae6-d125-48d2-a866-615d927dc525" containerName="neutron-api" containerID="cri-o://35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e" gracePeriod=30 Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.985919 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9eb7-account-create-update-f2s6t" Feb 25 07:10:30 crc kubenswrapper[4978]: I0225 07:10:30.986207 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5cdb588b97-7jkd8" podUID="d1ba0ae6-d125-48d2-a866-615d927dc525" containerName="neutron-httpd" containerID="cri-o://6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.018744 4978 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.018948 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data podName:a762aa77-9eb1-42eb-98e5-016cc3acac10 nodeName:}" failed. No retries permitted until 2026-02-25 07:10:33.018917751 +0000 UTC m=+1526.458174210 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data") pod "barbican-worker-7696768c55-mnqh9" (UID: "a762aa77-9eb1-42eb-98e5-016cc3acac10") : secret "barbican-config-data" not found Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.047123 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9373-account-create-update-lqszb" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.059900 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-cb34-account-create-update-2blwx"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.079205 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.083414 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-w7ztt"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.091448 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-w7ztt"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.104428 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8a8c-account-create-update-d7wm6"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.107548 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-99k22"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.117764 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-99k22"] Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.153263 4978 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.153564 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data podName:43bd8325-9408-4a32-9285-d522d497205e nodeName:}" failed. No retries permitted until 2026-02-25 07:10:31.653451756 +0000 UTC m=+1525.092708205 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data") pod "rabbitmq-cell1-server-0" (UID: "43bd8325-9408-4a32-9285-d522d497205e") : configmap "rabbitmq-cell1-config-data" not found Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.167701 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.168038 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-log" containerID="cri-o://2881499eed920adb23f199015c996a12e05fc38be2beb2a1bf0cc87bbbb6b34a" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.168671 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-metadata" containerID="cri-o://33f35943890b2d404b619750739644e060d0152191e4ffd0d1bb79cdebfb8e65" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.224020 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7696768c55-mnqh9"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.224599 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7696768c55-mnqh9" podUID="a762aa77-9eb1-42eb-98e5-016cc3acac10" containerName="barbican-worker-log" containerID="cri-o://47894865746e5efcf05b88359453cdd68cf29193d3fc760e152761c140f46884" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.224949 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7696768c55-mnqh9" podUID="a762aa77-9eb1-42eb-98e5-016cc3acac10" containerName="barbican-worker" containerID="cri-o://97d4b20d6d6a152207b32f06fed0f8b590209ec24594d4831ee695bb8b21c4ee" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.240136 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-sb\") pod \"e78e5151-267c-4b33-9473-6b249b01f12f\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.240193 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-swift-storage-0\") pod \"e78e5151-267c-4b33-9473-6b249b01f12f\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.240258 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-nb\") pod \"e78e5151-267c-4b33-9473-6b249b01f12f\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.240379 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k26lx\" (UniqueName: \"kubernetes.io/projected/e78e5151-267c-4b33-9473-6b249b01f12f-kube-api-access-k26lx\") pod \"e78e5151-267c-4b33-9473-6b249b01f12f\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.240432 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-config\") pod \"e78e5151-267c-4b33-9473-6b249b01f12f\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.240496 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-svc\") pod \"e78e5151-267c-4b33-9473-6b249b01f12f\" (UID: \"e78e5151-267c-4b33-9473-6b249b01f12f\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.263458 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9eb7-account-create-update-f2s6t"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.296309 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.300802 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78e5151-267c-4b33-9473-6b249b01f12f-kube-api-access-k26lx" (OuterVolumeSpecName: "kube-api-access-k26lx") pod "e78e5151-267c-4b33-9473-6b249b01f12f" (UID: "e78e5151-267c-4b33-9473-6b249b01f12f"). InnerVolumeSpecName "kube-api-access-k26lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.312679 4978 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Feb 25 07:10:31 crc kubenswrapper[4978]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Feb 25 07:10:31 crc kubenswrapper[4978]: + source /usr/local/bin/container-scripts/functions Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNBridge=br-int Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNRemote=tcp:localhost:6642 Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNEncapType=geneve Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNAvailabilityZones= Feb 25 07:10:31 crc kubenswrapper[4978]: ++ EnableChassisAsGateway=true Feb 25 07:10:31 crc kubenswrapper[4978]: ++ PhysicalNetworks= Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNHostName= Feb 25 07:10:31 crc kubenswrapper[4978]: ++ DB_FILE=/etc/openvswitch/conf.db Feb 25 07:10:31 crc kubenswrapper[4978]: ++ ovs_dir=/var/lib/openvswitch Feb 25 07:10:31 crc kubenswrapper[4978]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Feb 25 07:10:31 crc kubenswrapper[4978]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Feb 25 07:10:31 crc kubenswrapper[4978]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 25 07:10:31 crc kubenswrapper[4978]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 25 07:10:31 crc kubenswrapper[4978]: + sleep 0.5 Feb 25 07:10:31 crc kubenswrapper[4978]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 25 07:10:31 crc kubenswrapper[4978]: + sleep 0.5 Feb 25 07:10:31 crc kubenswrapper[4978]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 25 07:10:31 crc kubenswrapper[4978]: + cleanup_ovsdb_server_semaphore Feb 25 07:10:31 crc kubenswrapper[4978]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 25 07:10:31 crc kubenswrapper[4978]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Feb 25 07:10:31 crc kubenswrapper[4978]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-njndh" message=< Feb 25 07:10:31 crc kubenswrapper[4978]: Exiting ovsdb-server (5) [ OK ] Feb 25 07:10:31 crc kubenswrapper[4978]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Feb 25 07:10:31 crc kubenswrapper[4978]: + source /usr/local/bin/container-scripts/functions Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNBridge=br-int Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNRemote=tcp:localhost:6642 Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNEncapType=geneve Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNAvailabilityZones= Feb 25 07:10:31 crc kubenswrapper[4978]: ++ EnableChassisAsGateway=true Feb 25 07:10:31 crc kubenswrapper[4978]: ++ PhysicalNetworks= Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNHostName= Feb 25 07:10:31 crc kubenswrapper[4978]: ++ DB_FILE=/etc/openvswitch/conf.db Feb 25 07:10:31 crc kubenswrapper[4978]: ++ ovs_dir=/var/lib/openvswitch Feb 25 07:10:31 crc kubenswrapper[4978]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Feb 25 07:10:31 crc kubenswrapper[4978]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Feb 25 07:10:31 crc kubenswrapper[4978]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 25 07:10:31 crc kubenswrapper[4978]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 25 07:10:31 crc kubenswrapper[4978]: + sleep 0.5 Feb 25 07:10:31 crc kubenswrapper[4978]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 25 07:10:31 crc kubenswrapper[4978]: + sleep 0.5 Feb 25 07:10:31 crc kubenswrapper[4978]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 25 07:10:31 crc kubenswrapper[4978]: + cleanup_ovsdb_server_semaphore Feb 25 07:10:31 crc kubenswrapper[4978]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 25 07:10:31 crc kubenswrapper[4978]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Feb 25 07:10:31 crc kubenswrapper[4978]: > Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.312746 4978 kuberuntime_container.go:691] "PreStop hook failed" err=< Feb 25 07:10:31 crc kubenswrapper[4978]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Feb 25 07:10:31 crc kubenswrapper[4978]: + source /usr/local/bin/container-scripts/functions Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNBridge=br-int Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNRemote=tcp:localhost:6642 Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNEncapType=geneve Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNAvailabilityZones= Feb 25 07:10:31 crc kubenswrapper[4978]: ++ EnableChassisAsGateway=true Feb 25 07:10:31 crc kubenswrapper[4978]: ++ PhysicalNetworks= Feb 25 07:10:31 crc kubenswrapper[4978]: ++ OVNHostName= Feb 25 07:10:31 crc kubenswrapper[4978]: ++ DB_FILE=/etc/openvswitch/conf.db Feb 25 07:10:31 crc kubenswrapper[4978]: ++ ovs_dir=/var/lib/openvswitch Feb 25 07:10:31 crc kubenswrapper[4978]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Feb 25 07:10:31 crc kubenswrapper[4978]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Feb 25 07:10:31 crc kubenswrapper[4978]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 25 07:10:31 crc kubenswrapper[4978]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 25 07:10:31 crc kubenswrapper[4978]: + sleep 0.5 Feb 25 07:10:31 crc kubenswrapper[4978]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 25 07:10:31 crc kubenswrapper[4978]: + sleep 0.5 Feb 25 07:10:31 crc kubenswrapper[4978]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Feb 25 07:10:31 crc kubenswrapper[4978]: + cleanup_ovsdb_server_semaphore Feb 25 07:10:31 crc kubenswrapper[4978]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Feb 25 07:10:31 crc kubenswrapper[4978]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Feb 25 07:10:31 crc kubenswrapper[4978]: > pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" containerID="cri-o://f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.312783 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" containerID="cri-o://f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" gracePeriod=29 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.315138 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-77bc5d768-qx88q"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.320641 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-77bc5d768-qx88q" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerName="barbican-api-log" containerID="cri-o://3d765602f0864d6d998940f94e6b40eaa60dbf1a6c6bdb9427a47ca7993400a3" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.321476 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-77bc5d768-qx88q" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerName="barbican-api" containerID="cri-o://e3bb5987c1f9fda6b9f36d86116e1f3a943d9693a3917d0c5c728a4dd3429744" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.388732 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k26lx\" (UniqueName: \"kubernetes.io/projected/e78e5151-267c-4b33-9473-6b249b01f12f-kube-api-access-k26lx\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.423541 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovs-vswitchd" containerID="cri-o://02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" gracePeriod=29 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.424127 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ee32dc0-fe37-4be7-b87c-f605febf9692" path="/var/lib/kubelet/pods/2ee32dc0-fe37-4be7-b87c-f605febf9692/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.424696 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fcf1162-2ff4-4faa-89f0-272fd5b2669b" path="/var/lib/kubelet/pods/2fcf1162-2ff4-4faa-89f0-272fd5b2669b/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.426289 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e78e5151-267c-4b33-9473-6b249b01f12f" (UID: "e78e5151-267c-4b33-9473-6b249b01f12f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.427668 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e78e5151-267c-4b33-9473-6b249b01f12f" (UID: "e78e5151-267c-4b33-9473-6b249b01f12f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.429553 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.431216 4978 generic.go:334] "Generic (PLEG): container finished" podID="e78e5151-267c-4b33-9473-6b249b01f12f" containerID="4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.439849 4978 generic.go:334] "Generic (PLEG): container finished" podID="ecf946d2-e6a8-4df6-ab76-b53d03db5d27" containerID="b043db8fffc821daf3cf627f0556f152620f9dbbc236549845734482a005f759" exitCode=137 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.446610 4978 generic.go:334] "Generic (PLEG): container finished" podID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" containerID="23b908a2d2c8c5b19239c4d15a74e513b855d7ec6155c2aab43b358150b36fe5" exitCode=143 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.469821 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="330d6675-e003-4a9c-862a-88a54e67877e" path="/var/lib/kubelet/pods/330d6675-e003-4a9c-862a-88a54e67877e/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.470351 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="333c5a0b-874e-45cc-a44e-4b836b4bd850" path="/var/lib/kubelet/pods/333c5a0b-874e-45cc-a44e-4b836b4bd850/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.477688 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="4dfed42feb8ac7a42412c46dcae347795ab03fb0af477b3ffbb5372f2f1e92c3" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.481880 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="c5c66777c2f36490e61ea071a8cde68af4d8aa4113f9db174c104aae89ea3baf" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.481894 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="1ec7901b94aa973e76af694ac0b6f07b5c9788edfe7b888e4ae6f0ec6806b20b" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.481901 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="6fefdf96f9ea31d2462085c35bca5e00e4f192d43b6b9dbe2f57601d1afde09c" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.481911 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="f892ef9c825b59066a0c023cd3b507cce63da08b00077af4f81ced0830c89f1c" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.481919 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="c247f762435a5bf5a26de2224e27277501e932169aad0cc00bd5b31c4c134647" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.481926 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="36f71fa276b25e88398462e3bbc6d6b87e2bee8c2aade1e2f5f0ffdc61440473" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.481934 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="c196bca824e71c59dc1d9939514fc9f5d009c35f1cb8b40a2d2b99b8d722f60b" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.481941 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="25a3891e332bc2d905bcc5238aea824519156ba640e57709ba7a03f1986ab36b" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.481949 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="bb9960184024cbe585b528d58a3dd1a8b9fdeda2c6566bd7d76615473e70311e" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.481957 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="fe36f33cfa027ca4f84b465a94aa8c1bbd4436f1982627dab087ee9bdbe66214" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.482689 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bd97fe5-291b-4130-98c7-4ffbc0cba189" path="/var/lib/kubelet/pods/4bd97fe5-291b-4130-98c7-4ffbc0cba189/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.483925 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d7807bf-3cea-45ed-b534-29b26ebe2c12" path="/var/lib/kubelet/pods/4d7807bf-3cea-45ed-b534-29b26ebe2c12/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.484811 4978 generic.go:334] "Generic (PLEG): container finished" podID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" containerID="251a79da9fba47a65d4db2f16a1ee10da3ad7f02ffcf859eb40da544a9209781" exitCode=143 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.484948 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a73e3f-747f-4467-95c5-aebf41f4cc9b" path="/var/lib/kubelet/pods/51a73e3f-747f-4467-95c5-aebf41f4cc9b/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.485831 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a291e6-466e-456d-8674-690f1d8d4a1b" path="/var/lib/kubelet/pods/83a291e6-466e-456d-8674-690f1d8d4a1b/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.487042 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2cdb535-d30e-445f-97f9-259185cb4223" path="/var/lib/kubelet/pods/a2cdb535-d30e-445f-97f9-259185cb4223/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.488132 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a872e452-e198-4c3d-a8d4-85473516feab" path="/var/lib/kubelet/pods/a872e452-e198-4c3d-a8d4-85473516feab/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.489832 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.489857 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.492489 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ed9859d2-701a-4011-b9d6-d8cf6291265a/ovsdbserver-nb/0.log" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.492519 4978 generic.go:334] "Generic (PLEG): container finished" podID="ed9859d2-701a-4011-b9d6-d8cf6291265a" containerID="a15f02424a1013ac6691c287f99c51cc3e6aff1c7146b986bb694436609fc027" exitCode=2 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.492534 4978 generic.go:334] "Generic (PLEG): container finished" podID="ed9859d2-701a-4011-b9d6-d8cf6291265a" containerID="8baa6e062d410fe1594d058fc3924f5a1ed85894951c8f77a4918acff9d1e1bd" exitCode=143 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.494681 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5c931ba-06fd-41eb-9bdc-514ec3faef6e" path="/var/lib/kubelet/pods/b5c931ba-06fd-41eb-9bdc-514ec3faef6e/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.499270 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b86b8158-9329-44e8-9268-d5ed64155d48" path="/var/lib/kubelet/pods/b86b8158-9329-44e8-9268-d5ed64155d48/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.499924 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1" path="/var/lib/kubelet/pods/bcb2bbd3-7d86-4854-ab81-1e7ce97e8eb1/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.500500 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c183deb4-f193-4873-9617-5447f5c36e35" path="/var/lib/kubelet/pods/c183deb4-f193-4873-9617-5447f5c36e35/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.501043 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccea553d-1b65-436c-aa35-3b0217c92689" path="/var/lib/kubelet/pods/ccea553d-1b65-436c-aa35-3b0217c92689/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.503615 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0764b55-360c-46e0-a702-28ddb93b475d" path="/var/lib/kubelet/pods/e0764b55-360c-46e0-a702-28ddb93b475d/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.505709 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2a54702-6977-47b4-9e4e-11bc125b4981" path="/var/lib/kubelet/pods/e2a54702-6977-47b4-9e4e-11bc125b4981/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.506262 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f18ee3ff-994c-46c4-8f0c-76b1ee45da50" path="/var/lib/kubelet/pods/f18ee3ff-994c-46c4-8f0c-76b1ee45da50/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.507491 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3143797-ced6-46e4-a06a-b3790384dc8c" path="/var/lib/kubelet/pods/f3143797-ced6-46e4-a06a-b3790384dc8c/volumes" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.510608 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e78e5151-267c-4b33-9473-6b249b01f12f" (UID: "e78e5151-267c-4b33-9473-6b249b01f12f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.511678 4978 generic.go:334] "Generic (PLEG): container finished" podID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" containerID="8a5993b793c50295987a7b0055fc564f776c502b1b808c0832ceeaaacca5e945" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.521746 4978 generic.go:334] "Generic (PLEG): container finished" podID="c6a16d36-c629-483e-9b15-fbd7b8913883" containerID="9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af" exitCode=143 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.524239 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e78e5151-267c-4b33-9473-6b249b01f12f" (UID: "e78e5151-267c-4b33-9473-6b249b01f12f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.526968 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6df7a672-b889-40d3-abe5-271e4d36dd85/ovsdbserver-sb/0.log" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.527001 4978 generic.go:334] "Generic (PLEG): container finished" podID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerID="c227278c28e8d88bb7c9ddfa3eaa4e8de5003b5ef2d93630c76d010d8ca314e0" exitCode=2 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.527023 4978 generic.go:334] "Generic (PLEG): container finished" podID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerID="c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7" exitCode=143 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.536699 4978 generic.go:334] "Generic (PLEG): container finished" podID="9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" containerID="aefb1dba292711435eac1e73e9e529ab2e03a7b4863f808c55d3071e985aa85e" exitCode=0 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.596274 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.596808 4978 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.617236 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-config" (OuterVolumeSpecName: "config") pod "e78e5151-267c-4b33-9473-6b249b01f12f" (UID: "e78e5151-267c-4b33-9473-6b249b01f12f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699184 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" event={"ID":"e78e5151-267c-4b33-9473-6b249b01f12f","Type":"ContainerDied","Data":"4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699222 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c64474bcc-4rmct" event={"ID":"e78e5151-267c-4b33-9473-6b249b01f12f","Type":"ContainerDied","Data":"ad1c1aa647da09a023e8ffbe67cb46a2adb7b22fab5ea2529869b0f4806759e0"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699236 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5b4d65dff6-x8qrr"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699255 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-785ccc58d9-p75wv"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699266 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-wxrhg"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699277 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-wxrhg"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699289 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"58c891d9-fbca-406c-bd1a-ecb13f9d7444","Type":"ContainerDied","Data":"23b908a2d2c8c5b19239c4d15a74e513b855d7ec6155c2aab43b358150b36fe5"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699301 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwjnj" event={"ID":"b052d38e-931f-4eb1-85ea-142ef3da9881","Type":"ContainerStarted","Data":"cffdaf1ed686416ad7d1f7c33ca3c4b1c96e5970f3949c9b2071aaeffe295cbd"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699312 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699324 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"4dfed42feb8ac7a42412c46dcae347795ab03fb0af477b3ffbb5372f2f1e92c3"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699345 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-c75lr"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699355 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"c5c66777c2f36490e61ea071a8cde68af4d8aa4113f9db174c104aae89ea3baf"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699363 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"1ec7901b94aa973e76af694ac0b6f07b5c9788edfe7b888e4ae6f0ec6806b20b"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699396 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6ec4-account-create-update-649r6"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699408 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"6fefdf96f9ea31d2462085c35bca5e00e4f192d43b6b9dbe2f57601d1afde09c"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699419 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-c75lr"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699428 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"f892ef9c825b59066a0c023cd3b507cce63da08b00077af4f81ced0830c89f1c"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699437 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-xq9mn"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699445 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-swr2c"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699454 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"c247f762435a5bf5a26de2224e27277501e932169aad0cc00bd5b31c4c134647"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699463 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"36f71fa276b25e88398462e3bbc6d6b87e2bee8c2aade1e2f5f0ffdc61440473"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699474 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-xq9mn"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699484 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"c196bca824e71c59dc1d9939514fc9f5d009c35f1cb8b40a2d2b99b8d722f60b"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699493 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-9373-account-create-update-lqszb"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699505 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"25a3891e332bc2d905bcc5238aea824519156ba640e57709ba7a03f1986ab36b"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699516 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-swr2c"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699525 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699535 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699545 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-lhjvp"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699558 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699568 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"bb9960184024cbe585b528d58a3dd1a8b9fdeda2c6566bd7d76615473e70311e"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699576 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"fe36f33cfa027ca4f84b465a94aa8c1bbd4436f1982627dab087ee9bdbe66214"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699585 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7d28d9f8-de4f-4223-bfa6-fdf398f133bb","Type":"ContainerDied","Data":"251a79da9fba47a65d4db2f16a1ee10da3ad7f02ffcf859eb40da544a9209781"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699596 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ed9859d2-701a-4011-b9d6-d8cf6291265a","Type":"ContainerDied","Data":"a15f02424a1013ac6691c287f99c51cc3e6aff1c7146b986bb694436609fc027"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699606 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ed9859d2-701a-4011-b9d6-d8cf6291265a","Type":"ContainerDied","Data":"8baa6e062d410fe1594d058fc3924f5a1ed85894951c8f77a4918acff9d1e1bd"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699615 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8f7c6df-f16d-487d-98e8-cab2c89073b4","Type":"ContainerDied","Data":"8a5993b793c50295987a7b0055fc564f776c502b1b808c0832ceeaaacca5e945"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699625 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwmn" event={"ID":"34752acd-4270-408b-bc82-af9fc515da2b","Type":"ContainerStarted","Data":"0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699635 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69794fc6d6-zqnw6" event={"ID":"c6a16d36-c629-483e-9b15-fbd7b8913883","Type":"ContainerDied","Data":"9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699647 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6df7a672-b889-40d3-abe5-271e4d36dd85","Type":"ContainerDied","Data":"c227278c28e8d88bb7c9ddfa3eaa4e8de5003b5ef2d93630c76d010d8ca314e0"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699657 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6df7a672-b889-40d3-abe5-271e4d36dd85","Type":"ContainerDied","Data":"c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.699665 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r45vg" event={"ID":"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb","Type":"ContainerDied","Data":"aefb1dba292711435eac1e73e9e529ab2e03a7b4863f808c55d3071e985aa85e"} Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.700059 4978 scope.go:117] "RemoveContainer" containerID="4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.700202 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" podUID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" containerName="barbican-keystone-listener-log" containerID="cri-o://965888d2495c19353834efdc1dc9bee729c306c0310cc39f6260c432b9398f32" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.700355 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-785ccc58d9-p75wv" podUID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" containerName="proxy-httpd" containerID="cri-o://e0503e6f2f40da27125f61fc3830d5697085e44e0e23aa2a64cee749b3cf958e" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.701015 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" containerName="nova-api-log" containerID="cri-o://d48b870ae7f0075341f3f007177eb69704874efaa7faac91545002adf5db1de3" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.701132 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="185c1474-2c9a-408d-a3b1-8eb4820f6109" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://70fcc3b61d5c76d88b1ebc2b70685558693748a8e9f0734e5a69556982ffa4cb" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.701852 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" podUID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" containerName="barbican-keystone-listener" containerID="cri-o://7f25fb6fb4d79c681a829daa912270875a167f3eab8e635f499b9abf6ae1a887" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.701936 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-785ccc58d9-p75wv" podUID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" containerName="proxy-server" containerID="cri-o://d9e6c02d47ab88e1a99d79a917daf7a487b0b7f226c09ae65c011a4c5088f429" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.702014 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" containerName="nova-api-api" containerID="cri-o://ed748fe64533cd208cfaa954b8175f2c179485a3b36021fb250466d83a45c91e" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.702894 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78e5151-267c-4b33-9473-6b249b01f12f-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.702967 4978 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.702999 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data podName:43bd8325-9408-4a32-9285-d522d497205e nodeName:}" failed. No retries permitted until 2026-02-25 07:10:32.702988295 +0000 UTC m=+1526.142244754 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data") pod "rabbitmq-cell1-server-0" (UID: "43bd8325-9408-4a32-9285-d522d497205e") : configmap "rabbitmq-cell1-config-data" not found Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.724505 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4z765_a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682/openstack-network-exporter/0.log" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.724776 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.734009 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r45vg" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.742593 4978 scope.go:117] "RemoveContainer" containerID="772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.742591 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="185c1474-2c9a-408d-a3b1-8eb4820f6109" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.204:6080/vnc_lite.html\": dial tcp 10.217.0.204:6080: connect: connection refused" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.763472 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.765012 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bwjnj" podStartSLOduration=3.957328312 podStartE2EDuration="18.764996541s" podCreationTimestamp="2026-02-25 07:10:13 +0000 UTC" firstStartedPulling="2026-02-25 07:10:15.047617697 +0000 UTC m=+1508.486874156" lastFinishedPulling="2026-02-25 07:10:29.855285926 +0000 UTC m=+1523.294542385" observedRunningTime="2026-02-25 07:10:31.762576677 +0000 UTC m=+1525.201833136" watchObservedRunningTime="2026-02-25 07:10:31.764996541 +0000 UTC m=+1525.204253000" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.768915 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="43bd8325-9408-4a32-9285-d522d497205e" containerName="rabbitmq" containerID="cri-o://2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5" gracePeriod=604800 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.805588 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.805881 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="8a6b84e2-353d-47ef-82be-c7e182533a57" containerName="nova-cell1-conductor-conductor" containerID="cri-o://5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.807393 4978 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.807437 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data podName:09112202-a3ef-4742-bdbe-b46c1c9a55ba nodeName:}" failed. No retries permitted until 2026-02-25 07:10:33.807422725 +0000 UTC m=+1527.246679174 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data") pod "rabbitmq-server-0" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba") : configmap "rabbitmq-config-data" not found Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.811541 4978 scope.go:117] "RemoveContainer" containerID="4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.821786 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-b7zzg"] Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.823020 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706\": container with ID starting with 4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706 not found: ID does not exist" containerID="4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.823060 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706"} err="failed to get container status \"4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706\": rpc error: code = NotFound desc = could not find container \"4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706\": container with ID starting with 4150698ecdb1304a1961fdd09a6b3c72ddf09e37f5729f0413ffbfe3f8f51706 not found: ID does not exist" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.823088 4978 scope.go:117] "RemoveContainer" containerID="772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263" Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.825262 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263\": container with ID starting with 772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263 not found: ID does not exist" containerID="772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.825285 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263"} err="failed to get container status \"772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263\": rpc error: code = NotFound desc = could not find container \"772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263\": container with ID starting with 772a076e995ba1de564cade29d360bc49fc85b069be243ad164c4fd489e0f263 not found: ID does not exist" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.828062 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="09112202-a3ef-4742-bdbe-b46c1c9a55ba" containerName="rabbitmq" containerID="cri-o://0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33" gracePeriod=604800 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.834436 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-b7zzg"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.838154 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.838405 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="50c62987-be2b-4deb-a88c-107db5d00415" containerName="nova-cell0-conductor-conductor" containerID="cri-o://91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.843829 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tfnq8"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.874265 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.874523 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7456f008-69db-4a20-86c4-2ab52eb93aa7" containerName="nova-scheduler-scheduler" containerID="cri-o://b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.878587 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tfnq8"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.884582 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c64474bcc-4rmct"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.896889 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c64474bcc-4rmct"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.915892 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnbv5\" (UniqueName: \"kubernetes.io/projected/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-kube-api-access-jnbv5\") pod \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.915943 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-combined-ca-bundle\") pod \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.915962 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-ovn-controller-tls-certs\") pod \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.915982 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run-ovn\") pod \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916002 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config\") pod \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916023 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovs-rundir\") pod \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916061 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-config\") pod \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916092 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-scripts\") pod \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916138 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmr2q\" (UniqueName: \"kubernetes.io/projected/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-kube-api-access-pmr2q\") pod \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916155 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-combined-ca-bundle\") pod \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916169 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovn-rundir\") pod \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916197 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-combined-ca-bundle\") pod \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916224 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-log-ovn\") pod \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916244 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-metrics-certs-tls-certs\") pod \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\" (UID: \"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916276 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcnx7\" (UniqueName: \"kubernetes.io/projected/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-kube-api-access-xcnx7\") pod \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916296 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config-secret\") pod \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\" (UID: \"ecf946d2-e6a8-4df6-ab76-b53d03db5d27\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.916311 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run\") pod \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\" (UID: \"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb\") " Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.918933 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" (UID: "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.929926 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" (UID: "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.936564 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run" (OuterVolumeSpecName: "var-run") pod "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" (UID: "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.936643 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" (UID: "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.942528 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="c488aaac-88dd-4eb1-9020-e17749814ed9" containerName="galera" containerID="cri-o://df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79" gracePeriod=30 Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.942707 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" (UID: "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.945746 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-kube-api-access-pmr2q" (OuterVolumeSpecName: "kube-api-access-pmr2q") pod "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" (UID: "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb"). InnerVolumeSpecName "kube-api-access-pmr2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.946419 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-config" (OuterVolumeSpecName: "config") pod "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" (UID: "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.952815 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-scripts" (OuterVolumeSpecName: "scripts") pod "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" (UID: "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.966536 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-kube-api-access-xcnx7" (OuterVolumeSpecName: "kube-api-access-xcnx7") pod "ecf946d2-e6a8-4df6-ab76-b53d03db5d27" (UID: "ecf946d2-e6a8-4df6-ab76-b53d03db5d27"). InnerVolumeSpecName "kube-api-access-xcnx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.966537 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7 is running failed: container process not found" containerID="c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7" cmd=["/usr/bin/pidof","ovsdb-server"] Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.966670 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-kube-api-access-jnbv5" (OuterVolumeSpecName: "kube-api-access-jnbv5") pod "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" (UID: "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682"). InnerVolumeSpecName "kube-api-access-jnbv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: I0225 07:10:31.979032 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" (UID: "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.981546 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7 is running failed: container process not found" containerID="c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7" cmd=["/usr/bin/pidof","ovsdb-server"] Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.985013 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7 is running failed: container process not found" containerID="c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7" cmd=["/usr/bin/pidof","ovsdb-server"] Feb 25 07:10:31 crc kubenswrapper[4978]: E0225 07:10:31.985071 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerName="ovsdbserver-sb" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.009986 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" (UID: "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017455 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcnx7\" (UniqueName: \"kubernetes.io/projected/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-kube-api-access-xcnx7\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017480 4978 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017493 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnbv5\" (UniqueName: \"kubernetes.io/projected/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-kube-api-access-jnbv5\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017502 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017511 4978 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017519 4978 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovs-rundir\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017527 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017535 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017558 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmr2q\" (UniqueName: \"kubernetes.io/projected/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-kube-api-access-pmr2q\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017566 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017574 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-ovn-rundir\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.017582 4978 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.054870 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ecf946d2-e6a8-4df6-ab76-b53d03db5d27" (UID: "ecf946d2-e6a8-4df6-ab76-b53d03db5d27"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.081086 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6df7a672-b889-40d3-abe5-271e4d36dd85/ovsdbserver-sb/0.log" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.081160 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.112093 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecf946d2-e6a8-4df6-ab76-b53d03db5d27" (UID: "ecf946d2-e6a8-4df6-ab76-b53d03db5d27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.118841 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdpqt\" (UniqueName: \"kubernetes.io/projected/6df7a672-b889-40d3-abe5-271e4d36dd85-kube-api-access-fdpqt\") pod \"6df7a672-b889-40d3-abe5-271e4d36dd85\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.119504 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.119530 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.159011 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6df7a672-b889-40d3-abe5-271e4d36dd85-kube-api-access-fdpqt" (OuterVolumeSpecName: "kube-api-access-fdpqt") pod "6df7a672-b889-40d3-abe5-271e4d36dd85" (UID: "6df7a672-b889-40d3-abe5-271e4d36dd85"). InnerVolumeSpecName "kube-api-access-fdpqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.177022 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ecf946d2-e6a8-4df6-ab76-b53d03db5d27" (UID: "ecf946d2-e6a8-4df6-ab76-b53d03db5d27"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.227281 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" (UID: "a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.230806 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdb-rundir\") pod \"6df7a672-b889-40d3-abe5-271e4d36dd85\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.230855 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"6df7a672-b889-40d3-abe5-271e4d36dd85\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.230925 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-config\") pod \"6df7a672-b889-40d3-abe5-271e4d36dd85\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.230942 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-combined-ca-bundle\") pod \"6df7a672-b889-40d3-abe5-271e4d36dd85\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.230978 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-scripts\") pod \"6df7a672-b889-40d3-abe5-271e4d36dd85\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.230996 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-metrics-certs-tls-certs\") pod \"6df7a672-b889-40d3-abe5-271e4d36dd85\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.231108 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdbserver-sb-tls-certs\") pod \"6df7a672-b889-40d3-abe5-271e4d36dd85\" (UID: \"6df7a672-b889-40d3-abe5-271e4d36dd85\") " Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.231541 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ecf946d2-e6a8-4df6-ab76-b53d03db5d27-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.231554 4978 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.231563 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdpqt\" (UniqueName: \"kubernetes.io/projected/6df7a672-b889-40d3-abe5-271e4d36dd85-kube-api-access-fdpqt\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.234278 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-config" (OuterVolumeSpecName: "config") pod "6df7a672-b889-40d3-abe5-271e4d36dd85" (UID: "6df7a672-b889-40d3-abe5-271e4d36dd85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.234453 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "6df7a672-b889-40d3-abe5-271e4d36dd85" (UID: "6df7a672-b889-40d3-abe5-271e4d36dd85"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.234712 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-scripts" (OuterVolumeSpecName: "scripts") pod "6df7a672-b889-40d3-abe5-271e4d36dd85" (UID: "6df7a672-b889-40d3-abe5-271e4d36dd85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.242288 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "6df7a672-b889-40d3-abe5-271e4d36dd85" (UID: "6df7a672-b889-40d3-abe5-271e4d36dd85"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.290228 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" (UID: "9e95552a-ba97-44a7-9ce6-9fef9a84ecfb"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.303495 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-lhjvp"] Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.310586 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-cb34-account-create-update-2blwx"] Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.334603 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.334658 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.334670 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.334680 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6df7a672-b889-40d3-abe5-271e4d36dd85-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.334690 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.352528 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.372533 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.376548 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6df7a672-b889-40d3-abe5-271e4d36dd85" (UID: "6df7a672-b889-40d3-abe5-271e4d36dd85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.376966 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.383712 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.383764 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="8a6b84e2-353d-47ef-82be-c7e182533a57" containerName="nova-cell1-conductor-conductor" Feb 25 07:10:32 crc kubenswrapper[4978]: W0225 07:10:32.391678 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97e7baca_befe_4eba_9652_5474e0e88899.slice/crio-a51fee4a5ce94a963831a2282e192c0b87e3268ae1de07faec23d7dda17e50ba WatchSource:0}: Error finding container a51fee4a5ce94a963831a2282e192c0b87e3268ae1de07faec23d7dda17e50ba: Status 404 returned error can't find the container with id a51fee4a5ce94a963831a2282e192c0b87e3268ae1de07faec23d7dda17e50ba Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.409940 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "6df7a672-b889-40d3-abe5-271e4d36dd85" (UID: "6df7a672-b889-40d3-abe5-271e4d36dd85"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.420488 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "6df7a672-b889-40d3-abe5-271e4d36dd85" (UID: "6df7a672-b889-40d3-abe5-271e4d36dd85"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.435892 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.435921 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.435933 4978 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.435942 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df7a672-b889-40d3-abe5-271e4d36dd85-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.451063 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 07:10:32 crc kubenswrapper[4978]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658,Command:[/bin/sh -c #!/bin/bash Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: if [ -n "" ]; then Feb 25 07:10:32 crc kubenswrapper[4978]: GRANT_DATABASE="" Feb 25 07:10:32 crc kubenswrapper[4978]: else Feb 25 07:10:32 crc kubenswrapper[4978]: GRANT_DATABASE="*" Feb 25 07:10:32 crc kubenswrapper[4978]: fi Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: # going for maximum compatibility here: Feb 25 07:10:32 crc kubenswrapper[4978]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 25 07:10:32 crc kubenswrapper[4978]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 25 07:10:32 crc kubenswrapper[4978]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 25 07:10:32 crc kubenswrapper[4978]: # support updates Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: $MYSQL_CMD < logger="UnhandledError" Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.452289 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 07:10:32 crc kubenswrapper[4978]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658,Command:[/bin/sh -c #!/bin/bash Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: if [ -n "placement" ]; then Feb 25 07:10:32 crc kubenswrapper[4978]: GRANT_DATABASE="placement" Feb 25 07:10:32 crc kubenswrapper[4978]: else Feb 25 07:10:32 crc kubenswrapper[4978]: GRANT_DATABASE="*" Feb 25 07:10:32 crc kubenswrapper[4978]: fi Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: # going for maximum compatibility here: Feb 25 07:10:32 crc kubenswrapper[4978]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 25 07:10:32 crc kubenswrapper[4978]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 25 07:10:32 crc kubenswrapper[4978]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 25 07:10:32 crc kubenswrapper[4978]: # support updates Feb 25 07:10:32 crc kubenswrapper[4978]: Feb 25 07:10:32 crc kubenswrapper[4978]: $MYSQL_CMD < logger="UnhandledError" Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.452288 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"openstack-cell1-mariadb-root-db-secret\\\" not found\"" pod="openstack/root-account-create-update-lhjvp" podUID="97e7baca-befe-4eba-9652-5474e0e88899" Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.454401 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"placement-db-secret\\\" not found\"" pod="openstack/placement-cb34-account-create-update-2blwx" podUID="7cf4ce11-507f-4be4-9652-05035315d26d" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.568622 4978 generic.go:334] "Generic (PLEG): container finished" podID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" containerID="d9e6c02d47ab88e1a99d79a917daf7a487b0b7f226c09ae65c011a4c5088f429" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.568653 4978 generic.go:334] "Generic (PLEG): container finished" podID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" containerID="e0503e6f2f40da27125f61fc3830d5697085e44e0e23aa2a64cee749b3cf958e" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.568722 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-785ccc58d9-p75wv" event={"ID":"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0","Type":"ContainerDied","Data":"d9e6c02d47ab88e1a99d79a917daf7a487b0b7f226c09ae65c011a4c5088f429"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.568749 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-785ccc58d9-p75wv" event={"ID":"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0","Type":"ContainerDied","Data":"e0503e6f2f40da27125f61fc3830d5697085e44e0e23aa2a64cee749b3cf958e"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.573877 4978 generic.go:334] "Generic (PLEG): container finished" podID="34752acd-4270-408b-bc82-af9fc515da2b" containerID="0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.573950 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwmn" event={"ID":"34752acd-4270-408b-bc82-af9fc515da2b","Type":"ContainerDied","Data":"0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.606317 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="d908490f57d2f2a43ad75a1a8576681b6638c3b6cc2217e47bfbe190b3f31f3b" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.606349 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="cfdc985f7b5bb332d961a90b34a12545a7880322a9b3eaa992362b444ccba073" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.606356 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="2c1a63624536475bb718a04a658404433f192ece6f6682b1e5ca98142ac9eb79" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.606484 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"d908490f57d2f2a43ad75a1a8576681b6638c3b6cc2217e47bfbe190b3f31f3b"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.606511 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"cfdc985f7b5bb332d961a90b34a12545a7880322a9b3eaa992362b444ccba073"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.606530 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"2c1a63624536475bb718a04a658404433f192ece6f6682b1e5ca98142ac9eb79"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.609741 4978 generic.go:334] "Generic (PLEG): container finished" podID="d1ba0ae6-d125-48d2-a866-615d927dc525" containerID="6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.609797 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cdb588b97-7jkd8" event={"ID":"d1ba0ae6-d125-48d2-a866-615d927dc525","Type":"ContainerDied","Data":"6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.617107 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6df7a672-b889-40d3-abe5-271e4d36dd85/ovsdbserver-sb/0.log" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.617176 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6df7a672-b889-40d3-abe5-271e4d36dd85","Type":"ContainerDied","Data":"944d35b451cb47c9e602257957eeb480054d54b2e1346c5e1d27f74e82490a9c"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.617210 4978 scope.go:117] "RemoveContainer" containerID="c227278c28e8d88bb7c9ddfa3eaa4e8de5003b5ef2d93630c76d010d8ca314e0" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.617299 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.683278 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4z765_a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682/openstack-network-exporter/0.log" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.683813 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4z765" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.683797 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4z765" event={"ID":"a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682","Type":"ContainerDied","Data":"440ba6ed4d8946c310454eaecfb11414cacb380bf48e6d2fa2c5424f41b036e0"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.687342 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-lhjvp" event={"ID":"97e7baca-befe-4eba-9652-5474e0e88899","Type":"ContainerStarted","Data":"a51fee4a5ce94a963831a2282e192c0b87e3268ae1de07faec23d7dda17e50ba"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.689476 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.694501 4978 scope.go:117] "RemoveContainer" containerID="c9a1cd07339c94d03e839038bc0e07d5053222eb296ab4e420625a325c4f08c7" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.695899 4978 generic.go:334] "Generic (PLEG): container finished" podID="01eac805-ca08-46de-b79b-0ca5645f6955" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.695993 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-njndh" event={"ID":"01eac805-ca08-46de-b79b-0ca5645f6955","Type":"ContainerDied","Data":"f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.701541 4978 generic.go:334] "Generic (PLEG): container finished" podID="185c1474-2c9a-408d-a3b1-8eb4820f6109" containerID="70fcc3b61d5c76d88b1ebc2b70685558693748a8e9f0734e5a69556982ffa4cb" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.701610 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"185c1474-2c9a-408d-a3b1-8eb4820f6109","Type":"ContainerDied","Data":"70fcc3b61d5c76d88b1ebc2b70685558693748a8e9f0734e5a69556982ffa4cb"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.706419 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.728408 4978 generic.go:334] "Generic (PLEG): container finished" podID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" containerID="7f25fb6fb4d79c681a829daa912270875a167f3eab8e635f499b9abf6ae1a887" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.728440 4978 generic.go:334] "Generic (PLEG): container finished" podID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" containerID="965888d2495c19353834efdc1dc9bee729c306c0310cc39f6260c432b9398f32" exitCode=143 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.728538 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" event={"ID":"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a","Type":"ContainerDied","Data":"7f25fb6fb4d79c681a829daa912270875a167f3eab8e635f499b9abf6ae1a887"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.728603 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" event={"ID":"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a","Type":"ContainerDied","Data":"965888d2495c19353834efdc1dc9bee729c306c0310cc39f6260c432b9398f32"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.732788 4978 generic.go:334] "Generic (PLEG): container finished" podID="a762aa77-9eb1-42eb-98e5-016cc3acac10" containerID="97d4b20d6d6a152207b32f06fed0f8b590209ec24594d4831ee695bb8b21c4ee" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.732836 4978 generic.go:334] "Generic (PLEG): container finished" podID="a762aa77-9eb1-42eb-98e5-016cc3acac10" containerID="47894865746e5efcf05b88359453cdd68cf29193d3fc760e152761c140f46884" exitCode=143 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.732917 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7696768c55-mnqh9" event={"ID":"a762aa77-9eb1-42eb-98e5-016cc3acac10","Type":"ContainerDied","Data":"97d4b20d6d6a152207b32f06fed0f8b590209ec24594d4831ee695bb8b21c4ee"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.733003 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7696768c55-mnqh9" event={"ID":"a762aa77-9eb1-42eb-98e5-016cc3acac10","Type":"ContainerDied","Data":"47894865746e5efcf05b88359453cdd68cf29193d3fc760e152761c140f46884"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.734773 4978 generic.go:334] "Generic (PLEG): container finished" podID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerID="2881499eed920adb23f199015c996a12e05fc38be2beb2a1bf0cc87bbbb6b34a" exitCode=143 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.734839 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8","Type":"ContainerDied","Data":"2881499eed920adb23f199015c996a12e05fc38be2beb2a1bf0cc87bbbb6b34a"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.735811 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.738229 4978 generic.go:334] "Generic (PLEG): container finished" podID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerID="3d765602f0864d6d998940f94e6b40eaa60dbf1a6c6bdb9427a47ca7993400a3" exitCode=143 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.738747 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77bc5d768-qx88q" event={"ID":"3a2631ab-861d-40eb-9b52-b4fd269eb411","Type":"ContainerDied","Data":"3d765602f0864d6d998940f94e6b40eaa60dbf1a6c6bdb9427a47ca7993400a3"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.748249 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r45vg" event={"ID":"9e95552a-ba97-44a7-9ce6-9fef9a84ecfb","Type":"ContainerDied","Data":"df64dd81561667b23fde5ead3623cad904eb563e51098c9a81b12150ce28c753"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.748699 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r45vg" Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.749244 4978 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.749864 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data podName:43bd8325-9408-4a32-9285-d522d497205e nodeName:}" failed. No retries permitted until 2026-02-25 07:10:34.74984335 +0000 UTC m=+1528.189099809 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data") pod "rabbitmq-cell1-server-0" (UID: "43bd8325-9408-4a32-9285-d522d497205e") : configmap "rabbitmq-cell1-config-data" not found Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.760096 4978 generic.go:334] "Generic (PLEG): container finished" podID="2e39fd28-0520-44bc-9486-a118cec688c0" containerID="d48b870ae7f0075341f3f007177eb69704874efaa7faac91545002adf5db1de3" exitCode=143 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.760145 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e39fd28-0520-44bc-9486-a118cec688c0","Type":"ContainerDied","Data":"d48b870ae7f0075341f3f007177eb69704874efaa7faac91545002adf5db1de3"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.770695 4978 generic.go:334] "Generic (PLEG): container finished" podID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" containerID="abc550e17c1b943bad86597853b93ec464de90bcf8090229b1e9cb5d9aeca0d7" exitCode=0 Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.770753 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8f7c6df-f16d-487d-98e8-cab2c89073b4","Type":"ContainerDied","Data":"abc550e17c1b943bad86597853b93ec464de90bcf8090229b1e9cb5d9aeca0d7"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.796597 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cb34-account-create-update-2blwx" event={"ID":"7cf4ce11-507f-4be4-9652-05035315d26d","Type":"ContainerStarted","Data":"32c1928875454caef6ca3ab7aafd7711a1600bc85e947976a39e09864729b987"} Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.948813 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.963535 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-4z765"] Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.969477 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.969634 4978 scope.go:117] "RemoveContainer" containerID="986d5b46fb6a179bd4f7544f78f8396ed0327bcfeea74164711b40b9816145da" Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.969742 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.978120 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-4z765"] Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.978819 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.979054 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.980392 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.980439 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.985052 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:32 crc kubenswrapper[4978]: E0225 07:10:32.985078 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovs-vswitchd" Feb 25 07:10:32 crc kubenswrapper[4978]: I0225 07:10:32.989699 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-r45vg"] Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.003604 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ed9859d2-701a-4011-b9d6-d8cf6291265a/ovsdbserver-nb/0.log" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.003930 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.003726 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-r45vg"] Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.005412 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.019680 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.038744 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.043968 4978 scope.go:117] "RemoveContainer" containerID="b043db8fffc821daf3cf627f0556f152620f9dbbc236549845734482a005f759" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.057485 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-combined-ca-bundle\") pod \"ed9859d2-701a-4011-b9d6-d8cf6291265a\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.057661 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data-custom\") pod \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.057807 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8f7c6df-f16d-487d-98e8-cab2c89073b4-etc-machine-id\") pod \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.057925 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-combined-ca-bundle\") pod \"185c1474-2c9a-408d-a3b1-8eb4820f6109\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.057990 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2fhp\" (UniqueName: \"kubernetes.io/projected/f8f7c6df-f16d-487d-98e8-cab2c89073b4-kube-api-access-m2fhp\") pod \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058064 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdb-rundir\") pod \"ed9859d2-701a-4011-b9d6-d8cf6291265a\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058124 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-metrics-certs-tls-certs\") pod \"ed9859d2-701a-4011-b9d6-d8cf6291265a\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058199 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-logs\") pod \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058305 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-scripts\") pod \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058382 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-vencrypt-tls-certs\") pod \"185c1474-2c9a-408d-a3b1-8eb4820f6109\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058666 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-scripts\") pod \"ed9859d2-701a-4011-b9d6-d8cf6291265a\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058728 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ed9859d2-701a-4011-b9d6-d8cf6291265a\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058796 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-config-data\") pod \"185c1474-2c9a-408d-a3b1-8eb4820f6109\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058879 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfs5p\" (UniqueName: \"kubernetes.io/projected/185c1474-2c9a-408d-a3b1-8eb4820f6109-kube-api-access-zfs5p\") pod \"185c1474-2c9a-408d-a3b1-8eb4820f6109\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058943 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-nova-novncproxy-tls-certs\") pod \"185c1474-2c9a-408d-a3b1-8eb4820f6109\" (UID: \"185c1474-2c9a-408d-a3b1-8eb4820f6109\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.061047 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data-custom\") pod \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.061134 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data\") pod \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.061200 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdbserver-nb-tls-certs\") pod \"ed9859d2-701a-4011-b9d6-d8cf6291265a\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.061287 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-combined-ca-bundle\") pod \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.061348 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data\") pod \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.061449 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npq46\" (UniqueName: \"kubernetes.io/projected/ed9859d2-701a-4011-b9d6-d8cf6291265a-kube-api-access-npq46\") pod \"ed9859d2-701a-4011-b9d6-d8cf6291265a\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.061539 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6xjf\" (UniqueName: \"kubernetes.io/projected/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-kube-api-access-v6xjf\") pod \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\" (UID: \"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.061643 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-combined-ca-bundle\") pod \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\" (UID: \"f8f7c6df-f16d-487d-98e8-cab2c89073b4\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.061717 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-config\") pod \"ed9859d2-701a-4011-b9d6-d8cf6291265a\" (UID: \"ed9859d2-701a-4011-b9d6-d8cf6291265a\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.058067 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8f7c6df-f16d-487d-98e8-cab2c89073b4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f8f7c6df-f16d-487d-98e8-cab2c89073b4" (UID: "f8f7c6df-f16d-487d-98e8-cab2c89073b4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.067024 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-logs" (OuterVolumeSpecName: "logs") pod "d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" (UID: "d962a2e0-007e-46ca-9ad0-0ecf85a5e87a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.070089 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "ed9859d2-701a-4011-b9d6-d8cf6291265a" (UID: "ed9859d2-701a-4011-b9d6-d8cf6291265a"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.077724 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-config" (OuterVolumeSpecName: "config") pod "ed9859d2-701a-4011-b9d6-d8cf6291265a" (UID: "ed9859d2-701a-4011-b9d6-d8cf6291265a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.080209 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.080317 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.080434 4978 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8f7c6df-f16d-487d-98e8-cab2c89073b4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.080495 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.080605 4978 secret.go:188] Couldn't get secret openstack/barbican-config-data: secret "barbican-config-data" not found Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.080761 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data podName:a762aa77-9eb1-42eb-98e5-016cc3acac10 nodeName:}" failed. No retries permitted until 2026-02-25 07:10:37.0807474 +0000 UTC m=+1530.520003859 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data") pod "barbican-worker-7696768c55-mnqh9" (UID: "a762aa77-9eb1-42eb-98e5-016cc3acac10") : secret "barbican-config-data" not found Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.087575 4978 scope.go:117] "RemoveContainer" containerID="aefb1dba292711435eac1e73e9e529ab2e03a7b4863f808c55d3071e985aa85e" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.089567 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "ed9859d2-701a-4011-b9d6-d8cf6291265a" (UID: "ed9859d2-701a-4011-b9d6-d8cf6291265a"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.103711 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-scripts" (OuterVolumeSpecName: "scripts") pod "ed9859d2-701a-4011-b9d6-d8cf6291265a" (UID: "ed9859d2-701a-4011-b9d6-d8cf6291265a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.106656 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.108576 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.109491 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.110770 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.110804 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7456f008-69db-4a20-86c4-2ab52eb93aa7" containerName="nova-scheduler-scheduler" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.137505 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f8f7c6df-f16d-487d-98e8-cab2c89073b4" (UID: "f8f7c6df-f16d-487d-98e8-cab2c89073b4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.137815 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-scripts" (OuterVolumeSpecName: "scripts") pod "f8f7c6df-f16d-487d-98e8-cab2c89073b4" (UID: "f8f7c6df-f16d-487d-98e8-cab2c89073b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.138032 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185c1474-2c9a-408d-a3b1-8eb4820f6109-kube-api-access-zfs5p" (OuterVolumeSpecName: "kube-api-access-zfs5p") pod "185c1474-2c9a-408d-a3b1-8eb4820f6109" (UID: "185c1474-2c9a-408d-a3b1-8eb4820f6109"). InnerVolumeSpecName "kube-api-access-zfs5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.139669 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f7c6df-f16d-487d-98e8-cab2c89073b4-kube-api-access-m2fhp" (OuterVolumeSpecName: "kube-api-access-m2fhp") pod "f8f7c6df-f16d-487d-98e8-cab2c89073b4" (UID: "f8f7c6df-f16d-487d-98e8-cab2c89073b4"). InnerVolumeSpecName "kube-api-access-m2fhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.142879 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6ec4-account-create-update-649r6"] Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.157926 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-1b7b-account-create-update-qlxzw"] Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.163325 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed9859d2-701a-4011-b9d6-d8cf6291265a-kube-api-access-npq46" (OuterVolumeSpecName: "kube-api-access-npq46") pod "ed9859d2-701a-4011-b9d6-d8cf6291265a" (UID: "ed9859d2-701a-4011-b9d6-d8cf6291265a"). InnerVolumeSpecName "kube-api-access-npq46". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.167061 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" (UID: "d962a2e0-007e-46ca-9ad0-0ecf85a5e87a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.169616 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-kube-api-access-v6xjf" (OuterVolumeSpecName: "kube-api-access-v6xjf") pod "d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" (UID: "d962a2e0-007e-46ca-9ad0-0ecf85a5e87a"). InnerVolumeSpecName "kube-api-access-v6xjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.184296 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-config-data\") pod \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.184398 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd789\" (UniqueName: \"kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-kube-api-access-hd789\") pod \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.184422 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-internal-tls-certs\") pod \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.184444 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-run-httpd\") pod \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.184484 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data\") pod \"a762aa77-9eb1-42eb-98e5-016cc3acac10\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.184558 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data-custom\") pod \"a762aa77-9eb1-42eb-98e5-016cc3acac10\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.184584 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-combined-ca-bundle\") pod \"a762aa77-9eb1-42eb-98e5-016cc3acac10\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.184615 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a762aa77-9eb1-42eb-98e5-016cc3acac10-logs\") pod \"a762aa77-9eb1-42eb-98e5-016cc3acac10\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.184642 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-public-tls-certs\") pod \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.186674 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" (UID: "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.187953 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 07:10:33 crc kubenswrapper[4978]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658,Command:[/bin/sh -c #!/bin/bash Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: if [ -n "neutron" ]; then Feb 25 07:10:33 crc kubenswrapper[4978]: GRANT_DATABASE="neutron" Feb 25 07:10:33 crc kubenswrapper[4978]: else Feb 25 07:10:33 crc kubenswrapper[4978]: GRANT_DATABASE="*" Feb 25 07:10:33 crc kubenswrapper[4978]: fi Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: # going for maximum compatibility here: Feb 25 07:10:33 crc kubenswrapper[4978]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 25 07:10:33 crc kubenswrapper[4978]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 25 07:10:33 crc kubenswrapper[4978]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 25 07:10:33 crc kubenswrapper[4978]: # support updates Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: $MYSQL_CMD < logger="UnhandledError" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.191631 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-kube-api-access-hd789" (OuterVolumeSpecName: "kube-api-access-hd789") pod "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" (UID: "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0"). InnerVolumeSpecName "kube-api-access-hd789". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192039 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-etc-swift\") pod \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192099 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-log-httpd\") pod \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192116 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-combined-ca-bundle\") pod \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\" (UID: \"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192155 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltqc5\" (UniqueName: \"kubernetes.io/projected/a762aa77-9eb1-42eb-98e5-016cc3acac10-kube-api-access-ltqc5\") pod \"a762aa77-9eb1-42eb-98e5-016cc3acac10\" (UID: \"a762aa77-9eb1-42eb-98e5-016cc3acac10\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192902 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192932 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd789\" (UniqueName: \"kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-kube-api-access-hd789\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192943 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192952 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2fhp\" (UniqueName: \"kubernetes.io/projected/f8f7c6df-f16d-487d-98e8-cab2c89073b4-kube-api-access-m2fhp\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192961 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192970 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ed9859d2-701a-4011-b9d6-d8cf6291265a-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192988 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.192998 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfs5p\" (UniqueName: \"kubernetes.io/projected/185c1474-2c9a-408d-a3b1-8eb4820f6109-kube-api-access-zfs5p\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.193007 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.193015 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npq46\" (UniqueName: \"kubernetes.io/projected/ed9859d2-701a-4011-b9d6-d8cf6291265a-kube-api-access-npq46\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.193024 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6xjf\" (UniqueName: \"kubernetes.io/projected/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-kube-api-access-v6xjf\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.195259 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a762aa77-9eb1-42eb-98e5-016cc3acac10-logs" (OuterVolumeSpecName: "logs") pod "a762aa77-9eb1-42eb-98e5-016cc3acac10" (UID: "a762aa77-9eb1-42eb-98e5-016cc3acac10"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.198435 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a762aa77-9eb1-42eb-98e5-016cc3acac10" (UID: "a762aa77-9eb1-42eb-98e5-016cc3acac10"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.198753 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" (UID: "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.199062 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"neutron-db-secret\\\" not found\"" pod="openstack/neutron-6ec4-account-create-update-649r6" podUID="84e32f1f-5ad1-40a8-8f61-21025d84d003" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.202782 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" (UID: "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.205355 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lhjvp" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.206028 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8a8c-account-create-update-d7wm6"] Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.244397 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a762aa77-9eb1-42eb-98e5-016cc3acac10-kube-api-access-ltqc5" (OuterVolumeSpecName: "kube-api-access-ltqc5") pod "a762aa77-9eb1-42eb-98e5-016cc3acac10" (UID: "a762aa77-9eb1-42eb-98e5-016cc3acac10"). InnerVolumeSpecName "kube-api-access-ltqc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.246930 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 07:10:33 crc kubenswrapper[4978]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658,Command:[/bin/sh -c #!/bin/bash Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: if [ -n "glance" ]; then Feb 25 07:10:33 crc kubenswrapper[4978]: GRANT_DATABASE="glance" Feb 25 07:10:33 crc kubenswrapper[4978]: else Feb 25 07:10:33 crc kubenswrapper[4978]: GRANT_DATABASE="*" Feb 25 07:10:33 crc kubenswrapper[4978]: fi Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: # going for maximum compatibility here: Feb 25 07:10:33 crc kubenswrapper[4978]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 25 07:10:33 crc kubenswrapper[4978]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 25 07:10:33 crc kubenswrapper[4978]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 25 07:10:33 crc kubenswrapper[4978]: # support updates Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: $MYSQL_CMD < logger="UnhandledError" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.248795 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"glance-db-secret\\\" not found\"" pod="openstack/glance-1b7b-account-create-update-qlxzw" podUID="bc147cb6-a5d8-47f7-8ce6-168455f18512" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.257120 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 07:10:33 crc kubenswrapper[4978]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658,Command:[/bin/sh -c #!/bin/bash Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: if [ -n "barbican" ]; then Feb 25 07:10:33 crc kubenswrapper[4978]: GRANT_DATABASE="barbican" Feb 25 07:10:33 crc kubenswrapper[4978]: else Feb 25 07:10:33 crc kubenswrapper[4978]: GRANT_DATABASE="*" Feb 25 07:10:33 crc kubenswrapper[4978]: fi Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: # going for maximum compatibility here: Feb 25 07:10:33 crc kubenswrapper[4978]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 25 07:10:33 crc kubenswrapper[4978]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 25 07:10:33 crc kubenswrapper[4978]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 25 07:10:33 crc kubenswrapper[4978]: # support updates Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: $MYSQL_CMD < logger="UnhandledError" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.258257 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"barbican-db-secret\\\" not found\"" pod="openstack/barbican-8a8c-account-create-update-d7wm6" podUID="83bd2445-af69-432a-bc55-d5cee299162d" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.258418 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.259945 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="807eb4fa-5af0-450e-be50-f30c96731b46" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.169:8776/healthcheck\": read tcp 10.217.0.2:55374->10.217.0.169:8776: read: connection reset by peer" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.265061 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed9859d2-701a-4011-b9d6-d8cf6291265a" (UID: "ed9859d2-701a-4011-b9d6-d8cf6291265a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.273135 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-config-data" (OuterVolumeSpecName: "config-data") pod "185c1474-2c9a-408d-a3b1-8eb4820f6109" (UID: "185c1474-2c9a-408d-a3b1-8eb4820f6109"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.292526 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" (UID: "d962a2e0-007e-46ca-9ad0-0ecf85a5e87a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294080 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97e7baca-befe-4eba-9652-5474e0e88899-operator-scripts\") pod \"97e7baca-befe-4eba-9652-5474e0e88899\" (UID: \"97e7baca-befe-4eba-9652-5474e0e88899\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294315 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9vv7\" (UniqueName: \"kubernetes.io/projected/97e7baca-befe-4eba-9652-5474e0e88899-kube-api-access-b9vv7\") pod \"97e7baca-befe-4eba-9652-5474e0e88899\" (UID: \"97e7baca-befe-4eba-9652-5474e0e88899\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294493 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97e7baca-befe-4eba-9652-5474e0e88899-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "97e7baca-befe-4eba-9652-5474e0e88899" (UID: "97e7baca-befe-4eba-9652-5474e0e88899"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294675 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltqc5\" (UniqueName: \"kubernetes.io/projected/a762aa77-9eb1-42eb-98e5-016cc3acac10-kube-api-access-ltqc5\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294686 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294695 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97e7baca-befe-4eba-9652-5474e0e88899-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294704 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294714 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294724 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294732 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a762aa77-9eb1-42eb-98e5-016cc3acac10-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294741 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294749 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.294757 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.297016 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "185c1474-2c9a-408d-a3b1-8eb4820f6109" (UID: "185c1474-2c9a-408d-a3b1-8eb4820f6109"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.299952 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97e7baca-befe-4eba-9652-5474e0e88899-kube-api-access-b9vv7" (OuterVolumeSpecName: "kube-api-access-b9vv7") pod "97e7baca-befe-4eba-9652-5474e0e88899" (UID: "97e7baca-befe-4eba-9652-5474e0e88899"). InnerVolumeSpecName "kube-api-access-b9vv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.324057 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" (UID: "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.335724 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "185c1474-2c9a-408d-a3b1-8eb4820f6109" (UID: "185c1474-2c9a-408d-a3b1-8eb4820f6109"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.337672 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c4ce63-060f-4f53-8e1a-a70b6ba97255" path="/var/lib/kubelet/pods/12c4ce63-060f-4f53-8e1a-a70b6ba97255/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.338144 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1878ccac-42b1-4199-9008-eca6dcad41f6" path="/var/lib/kubelet/pods/1878ccac-42b1-4199-9008-eca6dcad41f6/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.338638 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5791f28c-d7ba-4aa4-99d5-319ecbd6bc45" path="/var/lib/kubelet/pods/5791f28c-d7ba-4aa4-99d5-319ecbd6bc45/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.339631 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6df7a672-b889-40d3-abe5-271e4d36dd85" path="/var/lib/kubelet/pods/6df7a672-b889-40d3-abe5-271e4d36dd85/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.340170 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" path="/var/lib/kubelet/pods/9e95552a-ba97-44a7-9ce6-9fef9a84ecfb/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.341022 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" path="/var/lib/kubelet/pods/a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.341150 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "185c1474-2c9a-408d-a3b1-8eb4820f6109" (UID: "185c1474-2c9a-408d-a3b1-8eb4820f6109"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.342045 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2885313-ac3c-44af-9fb9-941071137b56" path="/var/lib/kubelet/pods/b2885313-ac3c-44af-9fb9-941071137b56/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.342544 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be1ecfff-67b5-4cc5-90f0-d833276a5b62" path="/var/lib/kubelet/pods/be1ecfff-67b5-4cc5-90f0-d833276a5b62/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.343093 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78e5151-267c-4b33-9473-6b249b01f12f" path="/var/lib/kubelet/pods/e78e5151-267c-4b33-9473-6b249b01f12f/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.344432 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecf946d2-e6a8-4df6-ab76-b53d03db5d27" path="/var/lib/kubelet/pods/ecf946d2-e6a8-4df6-ab76-b53d03db5d27/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.344917 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5a1b5fa-aba9-4232-b5b7-96365cebef1a" path="/var/lib/kubelet/pods/f5a1b5fa-aba9-4232-b5b7-96365cebef1a/volumes" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.370621 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-config-data" (OuterVolumeSpecName: "config-data") pod "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" (UID: "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.394260 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a762aa77-9eb1-42eb-98e5-016cc3acac10" (UID: "a762aa77-9eb1-42eb-98e5-016cc3acac10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.396696 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.396723 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9vv7\" (UniqueName: \"kubernetes.io/projected/97e7baca-befe-4eba-9652-5474e0e88899-kube-api-access-b9vv7\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.396735 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.396744 4978 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.396753 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.396761 4978 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/185c1474-2c9a-408d-a3b1-8eb4820f6109-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.396769 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.411557 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" (UID: "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.411593 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" (UID: "f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.411648 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8f7c6df-f16d-487d-98e8-cab2c89073b4" (UID: "f8f7c6df-f16d-487d-98e8-cab2c89073b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.411640 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "ed9859d2-701a-4011-b9d6-d8cf6291265a" (UID: "ed9859d2-701a-4011-b9d6-d8cf6291265a"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.418042 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data" (OuterVolumeSpecName: "config-data") pod "a762aa77-9eb1-42eb-98e5-016cc3acac10" (UID: "a762aa77-9eb1-42eb-98e5-016cc3acac10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.421453 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "ed9859d2-701a-4011-b9d6-d8cf6291265a" (UID: "ed9859d2-701a-4011-b9d6-d8cf6291265a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.433596 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data" (OuterVolumeSpecName: "config-data") pod "d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" (UID: "d962a2e0-007e-46ca-9ad0-0ecf85a5e87a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.485650 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data" (OuterVolumeSpecName: "config-data") pod "f8f7c6df-f16d-487d-98e8-cab2c89073b4" (UID: "f8f7c6df-f16d-487d-98e8-cab2c89073b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.498614 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.498642 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.498651 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.498660 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.498671 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8f7c6df-f16d-487d-98e8-cab2c89073b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.498679 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.498687 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a762aa77-9eb1-42eb-98e5-016cc3acac10-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.498695 4978 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed9859d2-701a-4011-b9d6-d8cf6291265a-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.588658 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cb34-account-create-update-2blwx" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.701260 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s8d7\" (UniqueName: \"kubernetes.io/projected/7cf4ce11-507f-4be4-9652-05035315d26d-kube-api-access-9s8d7\") pod \"7cf4ce11-507f-4be4-9652-05035315d26d\" (UID: \"7cf4ce11-507f-4be4-9652-05035315d26d\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.701415 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cf4ce11-507f-4be4-9652-05035315d26d-operator-scripts\") pod \"7cf4ce11-507f-4be4-9652-05035315d26d\" (UID: \"7cf4ce11-507f-4be4-9652-05035315d26d\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.702305 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cf4ce11-507f-4be4-9652-05035315d26d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7cf4ce11-507f-4be4-9652-05035315d26d" (UID: "7cf4ce11-507f-4be4-9652-05035315d26d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.706517 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cf4ce11-507f-4be4-9652-05035315d26d-kube-api-access-9s8d7" (OuterVolumeSpecName: "kube-api-access-9s8d7") pod "7cf4ce11-507f-4be4-9652-05035315d26d" (UID: "7cf4ce11-507f-4be4-9652-05035315d26d"). InnerVolumeSpecName "kube-api-access-9s8d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.741104 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9eb7-account-create-update-f2s6t"] Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.751653 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 07:10:33 crc kubenswrapper[4978]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658,Command:[/bin/sh -c #!/bin/bash Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: if [ -n "cinder" ]; then Feb 25 07:10:33 crc kubenswrapper[4978]: GRANT_DATABASE="cinder" Feb 25 07:10:33 crc kubenswrapper[4978]: else Feb 25 07:10:33 crc kubenswrapper[4978]: GRANT_DATABASE="*" Feb 25 07:10:33 crc kubenswrapper[4978]: fi Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: # going for maximum compatibility here: Feb 25 07:10:33 crc kubenswrapper[4978]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 25 07:10:33 crc kubenswrapper[4978]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 25 07:10:33 crc kubenswrapper[4978]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 25 07:10:33 crc kubenswrapper[4978]: # support updates Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: $MYSQL_CMD < logger="UnhandledError" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.753160 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"cinder-db-secret\\\" not found\"" pod="openstack/cinder-9eb7-account-create-update-f2s6t" podUID="c7bd49f1-22d2-4ef1-995a-b98bf91a60ce" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.759466 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.763100 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.803161 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s8d7\" (UniqueName: \"kubernetes.io/projected/7cf4ce11-507f-4be4-9652-05035315d26d-kube-api-access-9s8d7\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.803637 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7cf4ce11-507f-4be4-9652-05035315d26d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.857818 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.857860 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b4d65dff6-x8qrr" event={"ID":"d962a2e0-007e-46ca-9ad0-0ecf85a5e87a","Type":"ContainerDied","Data":"970e5ba4ea94db9181506a1fcfbd970c41876a8951f71cdc8d5839d4046099bf"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.857921 4978 scope.go:117] "RemoveContainer" containerID="7f25fb6fb4d79c681a829daa912270875a167f3eab8e635f499b9abf6ae1a887" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.858566 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 07:10:33 crc kubenswrapper[4978]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658,Command:[/bin/sh -c #!/bin/bash Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: if [ -n "nova_api" ]; then Feb 25 07:10:33 crc kubenswrapper[4978]: GRANT_DATABASE="nova_api" Feb 25 07:10:33 crc kubenswrapper[4978]: else Feb 25 07:10:33 crc kubenswrapper[4978]: GRANT_DATABASE="*" Feb 25 07:10:33 crc kubenswrapper[4978]: fi Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: # going for maximum compatibility here: Feb 25 07:10:33 crc kubenswrapper[4978]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 25 07:10:33 crc kubenswrapper[4978]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 25 07:10:33 crc kubenswrapper[4978]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 25 07:10:33 crc kubenswrapper[4978]: # support updates Feb 25 07:10:33 crc kubenswrapper[4978]: Feb 25 07:10:33 crc kubenswrapper[4978]: $MYSQL_CMD < logger="UnhandledError" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.859953 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"nova-api-db-secret\\\" not found\"" pod="openstack/nova-api-9373-account-create-update-lqszb" podUID="cbba37d2-4c5c-467c-9e1c-90c2c3aea110" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.870278 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-9373-account-create-update-lqszb"] Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.878873 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-785ccc58d9-p75wv" event={"ID":"f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0","Type":"ContainerDied","Data":"fdf036e6392d60dcd9126e8debfaca349560b2793a042d255bcad5aac174a529"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.881908 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-785ccc58d9-p75wv" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.884310 4978 generic.go:334] "Generic (PLEG): container finished" podID="34752acd-4270-408b-bc82-af9fc515da2b" containerID="62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd" exitCode=0 Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.884550 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwmn" event={"ID":"34752acd-4270-408b-bc82-af9fc515da2b","Type":"ContainerDied","Data":"62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.901184 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ed9859d2-701a-4011-b9d6-d8cf6291265a/ovsdbserver-nb/0.log" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.901246 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ed9859d2-701a-4011-b9d6-d8cf6291265a","Type":"ContainerDied","Data":"31a02b14eff08a3a2fb7cfd46bb47a4dea8b566f67dc09dede386d4454c64379"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.901355 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.904720 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-rzj6b"] Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905083 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" containerName="probe" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905099 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" containerName="probe" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905108 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" containerName="barbican-keystone-listener" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905115 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" containerName="barbican-keystone-listener" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905125 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" containerName="proxy-httpd" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905131 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" containerName="proxy-httpd" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905138 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9859d2-701a-4011-b9d6-d8cf6291265a" containerName="ovsdbserver-nb" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905144 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9859d2-701a-4011-b9d6-d8cf6291265a" containerName="ovsdbserver-nb" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905152 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9859d2-701a-4011-b9d6-d8cf6291265a" containerName="openstack-network-exporter" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905158 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9859d2-701a-4011-b9d6-d8cf6291265a" containerName="openstack-network-exporter" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905170 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerName="openstack-network-exporter" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905175 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerName="openstack-network-exporter" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905182 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" containerName="ovn-controller" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905188 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" containerName="ovn-controller" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905196 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" containerName="cinder-scheduler" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905201 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" containerName="cinder-scheduler" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905217 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78e5151-267c-4b33-9473-6b249b01f12f" containerName="init" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905222 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78e5151-267c-4b33-9473-6b249b01f12f" containerName="init" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905232 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" containerName="openstack-network-exporter" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905239 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" containerName="openstack-network-exporter" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905253 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a762aa77-9eb1-42eb-98e5-016cc3acac10" containerName="barbican-worker-log" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905259 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a762aa77-9eb1-42eb-98e5-016cc3acac10" containerName="barbican-worker-log" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905270 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a762aa77-9eb1-42eb-98e5-016cc3acac10" containerName="barbican-worker" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905275 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a762aa77-9eb1-42eb-98e5-016cc3acac10" containerName="barbican-worker" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905286 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c488aaac-88dd-4eb1-9020-e17749814ed9" containerName="mysql-bootstrap" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905291 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c488aaac-88dd-4eb1-9020-e17749814ed9" containerName="mysql-bootstrap" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905301 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" containerName="barbican-keystone-listener-log" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905306 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" containerName="barbican-keystone-listener-log" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905316 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerName="ovsdbserver-sb" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905322 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerName="ovsdbserver-sb" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905331 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="807eb4fa-5af0-450e-be50-f30c96731b46" containerName="cinder-api" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905336 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="807eb4fa-5af0-450e-be50-f30c96731b46" containerName="cinder-api" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905347 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185c1474-2c9a-408d-a3b1-8eb4820f6109" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905353 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="185c1474-2c9a-408d-a3b1-8eb4820f6109" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905362 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="807eb4fa-5af0-450e-be50-f30c96731b46" containerName="cinder-api-log" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905378 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="807eb4fa-5af0-450e-be50-f30c96731b46" containerName="cinder-api-log" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905388 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c488aaac-88dd-4eb1-9020-e17749814ed9" containerName="galera" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905394 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c488aaac-88dd-4eb1-9020-e17749814ed9" containerName="galera" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905406 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" containerName="proxy-server" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905412 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" containerName="proxy-server" Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.905424 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78e5151-267c-4b33-9473-6b249b01f12f" containerName="dnsmasq-dns" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905430 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78e5151-267c-4b33-9473-6b249b01f12f" containerName="dnsmasq-dns" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905582 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a762aa77-9eb1-42eb-98e5-016cc3acac10" containerName="barbican-worker" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905592 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerName="ovsdbserver-sb" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905603 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed9859d2-701a-4011-b9d6-d8cf6291265a" containerName="ovsdbserver-nb" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905611 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" containerName="probe" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905621 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78e5151-267c-4b33-9473-6b249b01f12f" containerName="dnsmasq-dns" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905630 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df7a672-b889-40d3-abe5-271e4d36dd85" containerName="openstack-network-exporter" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905638 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="807eb4fa-5af0-450e-be50-f30c96731b46" containerName="cinder-api" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905649 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="807eb4fa-5af0-450e-be50-f30c96731b46" containerName="cinder-api-log" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905658 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" containerName="cinder-scheduler" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905665 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e95552a-ba97-44a7-9ce6-9fef9a84ecfb" containerName="ovn-controller" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905672 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" containerName="barbican-keystone-listener" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905681 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8bf93c4-6b0b-4e09-81b5-d6f36c3f6682" containerName="openstack-network-exporter" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905690 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c488aaac-88dd-4eb1-9020-e17749814ed9" containerName="galera" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905697 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a762aa77-9eb1-42eb-98e5-016cc3acac10" containerName="barbican-worker-log" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905706 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="185c1474-2c9a-408d-a3b1-8eb4820f6109" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905715 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" containerName="proxy-httpd" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905723 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" containerName="barbican-keystone-listener-log" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905730 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed9859d2-701a-4011-b9d6-d8cf6291265a" containerName="openstack-network-exporter" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.905737 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" containerName="proxy-server" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.906259 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rzj6b" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.906469 4978 scope.go:117] "RemoveContainer" containerID="965888d2495c19353834efdc1dc9bee729c306c0310cc39f6260c432b9398f32" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.908100 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-596qq\" (UniqueName: \"kubernetes.io/projected/c488aaac-88dd-4eb1-9020-e17749814ed9-kube-api-access-596qq\") pod \"c488aaac-88dd-4eb1-9020-e17749814ed9\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.908130 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-internal-tls-certs\") pod \"807eb4fa-5af0-450e-be50-f30c96731b46\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.911740 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-lhjvp" event={"ID":"97e7baca-befe-4eba-9652-5474e0e88899","Type":"ContainerDied","Data":"a51fee4a5ce94a963831a2282e192c0b87e3268ae1de07faec23d7dda17e50ba"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.911862 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lhjvp" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.914066 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.915126 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-generated\") pod \"c488aaac-88dd-4eb1-9020-e17749814ed9\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.916215 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "c488aaac-88dd-4eb1-9020-e17749814ed9" (UID: "c488aaac-88dd-4eb1-9020-e17749814ed9"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.917585 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-operator-scripts\") pod \"c488aaac-88dd-4eb1-9020-e17749814ed9\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.917907 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/807eb4fa-5af0-450e-be50-f30c96731b46-logs\") pod \"807eb4fa-5af0-450e-be50-f30c96731b46\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.917961 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85rdz\" (UniqueName: \"kubernetes.io/projected/807eb4fa-5af0-450e-be50-f30c96731b46-kube-api-access-85rdz\") pod \"807eb4fa-5af0-450e-be50-f30c96731b46\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.917980 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-galera-tls-certs\") pod \"c488aaac-88dd-4eb1-9020-e17749814ed9\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918000 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-kolla-config\") pod \"c488aaac-88dd-4eb1-9020-e17749814ed9\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918019 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-combined-ca-bundle\") pod \"807eb4fa-5af0-450e-be50-f30c96731b46\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918038 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data\") pod \"807eb4fa-5af0-450e-be50-f30c96731b46\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918120 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-scripts\") pod \"807eb4fa-5af0-450e-be50-f30c96731b46\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918143 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c488aaac-88dd-4eb1-9020-e17749814ed9\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918181 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/807eb4fa-5af0-450e-be50-f30c96731b46-etc-machine-id\") pod \"807eb4fa-5af0-450e-be50-f30c96731b46\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918213 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data-custom\") pod \"807eb4fa-5af0-450e-be50-f30c96731b46\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918257 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-combined-ca-bundle\") pod \"c488aaac-88dd-4eb1-9020-e17749814ed9\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918275 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-default\") pod \"c488aaac-88dd-4eb1-9020-e17749814ed9\" (UID: \"c488aaac-88dd-4eb1-9020-e17749814ed9\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918293 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-public-tls-certs\") pod \"807eb4fa-5af0-450e-be50-f30c96731b46\" (UID: \"807eb4fa-5af0-450e-be50-f30c96731b46\") " Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918474 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c488aaac-88dd-4eb1-9020-e17749814ed9-kube-api-access-596qq" (OuterVolumeSpecName: "kube-api-access-596qq") pod "c488aaac-88dd-4eb1-9020-e17749814ed9" (UID: "c488aaac-88dd-4eb1-9020-e17749814ed9"). InnerVolumeSpecName "kube-api-access-596qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.918522 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/807eb4fa-5af0-450e-be50-f30c96731b46-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "807eb4fa-5af0-450e-be50-f30c96731b46" (UID: "807eb4fa-5af0-450e-be50-f30c96731b46"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.919960 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ec4-account-create-update-649r6" event={"ID":"84e32f1f-5ad1-40a8-8f61-21025d84d003","Type":"ContainerStarted","Data":"36cd594e8e8fab32c5d5c0941858a59337834b174129f331eade37a79dfd42e8"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.925356 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "c488aaac-88dd-4eb1-9020-e17749814ed9" (UID: "c488aaac-88dd-4eb1-9020-e17749814ed9"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.925942 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/807eb4fa-5af0-450e-be50-f30c96731b46-logs" (OuterVolumeSpecName: "logs") pod "807eb4fa-5af0-450e-be50-f30c96731b46" (UID: "807eb4fa-5af0-450e-be50-f30c96731b46"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.925943 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-rzj6b"] Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.926294 4978 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Feb 25 07:10:33 crc kubenswrapper[4978]: E0225 07:10:33.926483 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data podName:09112202-a3ef-4742-bdbe-b46c1c9a55ba nodeName:}" failed. No retries permitted until 2026-02-25 07:10:37.926457073 +0000 UTC m=+1531.365713532 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data") pod "rabbitmq-server-0" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba") : configmap "rabbitmq-config-data" not found Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.926777 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-596qq\" (UniqueName: \"kubernetes.io/projected/c488aaac-88dd-4eb1-9020-e17749814ed9-kube-api-access-596qq\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.926804 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-generated\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.926814 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/807eb4fa-5af0-450e-be50-f30c96731b46-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.926825 4978 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/807eb4fa-5af0-450e-be50-f30c96731b46-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.926836 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-config-data-default\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.926855 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c488aaac-88dd-4eb1-9020-e17749814ed9" (UID: "c488aaac-88dd-4eb1-9020-e17749814ed9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.929689 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "c488aaac-88dd-4eb1-9020-e17749814ed9" (UID: "c488aaac-88dd-4eb1-9020-e17749814ed9"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.932520 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "807eb4fa-5af0-450e-be50-f30c96731b46" (UID: "807eb4fa-5af0-450e-be50-f30c96731b46"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.932954 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-scripts" (OuterVolumeSpecName: "scripts") pod "807eb4fa-5af0-450e-be50-f30c96731b46" (UID: "807eb4fa-5af0-450e-be50-f30c96731b46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.933003 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-5b4d65dff6-x8qrr"] Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.947898 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-5b4d65dff6-x8qrr"] Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.956900 4978 generic.go:334] "Generic (PLEG): container finished" podID="807eb4fa-5af0-450e-be50-f30c96731b46" containerID="e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644" exitCode=0 Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.956967 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"807eb4fa-5af0-450e-be50-f30c96731b46","Type":"ContainerDied","Data":"e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.956993 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"807eb4fa-5af0-450e-be50-f30c96731b46","Type":"ContainerDied","Data":"32c85db95df224855ea938a7986cce424a25a22f5af89d58ab60ae8b5161bd02"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.957068 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.957537 4978 scope.go:117] "RemoveContainer" containerID="d9e6c02d47ab88e1a99d79a917daf7a487b0b7f226c09ae65c011a4c5088f429" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.961634 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/807eb4fa-5af0-450e-be50-f30c96731b46-kube-api-access-85rdz" (OuterVolumeSpecName: "kube-api-access-85rdz") pod "807eb4fa-5af0-450e-be50-f30c96731b46" (UID: "807eb4fa-5af0-450e-be50-f30c96731b46"). InnerVolumeSpecName "kube-api-access-85rdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.962635 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a8c-account-create-update-d7wm6" event={"ID":"83bd2445-af69-432a-bc55-d5cee299162d","Type":"ContainerStarted","Data":"7c7f63cf88ea4ea3379a36a8d4d157cdf1918c77c3bf5759473a8b46ee2aa4e2"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.985995 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8f7c6df-f16d-487d-98e8-cab2c89073b4","Type":"ContainerDied","Data":"097ec97b849c38518ace2fc78c37ad724fd0b060ecec088a0e316d7fa0266f50"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.986132 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.987061 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "c488aaac-88dd-4eb1-9020-e17749814ed9" (UID: "c488aaac-88dd-4eb1-9020-e17749814ed9"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.996182 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-cb34-account-create-update-2blwx" event={"ID":"7cf4ce11-507f-4be4-9652-05035315d26d","Type":"ContainerDied","Data":"32c1928875454caef6ca3ab7aafd7711a1600bc85e947976a39e09864729b987"} Feb 25 07:10:33 crc kubenswrapper[4978]: I0225 07:10:33.996278 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-cb34-account-create-update-2blwx" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.002080 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7696768c55-mnqh9" event={"ID":"a762aa77-9eb1-42eb-98e5-016cc3acac10","Type":"ContainerDied","Data":"5e7e5ca67e7e0604612baa166176d1957dd92e7cfb16bf5f97d6305b191dea2f"} Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.002137 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7696768c55-mnqh9" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.006541 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"185c1474-2c9a-408d-a3b1-8eb4820f6109","Type":"ContainerDied","Data":"fddc28cae6f8e09ad937d7d26c7a1b34f83080dc03593ec85877b1767e65396d"} Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.006580 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.009846 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "807eb4fa-5af0-450e-be50-f30c96731b46" (UID: "807eb4fa-5af0-450e-be50-f30c96731b46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.038965 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9eb7-account-create-update-f2s6t" event={"ID":"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce","Type":"ContainerStarted","Data":"a0e1b71f6970f985eaccadb4987a0c7254c6b4f1cb4b8a22c5a89705546dc3b9"} Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.060586 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w9bx\" (UniqueName: \"kubernetes.io/projected/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-kube-api-access-2w9bx\") pod \"root-account-create-update-rzj6b\" (UID: \"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c\") " pod="openstack/root-account-create-update-rzj6b" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.060700 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-operator-scripts\") pod \"root-account-create-update-rzj6b\" (UID: \"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c\") " pod="openstack/root-account-create-update-rzj6b" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.060853 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.060875 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.060885 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.060894 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.060903 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85rdz\" (UniqueName: \"kubernetes.io/projected/807eb4fa-5af0-450e-be50-f30c96731b46-kube-api-access-85rdz\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.060911 4978 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c488aaac-88dd-4eb1-9020-e17749814ed9-kolla-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.060919 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.076541 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data" (OuterVolumeSpecName: "config-data") pod "807eb4fa-5af0-450e-be50-f30c96731b46" (UID: "807eb4fa-5af0-450e-be50-f30c96731b46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.117475 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "807eb4fa-5af0-450e-be50-f30c96731b46" (UID: "807eb4fa-5af0-450e-be50-f30c96731b46"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.151711 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.161910 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "c488aaac-88dd-4eb1-9020-e17749814ed9" (UID: "c488aaac-88dd-4eb1-9020-e17749814ed9"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.162039 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w9bx\" (UniqueName: \"kubernetes.io/projected/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-kube-api-access-2w9bx\") pod \"root-account-create-update-rzj6b\" (UID: \"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c\") " pod="openstack/root-account-create-update-rzj6b" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.162097 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-operator-scripts\") pod \"root-account-create-update-rzj6b\" (UID: \"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c\") " pod="openstack/root-account-create-update-rzj6b" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.162229 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.162242 4978 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.162251 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.162259 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.162907 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-operator-scripts\") pod \"root-account-create-update-rzj6b\" (UID: \"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c\") " pod="openstack/root-account-create-update-rzj6b" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.164496 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c488aaac-88dd-4eb1-9020-e17749814ed9" (UID: "c488aaac-88dd-4eb1-9020-e17749814ed9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.171705 4978 generic.go:334] "Generic (PLEG): container finished" podID="c488aaac-88dd-4eb1-9020-e17749814ed9" containerID="df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79" exitCode=0 Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.171765 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c488aaac-88dd-4eb1-9020-e17749814ed9","Type":"ContainerDied","Data":"df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79"} Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.171792 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c488aaac-88dd-4eb1-9020-e17749814ed9","Type":"ContainerDied","Data":"58b5d1d02d020975779c0d2f0eae1010d9bf42641ae62f59a5cbe5a067fc115d"} Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.171862 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.183112 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w9bx\" (UniqueName: \"kubernetes.io/projected/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-kube-api-access-2w9bx\") pod \"root-account-create-update-rzj6b\" (UID: \"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c\") " pod="openstack/root-account-create-update-rzj6b" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.183176 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-lhjvp"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.187600 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.189719 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.211900 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-lhjvp"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.227740 4978 scope.go:117] "RemoveContainer" containerID="e0503e6f2f40da27125f61fc3830d5697085e44e0e23aa2a64cee749b3cf958e" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.228457 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-785ccc58d9-p75wv"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.242520 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-785ccc58d9-p75wv"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.242676 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "807eb4fa-5af0-450e-be50-f30c96731b46" (UID: "807eb4fa-5af0-450e-be50-f30c96731b46"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.254485 4978 scope.go:117] "RemoveContainer" containerID="a15f02424a1013ac6691c287f99c51cc3e6aff1c7146b986bb694436609fc027" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.255098 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rzj6b" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.260449 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1b7b-account-create-update-qlxzw" event={"ID":"bc147cb6-a5d8-47f7-8ce6-168455f18512","Type":"ContainerStarted","Data":"db33e38cd6265a879b5db49dc5a0c9b860d27722dd6b77376d11042cdf69fa62"} Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.264531 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c488aaac-88dd-4eb1-9020-e17749814ed9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.264559 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/807eb4fa-5af0-450e-be50-f30c96731b46-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.291070 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.302523 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.315953 4978 scope.go:117] "RemoveContainer" containerID="8baa6e062d410fe1594d058fc3924f5a1ed85894951c8f77a4918acff9d1e1bd" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.316111 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7696768c55-mnqh9"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.331971 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-7696768c55-mnqh9"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.350522 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.353725 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.368205 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-cb34-account-create-update-2blwx"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.381443 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-cb34-account-create-update-2blwx"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.411445 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.429486 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.441577 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.447184 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.453399 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.467204 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.606689 4978 scope.go:117] "RemoveContainer" containerID="e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.627855 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.630007 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="ceilometer-central-agent" containerID="cri-o://a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a" gracePeriod=30 Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.630400 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="proxy-httpd" containerID="cri-o://f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b" gracePeriod=30 Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.630455 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="sg-core" containerID="cri-o://887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d" gracePeriod=30 Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.630484 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="ceilometer-notification-agent" containerID="cri-o://48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae" gracePeriod=30 Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.668704 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.668882 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3be5007a-36e1-4b55-a8d8-5fd560efb8fe" containerName="kube-state-metrics" containerID="cri-o://78414f7a4e1d623e9a11a84d9c8ca28b022bcdb53fed8a932a715712c16abf12" gracePeriod=30 Feb 25 07:10:34 crc kubenswrapper[4978]: E0225 07:10:34.684196 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 07:10:34 crc kubenswrapper[4978]: E0225 07:10:34.717515 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.722117 4978 scope.go:117] "RemoveContainer" containerID="05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc" Feb 25 07:10:34 crc kubenswrapper[4978]: E0225 07:10:34.736883 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 07:10:34 crc kubenswrapper[4978]: E0225 07:10:34.736931 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="50c62987-be2b-4deb-a88c-107db5d00415" containerName="nova-cell0-conductor-conductor" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.747098 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-d7wm6" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.772110 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-77bc5d768-qx88q" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.164:9311/healthcheck\": read tcp 10.217.0.2:38310->10.217.0.164:9311: read: connection reset by peer" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.772365 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-77bc5d768-qx88q" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.164:9311/healthcheck\": read tcp 10.217.0.2:38324->10.217.0.164:9311: read: connection reset by peer" Feb 25 07:10:34 crc kubenswrapper[4978]: E0225 07:10:34.780534 4978 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Feb 25 07:10:34 crc kubenswrapper[4978]: E0225 07:10:34.780581 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data podName:43bd8325-9408-4a32-9285-d522d497205e nodeName:}" failed. No retries permitted until 2026-02-25 07:10:38.780568844 +0000 UTC m=+1532.219825293 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data") pod "rabbitmq-cell1-server-0" (UID: "43bd8325-9408-4a32-9285-d522d497205e") : configmap "rabbitmq-cell1-config-data" not found Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.789592 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.789815 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="85624b7d-192e-42c3-9c5b-18cfd91d37e9" containerName="memcached" containerID="cri-o://22db8e67ef3c374833a5a7d3778dc19f3cb69c3620c3009d7badbe8c31ef7006" gracePeriod=30 Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.845676 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-56f1-account-create-update-5zpqd"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.863466 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-56f1-account-create-update-5zpqd"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.876900 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-56f1-account-create-update-dvwqn"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.878079 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.880072 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-56f1-account-create-update-dvwqn"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.883623 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.884314 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59csl\" (UniqueName: \"kubernetes.io/projected/83bd2445-af69-432a-bc55-d5cee299162d-kube-api-access-59csl\") pod \"83bd2445-af69-432a-bc55-d5cee299162d\" (UID: \"83bd2445-af69-432a-bc55-d5cee299162d\") " Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.884404 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bd2445-af69-432a-bc55-d5cee299162d-operator-scripts\") pod \"83bd2445-af69-432a-bc55-d5cee299162d\" (UID: \"83bd2445-af69-432a-bc55-d5cee299162d\") " Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.893155 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-q4lmf"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.894622 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83bd2445-af69-432a-bc55-d5cee299162d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83bd2445-af69-432a-bc55-d5cee299162d" (UID: "83bd2445-af69-432a-bc55-d5cee299162d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.910303 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-q4lmf"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.915090 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-phpsb"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.916539 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83bd2445-af69-432a-bc55-d5cee299162d-kube-api-access-59csl" (OuterVolumeSpecName: "kube-api-access-59csl") pod "83bd2445-af69-432a-bc55-d5cee299162d" (UID: "83bd2445-af69-432a-bc55-d5cee299162d"). InnerVolumeSpecName "kube-api-access-59csl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.933825 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-phpsb"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.947495 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7dd4fddc6c-nv49c"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.947732 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-7dd4fddc6c-nv49c" podUID="3ee398cf-5bd3-4905-80f2-b66d67c57f3d" containerName="keystone-api" containerID="cri-o://3b71dd25e21037082b24f81353af12fb7fe11432d9ea70560778f8494ba5ae98" gracePeriod=30 Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.975442 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.985703 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts\") pod \"keystone-56f1-account-create-update-dvwqn\" (UID: \"4e53c877-f109-4d98-9388-25dccf9bd229\") " pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.990174 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcnqg\" (UniqueName: \"kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg\") pod \"keystone-56f1-account-create-update-dvwqn\" (UID: \"4e53c877-f109-4d98-9388-25dccf9bd229\") " pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.990284 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59csl\" (UniqueName: \"kubernetes.io/projected/83bd2445-af69-432a-bc55-d5cee299162d-kube-api-access-59csl\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.990299 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83bd2445-af69-432a-bc55-d5cee299162d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:34 crc kubenswrapper[4978]: I0225 07:10:34.990413 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rzj6b"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:34.997027 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-kvbg5"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.008315 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-kvbg5"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.011120 4978 scope.go:117] "RemoveContainer" containerID="e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644" Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.012783 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644\": container with ID starting with e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644 not found: ID does not exist" containerID="e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.012825 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644"} err="failed to get container status \"e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644\": rpc error: code = NotFound desc = could not find container \"e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644\": container with ID starting with e0ba4ed2ef8fb5d221b71071988aa6c415e1aaee05f023ceeefa0779a238d644 not found: ID does not exist" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.012847 4978 scope.go:117] "RemoveContainer" containerID="05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.013041 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.206:3000/\": read tcp 10.217.0.2:40502->10.217.0.206:3000: read: connection reset by peer" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.020745 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-56f1-account-create-update-dvwqn"] Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.021278 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-rcnqg operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone-56f1-account-create-update-dvwqn" podUID="4e53c877-f109-4d98-9388-25dccf9bd229" Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.039789 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc\": container with ID starting with 05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc not found: ID does not exist" containerID="05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.039828 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc"} err="failed to get container status \"05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc\": rpc error: code = NotFound desc = could not find container \"05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc\": container with ID starting with 05f82d930ba453e0982ea74218677bae48f8366595a5a2b2e936f2ad139516fc not found: ID does not exist" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.039853 4978 scope.go:117] "RemoveContainer" containerID="8a5993b793c50295987a7b0055fc564f776c502b1b808c0832ceeaaacca5e945" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.091428 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcnqg\" (UniqueName: \"kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg\") pod \"keystone-56f1-account-create-update-dvwqn\" (UID: \"4e53c877-f109-4d98-9388-25dccf9bd229\") " pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.091516 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts\") pod \"keystone-56f1-account-create-update-dvwqn\" (UID: \"4e53c877-f109-4d98-9388-25dccf9bd229\") " pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.091643 4978 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.091684 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts podName:4e53c877-f109-4d98-9388-25dccf9bd229 nodeName:}" failed. No retries permitted until 2026-02-25 07:10:35.591671286 +0000 UTC m=+1529.030927745 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts") pod "keystone-56f1-account-create-update-dvwqn" (UID: "4e53c877-f109-4d98-9388-25dccf9bd229") : configmap "openstack-scripts" not found Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.095662 4978 projected.go:194] Error preparing data for projected volume kube-api-access-rcnqg for pod openstack/keystone-56f1-account-create-update-dvwqn: failed to fetch token: serviceaccounts "galera-openstack" not found Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.095690 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg podName:4e53c877-f109-4d98-9388-25dccf9bd229 nodeName:}" failed. No retries permitted until 2026-02-25 07:10:35.59568206 +0000 UTC m=+1529.034938519 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-rcnqg" (UniqueName: "kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg") pod "keystone-56f1-account-create-update-dvwqn" (UID: "4e53c877-f109-4d98-9388-25dccf9bd229") : failed to fetch token: serviceaccounts "galera-openstack" not found Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.111811 4978 scope.go:117] "RemoveContainer" containerID="abc550e17c1b943bad86597853b93ec464de90bcf8090229b1e9cb5d9aeca0d7" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.138726 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.154471 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": dial tcp 10.217.0.211:8775: connect: connection refused" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.154549 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": dial tcp 10.217.0.211:8775: connect: connection refused" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.180584 4978 scope.go:117] "RemoveContainer" containerID="97d4b20d6d6a152207b32f06fed0f8b590209ec24594d4831ee695bb8b21c4ee" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.191934 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-combined-ca-bundle\") pod \"c6a16d36-c629-483e-9b15-fbd7b8913883\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.192018 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-config-data\") pod \"c6a16d36-c629-483e-9b15-fbd7b8913883\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.192054 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7g6j\" (UniqueName: \"kubernetes.io/projected/c6a16d36-c629-483e-9b15-fbd7b8913883-kube-api-access-w7g6j\") pod \"c6a16d36-c629-483e-9b15-fbd7b8913883\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.192115 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-public-tls-certs\") pod \"c6a16d36-c629-483e-9b15-fbd7b8913883\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.192155 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a16d36-c629-483e-9b15-fbd7b8913883-logs\") pod \"c6a16d36-c629-483e-9b15-fbd7b8913883\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.192187 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-scripts\") pod \"c6a16d36-c629-483e-9b15-fbd7b8913883\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.192226 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-internal-tls-certs\") pod \"c6a16d36-c629-483e-9b15-fbd7b8913883\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.196777 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6a16d36-c629-483e-9b15-fbd7b8913883-logs" (OuterVolumeSpecName: "logs") pod "c6a16d36-c629-483e-9b15-fbd7b8913883" (UID: "c6a16d36-c629-483e-9b15-fbd7b8913883"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.204608 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6a16d36-c629-483e-9b15-fbd7b8913883-kube-api-access-w7g6j" (OuterVolumeSpecName: "kube-api-access-w7g6j") pod "c6a16d36-c629-483e-9b15-fbd7b8913883" (UID: "c6a16d36-c629-483e-9b15-fbd7b8913883"). InnerVolumeSpecName "kube-api-access-w7g6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.230215 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-scripts" (OuterVolumeSpecName: "scripts") pod "c6a16d36-c629-483e-9b15-fbd7b8913883" (UID: "c6a16d36-c629-483e-9b15-fbd7b8913883"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.235423 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1b7b-account-create-update-qlxzw" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.250296 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9eb7-account-create-update-f2s6t" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.269768 4978 scope.go:117] "RemoveContainer" containerID="47894865746e5efcf05b88359453cdd68cf29193d3fc760e152761c140f46884" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.277312 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bwjnj" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerName="registry-server" probeResult="failure" output=< Feb 25 07:10:35 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 07:10:35 crc kubenswrapper[4978]: > Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.279172 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ec4-account-create-update-649r6" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.287438 4978 generic.go:334] "Generic (PLEG): container finished" podID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerID="f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b" exitCode=0 Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.287468 4978 generic.go:334] "Generic (PLEG): container finished" podID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerID="887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d" exitCode=2 Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.287524 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ddb624b-1c38-44b8-8f02-bd177937ef4d","Type":"ContainerDied","Data":"f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.287556 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ddb624b-1c38-44b8-8f02-bd177937ef4d","Type":"ContainerDied","Data":"887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.292853 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-operator-scripts\") pod \"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce\" (UID: \"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.292935 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzlkq\" (UniqueName: \"kubernetes.io/projected/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-kube-api-access-mzlkq\") pod \"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce\" (UID: \"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.296861 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-config-data" (OuterVolumeSpecName: "config-data") pod "c6a16d36-c629-483e-9b15-fbd7b8913883" (UID: "c6a16d36-c629-483e-9b15-fbd7b8913883"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.292957 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkzsx\" (UniqueName: \"kubernetes.io/projected/bc147cb6-a5d8-47f7-8ce6-168455f18512-kube-api-access-zkzsx\") pod \"bc147cb6-a5d8-47f7-8ce6-168455f18512\" (UID: \"bc147cb6-a5d8-47f7-8ce6-168455f18512\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.308427 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q7kb\" (UniqueName: \"kubernetes.io/projected/84e32f1f-5ad1-40a8-8f61-21025d84d003-kube-api-access-8q7kb\") pod \"84e32f1f-5ad1-40a8-8f61-21025d84d003\" (UID: \"84e32f1f-5ad1-40a8-8f61-21025d84d003\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.308460 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-config-data\") pod \"c6a16d36-c629-483e-9b15-fbd7b8913883\" (UID: \"c6a16d36-c629-483e-9b15-fbd7b8913883\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.308487 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc147cb6-a5d8-47f7-8ce6-168455f18512-operator-scripts\") pod \"bc147cb6-a5d8-47f7-8ce6-168455f18512\" (UID: \"bc147cb6-a5d8-47f7-8ce6-168455f18512\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.308554 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84e32f1f-5ad1-40a8-8f61-21025d84d003-operator-scripts\") pod \"84e32f1f-5ad1-40a8-8f61-21025d84d003\" (UID: \"84e32f1f-5ad1-40a8-8f61-21025d84d003\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.309143 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6a16d36-c629-483e-9b15-fbd7b8913883-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.309154 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.309165 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7g6j\" (UniqueName: \"kubernetes.io/projected/c6a16d36-c629-483e-9b15-fbd7b8913883-kube-api-access-w7g6j\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.309517 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84e32f1f-5ad1-40a8-8f61-21025d84d003-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "84e32f1f-5ad1-40a8-8f61-21025d84d003" (UID: "84e32f1f-5ad1-40a8-8f61-21025d84d003"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: W0225 07:10:35.312390 4978 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c6a16d36-c629-483e-9b15-fbd7b8913883/volumes/kubernetes.io~secret/config-data Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.312473 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-config-data" (OuterVolumeSpecName: "config-data") pod "c6a16d36-c629-483e-9b15-fbd7b8913883" (UID: "c6a16d36-c629-483e-9b15-fbd7b8913883"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.312909 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc147cb6-a5d8-47f7-8ce6-168455f18512-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc147cb6-a5d8-47f7-8ce6-168455f18512" (UID: "bc147cb6-a5d8-47f7-8ce6-168455f18512"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.338150 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c7bd49f1-22d2-4ef1-995a-b98bf91a60ce" (UID: "c7bd49f1-22d2-4ef1-995a-b98bf91a60ce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.355804 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="185c1474-2c9a-408d-a3b1-8eb4820f6109" path="/var/lib/kubelet/pods/185c1474-2c9a-408d-a3b1-8eb4820f6109/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.356303 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cf4ce11-507f-4be4-9652-05035315d26d" path="/var/lib/kubelet/pods/7cf4ce11-507f-4be4-9652-05035315d26d/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.356750 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="807eb4fa-5af0-450e-be50-f30c96731b46" path="/var/lib/kubelet/pods/807eb4fa-5af0-450e-be50-f30c96731b46/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.357365 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97e7baca-befe-4eba-9652-5474e0e88899" path="/var/lib/kubelet/pods/97e7baca-befe-4eba-9652-5474e0e88899/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.358015 4978 generic.go:334] "Generic (PLEG): container finished" podID="3be5007a-36e1-4b55-a8d8-5fd560efb8fe" containerID="78414f7a4e1d623e9a11a84d9c8ca28b022bcdb53fed8a932a715712c16abf12" exitCode=2 Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.360664 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a762aa77-9eb1-42eb-98e5-016cc3acac10" path="/var/lib/kubelet/pods/a762aa77-9eb1-42eb-98e5-016cc3acac10/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.360925 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e32f1f-5ad1-40a8-8f61-21025d84d003-kube-api-access-8q7kb" (OuterVolumeSpecName: "kube-api-access-8q7kb") pod "84e32f1f-5ad1-40a8-8f61-21025d84d003" (UID: "84e32f1f-5ad1-40a8-8f61-21025d84d003"). InnerVolumeSpecName "kube-api-access-8q7kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.361355 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c488aaac-88dd-4eb1-9020-e17749814ed9" path="/var/lib/kubelet/pods/c488aaac-88dd-4eb1-9020-e17749814ed9/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.361940 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c953f23d-4f0e-4be8-9b37-b26b7de7cad4" path="/var/lib/kubelet/pods/c953f23d-4f0e-4be8-9b37-b26b7de7cad4/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.362933 4978 generic.go:334] "Generic (PLEG): container finished" podID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" containerID="4b497e57dce20bdb93e7a82daed8baa0e9acd2b6048c9e18954a4341399c8f5b" exitCode=0 Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.363230 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d31c2c4a-d61c-496d-aef6-31895fee244f" path="/var/lib/kubelet/pods/d31c2c4a-d61c-496d-aef6-31895fee244f/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.363789 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d962a2e0-007e-46ca-9ad0-0ecf85a5e87a" path="/var/lib/kubelet/pods/d962a2e0-007e-46ca-9ad0-0ecf85a5e87a/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.364302 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df4bb775-9429-45bf-af46-93dcf3b4a82d" path="/var/lib/kubelet/pods/df4bb775-9429-45bf-af46-93dcf3b4a82d/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.365128 4978 generic.go:334] "Generic (PLEG): container finished" podID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" containerID="231167ca97b4be8ade22723cde0ed17906f6163a3cedd000bd07b8e2dddaaa83" exitCode=0 Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.365601 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed9859d2-701a-4011-b9d6-d8cf6291265a" path="/var/lib/kubelet/pods/ed9859d2-701a-4011-b9d6-d8cf6291265a/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.366276 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0" path="/var/lib/kubelet/pods/f5aee89f-a16a-4c7f-aa7d-d97b5f9f8cd0/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.367263 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8f7c6df-f16d-487d-98e8-cab2c89073b4" path="/var/lib/kubelet/pods/f8f7c6df-f16d-487d-98e8-cab2c89073b4/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.367911 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9" path="/var/lib/kubelet/pods/fe5d5de6-2bd9-4bba-8730-bc87bb2e27e9/volumes" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.369151 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc147cb6-a5d8-47f7-8ce6-168455f18512-kube-api-access-zkzsx" (OuterVolumeSpecName: "kube-api-access-zkzsx") pod "bc147cb6-a5d8-47f7-8ce6-168455f18512" (UID: "bc147cb6-a5d8-47f7-8ce6-168455f18512"). InnerVolumeSpecName "kube-api-access-zkzsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.370850 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-kube-api-access-mzlkq" (OuterVolumeSpecName: "kube-api-access-mzlkq") pod "c7bd49f1-22d2-4ef1-995a-b98bf91a60ce" (UID: "c7bd49f1-22d2-4ef1-995a-b98bf91a60ce"). InnerVolumeSpecName "kube-api-access-mzlkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.410265 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6a16d36-c629-483e-9b15-fbd7b8913883" (UID: "c6a16d36-c629-483e-9b15-fbd7b8913883"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.410458 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.410473 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.410483 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzlkq\" (UniqueName: \"kubernetes.io/projected/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce-kube-api-access-mzlkq\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.410493 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkzsx\" (UniqueName: \"kubernetes.io/projected/bc147cb6-a5d8-47f7-8ce6-168455f18512-kube-api-access-zkzsx\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.410502 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q7kb\" (UniqueName: \"kubernetes.io/projected/84e32f1f-5ad1-40a8-8f61-21025d84d003-kube-api-access-8q7kb\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.410512 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.410521 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc147cb6-a5d8-47f7-8ce6-168455f18512-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.410529 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84e32f1f-5ad1-40a8-8f61-21025d84d003-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.425734 4978 generic.go:334] "Generic (PLEG): container finished" podID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerID="33f35943890b2d404b619750739644e060d0152191e4ffd0d1bb79cdebfb8e65" exitCode=0 Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.426904 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8a8c-account-create-update-d7wm6" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.428252 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9eb7-account-create-update-f2s6t" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.433924 4978 generic.go:334] "Generic (PLEG): container finished" podID="c6a16d36-c629-483e-9b15-fbd7b8913883" containerID="03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6" exitCode=0 Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.434002 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69794fc6d6-zqnw6" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.438873 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6ec4-account-create-update-649r6" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.463913 4978 scope.go:117] "RemoveContainer" containerID="70fcc3b61d5c76d88b1ebc2b70685558693748a8e9f0734e5a69556982ffa4cb" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.464174 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-1b7b-account-create-update-qlxzw" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.496396 4978 generic.go:334] "Generic (PLEG): container finished" podID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerID="e3bb5987c1f9fda6b9f36d86116e1f3a943d9693a3917d0c5c728a4dd3429744" exitCode=0 Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.503027 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.522108 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nxwmn" podStartSLOduration=4.661535166 podStartE2EDuration="7.522088965s" podCreationTimestamp="2026-02-25 07:10:28 +0000 UTC" firstStartedPulling="2026-02-25 07:10:31.532595658 +0000 UTC m=+1524.971852117" lastFinishedPulling="2026-02-25 07:10:34.393149457 +0000 UTC m=+1527.832405916" observedRunningTime="2026-02-25 07:10:35.423291509 +0000 UTC m=+1528.862547978" watchObservedRunningTime="2026-02-25 07:10:35.522088965 +0000 UTC m=+1528.961345424" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.551113 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c6a16d36-c629-483e-9b15-fbd7b8913883" (UID: "c6a16d36-c629-483e-9b15-fbd7b8913883"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.567084 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c6a16d36-c629-483e-9b15-fbd7b8913883" (UID: "c6a16d36-c629-483e-9b15-fbd7b8913883"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.589892 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="315eefbb-24fb-4e45-97ed-03caf894bc08" containerName="galera" containerID="cri-o://4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f" gracePeriod=30 Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.620046 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcnqg\" (UniqueName: \"kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg\") pod \"keystone-56f1-account-create-update-dvwqn\" (UID: \"4e53c877-f109-4d98-9388-25dccf9bd229\") " pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.621438 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts\") pod \"keystone-56f1-account-create-update-dvwqn\" (UID: \"4e53c877-f109-4d98-9388-25dccf9bd229\") " pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.621696 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.621715 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6a16d36-c629-483e-9b15-fbd7b8913883-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.621780 4978 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.621832 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts podName:4e53c877-f109-4d98-9388-25dccf9bd229 nodeName:}" failed. No retries permitted until 2026-02-25 07:10:36.62181568 +0000 UTC m=+1530.061072139 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts") pod "keystone-56f1-account-create-update-dvwqn" (UID: "4e53c877-f109-4d98-9388-25dccf9bd229") : configmap "openstack-scripts" not found Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.630608 4978 projected.go:194] Error preparing data for projected volume kube-api-access-rcnqg for pod openstack/keystone-56f1-account-create-update-dvwqn: failed to fetch token: serviceaccounts "galera-openstack" not found Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.630666 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg podName:4e53c877-f109-4d98-9388-25dccf9bd229 nodeName:}" failed. No retries permitted until 2026-02-25 07:10:36.630648081 +0000 UTC m=+1530.069904540 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-rcnqg" (UniqueName: "kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg") pod "keystone-56f1-account-create-update-dvwqn" (UID: "4e53c877-f109-4d98-9388-25dccf9bd229") : failed to fetch token: serviceaccounts "galera-openstack" not found Feb 25 07:10:35 crc kubenswrapper[4978]: W0225 07:10:35.695200 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ccfe3b6_2fd4_432e_b6d4_71869edbfe4c.slice/crio-b74a748b253680333605d35b10a423526a41b81f2dcdc68c37a40527bc86cb3e WatchSource:0}: Error finding container b74a748b253680333605d35b10a423526a41b81f2dcdc68c37a40527bc86cb3e: Status 404 returned error can't find the container with id b74a748b253680333605d35b10a423526a41b81f2dcdc68c37a40527bc86cb3e Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.699140 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 07:10:35 crc kubenswrapper[4978]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:bafa959fd4a24c80de0c6b1c5adbf2b44992312068ca741c6a0717d49c919658,Command:[/bin/sh -c #!/bin/bash Feb 25 07:10:35 crc kubenswrapper[4978]: Feb 25 07:10:35 crc kubenswrapper[4978]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Feb 25 07:10:35 crc kubenswrapper[4978]: Feb 25 07:10:35 crc kubenswrapper[4978]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Feb 25 07:10:35 crc kubenswrapper[4978]: Feb 25 07:10:35 crc kubenswrapper[4978]: MYSQL_CMD="mysql -h -u root -P 3306" Feb 25 07:10:35 crc kubenswrapper[4978]: Feb 25 07:10:35 crc kubenswrapper[4978]: if [ -n "" ]; then Feb 25 07:10:35 crc kubenswrapper[4978]: GRANT_DATABASE="" Feb 25 07:10:35 crc kubenswrapper[4978]: else Feb 25 07:10:35 crc kubenswrapper[4978]: GRANT_DATABASE="*" Feb 25 07:10:35 crc kubenswrapper[4978]: fi Feb 25 07:10:35 crc kubenswrapper[4978]: Feb 25 07:10:35 crc kubenswrapper[4978]: # going for maximum compatibility here: Feb 25 07:10:35 crc kubenswrapper[4978]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Feb 25 07:10:35 crc kubenswrapper[4978]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Feb 25 07:10:35 crc kubenswrapper[4978]: # 3. create user with CREATE but then do all password and TLS with ALTER to Feb 25 07:10:35 crc kubenswrapper[4978]: # support updates Feb 25 07:10:35 crc kubenswrapper[4978]: Feb 25 07:10:35 crc kubenswrapper[4978]: $MYSQL_CMD < logger="UnhandledError" Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.700964 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"openstack-mariadb-root-db-secret\\\" not found\"" pod="openstack/root-account-create-update-rzj6b" podUID="6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.710960 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rzj6b"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711005 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3be5007a-36e1-4b55-a8d8-5fd560efb8fe","Type":"ContainerDied","Data":"78414f7a4e1d623e9a11a84d9c8ca28b022bcdb53fed8a932a715712c16abf12"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711037 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"58c891d9-fbca-406c-bd1a-ecb13f9d7444","Type":"ContainerDied","Data":"4b497e57dce20bdb93e7a82daed8baa0e9acd2b6048c9e18954a4341399c8f5b"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711054 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7d28d9f8-de4f-4223-bfa6-fdf398f133bb","Type":"ContainerDied","Data":"231167ca97b4be8ade22723cde0ed17906f6163a3cedd000bd07b8e2dddaaa83"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711068 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwmn" event={"ID":"34752acd-4270-408b-bc82-af9fc515da2b","Type":"ContainerStarted","Data":"9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711086 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8","Type":"ContainerDied","Data":"33f35943890b2d404b619750739644e060d0152191e4ffd0d1bb79cdebfb8e65"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711101 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8a8c-account-create-update-d7wm6" event={"ID":"83bd2445-af69-432a-bc55-d5cee299162d","Type":"ContainerDied","Data":"7c7f63cf88ea4ea3379a36a8d4d157cdf1918c77c3bf5759473a8b46ee2aa4e2"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711115 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9eb7-account-create-update-f2s6t" event={"ID":"c7bd49f1-22d2-4ef1-995a-b98bf91a60ce","Type":"ContainerDied","Data":"a0e1b71f6970f985eaccadb4987a0c7254c6b4f1cb4b8a22c5a89705546dc3b9"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711128 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69794fc6d6-zqnw6" event={"ID":"c6a16d36-c629-483e-9b15-fbd7b8913883","Type":"ContainerDied","Data":"03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711143 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69794fc6d6-zqnw6" event={"ID":"c6a16d36-c629-483e-9b15-fbd7b8913883","Type":"ContainerDied","Data":"52264c07775f9aaabcc27a16a6dbeb0d55deb95b0d2bc1cda2b362cb0db62e1d"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711154 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6ec4-account-create-update-649r6" event={"ID":"84e32f1f-5ad1-40a8-8f61-21025d84d003","Type":"ContainerDied","Data":"36cd594e8e8fab32c5d5c0941858a59337834b174129f331eade37a79dfd42e8"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711166 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-1b7b-account-create-update-qlxzw" event={"ID":"bc147cb6-a5d8-47f7-8ce6-168455f18512","Type":"ContainerDied","Data":"db33e38cd6265a879b5db49dc5a0c9b860d27722dd6b77376d11042cdf69fa62"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711177 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9373-account-create-update-lqszb" event={"ID":"cbba37d2-4c5c-467c-9e1c-90c2c3aea110","Type":"ContainerStarted","Data":"a1d022e266802e5a68762d39becfd60239b0242f1649c9c4be4e18ec3faefe60"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.711188 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77bc5d768-qx88q" event={"ID":"3a2631ab-861d-40eb-9b52-b4fd269eb411","Type":"ContainerDied","Data":"e3bb5987c1f9fda6b9f36d86116e1f3a943d9693a3917d0c5c728a4dd3429744"} Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.779720 4978 scope.go:117] "RemoveContainer" containerID="df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.788955 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.825132 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwr48\" (UniqueName: \"kubernetes.io/projected/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-kube-api-access-wwr48\") pod \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.825250 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-httpd-run\") pod \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.825812 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7d28d9f8-de4f-4223-bfa6-fdf398f133bb" (UID: "7d28d9f8-de4f-4223-bfa6-fdf398f133bb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.826823 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-public-tls-certs\") pod \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.826865 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-scripts\") pod \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.826922 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-combined-ca-bundle\") pod \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.829029 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.830956 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-logs\") pod \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.834738 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.837514 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-logs" (OuterVolumeSpecName: "logs") pod "7d28d9f8-de4f-4223-bfa6-fdf398f133bb" (UID: "7d28d9f8-de4f-4223-bfa6-fdf398f133bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.837588 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-config-data\") pod \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.837634 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\" (UID: \"7d28d9f8-de4f-4223-bfa6-fdf398f133bb\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.838272 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.838282 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.839734 4978 scope.go:117] "RemoveContainer" containerID="3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.840651 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6ec4-account-create-update-649r6"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.844780 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-scripts" (OuterVolumeSpecName: "scripts") pod "7d28d9f8-de4f-4223-bfa6-fdf398f133bb" (UID: "7d28d9f8-de4f-4223-bfa6-fdf398f133bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.850677 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6ec4-account-create-update-649r6"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.863688 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-kube-api-access-wwr48" (OuterVolumeSpecName: "kube-api-access-wwr48") pod "7d28d9f8-de4f-4223-bfa6-fdf398f133bb" (UID: "7d28d9f8-de4f-4223-bfa6-fdf398f133bb"). InnerVolumeSpecName "kube-api-access-wwr48". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.876884 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "7d28d9f8-de4f-4223-bfa6-fdf398f133bb" (UID: "7d28d9f8-de4f-4223-bfa6-fdf398f133bb"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.886874 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9eb7-account-create-update-f2s6t"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.893557 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d28d9f8-de4f-4223-bfa6-fdf398f133bb" (UID: "7d28d9f8-de4f-4223-bfa6-fdf398f133bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.894693 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-9eb7-account-create-update-f2s6t"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.902374 4978 scope.go:117] "RemoveContainer" containerID="df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79" Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.912584 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79\": container with ID starting with df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79 not found: ID does not exist" containerID="df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.912638 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79"} err="failed to get container status \"df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79\": rpc error: code = NotFound desc = could not find container \"df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79\": container with ID starting with df06467441359f0eaaf56e3684db1ca9656d3086146424ec9951c6ba6d673e79 not found: ID does not exist" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.912668 4978 scope.go:117] "RemoveContainer" containerID="3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.914718 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7d28d9f8-de4f-4223-bfa6-fdf398f133bb" (UID: "7d28d9f8-de4f-4223-bfa6-fdf398f133bb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.919241 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:10:35 crc kubenswrapper[4978]: E0225 07:10:35.919308 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658\": container with ID starting with 3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658 not found: ID does not exist" containerID="3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.919331 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658"} err="failed to get container status \"3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658\": rpc error: code = NotFound desc = could not find container \"3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658\": container with ID starting with 3524def883b5b5e97f2760e3c2a9724e786009167672fec1904a75533eb2e658 not found: ID does not exist" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.919350 4978 scope.go:117] "RemoveContainer" containerID="03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.919772 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.928778 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.936441 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8a8c-account-create-update-d7wm6"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.939117 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.939187 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-httpd-run\") pod \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.939229 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-combined-ca-bundle\") pod \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.939766 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-scripts\") pod \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.939812 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-config-data\") pod \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.939851 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-internal-tls-certs\") pod \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.939878 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-logs\") pod \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.939937 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2hkq\" (UniqueName: \"kubernetes.io/projected/58c891d9-fbca-406c-bd1a-ecb13f9d7444-kube-api-access-g2hkq\") pod \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.940573 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.940666 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.940736 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.940818 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.940890 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwr48\" (UniqueName: \"kubernetes.io/projected/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-kube-api-access-wwr48\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.941890 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-logs" (OuterVolumeSpecName: "logs") pod "58c891d9-fbca-406c-bd1a-ecb13f9d7444" (UID: "58c891d9-fbca-406c-bd1a-ecb13f9d7444"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.945550 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "58c891d9-fbca-406c-bd1a-ecb13f9d7444" (UID: "58c891d9-fbca-406c-bd1a-ecb13f9d7444"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.956600 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "58c891d9-fbca-406c-bd1a-ecb13f9d7444" (UID: "58c891d9-fbca-406c-bd1a-ecb13f9d7444"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.957635 4978 scope.go:117] "RemoveContainer" containerID="9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.962060 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58c891d9-fbca-406c-bd1a-ecb13f9d7444-kube-api-access-g2hkq" (OuterVolumeSpecName: "kube-api-access-g2hkq") pod "58c891d9-fbca-406c-bd1a-ecb13f9d7444" (UID: "58c891d9-fbca-406c-bd1a-ecb13f9d7444"). InnerVolumeSpecName "kube-api-access-g2hkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.962603 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8a8c-account-create-update-d7wm6"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.976340 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-scripts" (OuterVolumeSpecName: "scripts") pod "58c891d9-fbca-406c-bd1a-ecb13f9d7444" (UID: "58c891d9-fbca-406c-bd1a-ecb13f9d7444"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.976672 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-69794fc6d6-zqnw6"] Feb 25 07:10:35 crc kubenswrapper[4978]: I0225 07:10:35.976739 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-69794fc6d6-zqnw6"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.059704 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.064970 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58c891d9-fbca-406c-bd1a-ecb13f9d7444" (UID: "58c891d9-fbca-406c-bd1a-ecb13f9d7444"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.066382 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "58c891d9-fbca-406c-bd1a-ecb13f9d7444" (UID: "58c891d9-fbca-406c-bd1a-ecb13f9d7444"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.069602 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-certs\") pod \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093491 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-combined-ca-bundle\") pod \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093554 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-internal-tls-certs\") pod \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093580 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data-custom\") pod \"3a2631ab-861d-40eb-9b52-b4fd269eb411\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093630 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-combined-ca-bundle\") pod \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093652 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-logs\") pod \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093712 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-combined-ca-bundle\") pod \"3a2631ab-861d-40eb-9b52-b4fd269eb411\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093761 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fgrc\" (UniqueName: \"kubernetes.io/projected/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-kube-api-access-6fgrc\") pod \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093809 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-combined-ca-bundle\") pod \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\" (UID: \"58c891d9-fbca-406c-bd1a-ecb13f9d7444\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093829 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2631ab-861d-40eb-9b52-b4fd269eb411-logs\") pod \"3a2631ab-861d-40eb-9b52-b4fd269eb411\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093851 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-config-data\") pod \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093884 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-config\") pod \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093913 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-nova-metadata-tls-certs\") pod \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\" (UID: \"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093943 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data\") pod \"3a2631ab-861d-40eb-9b52-b4fd269eb411\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093960 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-public-tls-certs\") pod \"3a2631ab-861d-40eb-9b52-b4fd269eb411\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.093979 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vsbp\" (UniqueName: \"kubernetes.io/projected/3a2631ab-861d-40eb-9b52-b4fd269eb411-kube-api-access-5vsbp\") pod \"3a2631ab-861d-40eb-9b52-b4fd269eb411\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.094010 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jh5v\" (UniqueName: \"kubernetes.io/projected/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-api-access-9jh5v\") pod \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\" (UID: \"3be5007a-36e1-4b55-a8d8-5fd560efb8fe\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.094033 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-internal-tls-certs\") pod \"3a2631ab-861d-40eb-9b52-b4fd269eb411\" (UID: \"3a2631ab-861d-40eb-9b52-b4fd269eb411\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.094769 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.094787 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.094796 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.094805 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.094814 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58c891d9-fbca-406c-bd1a-ecb13f9d7444-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.094823 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2hkq\" (UniqueName: \"kubernetes.io/projected/58c891d9-fbca-406c-bd1a-ecb13f9d7444-kube-api-access-g2hkq\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.090963 4978 scope.go:117] "RemoveContainer" containerID="03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.076213 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-config-data" (OuterVolumeSpecName: "config-data") pod "58c891d9-fbca-406c-bd1a-ecb13f9d7444" (UID: "58c891d9-fbca-406c-bd1a-ecb13f9d7444"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.092631 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-config-data" (OuterVolumeSpecName: "config-data") pod "7d28d9f8-de4f-4223-bfa6-fdf398f133bb" (UID: "7d28d9f8-de4f-4223-bfa6-fdf398f133bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.098017 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-kube-api-access-6fgrc" (OuterVolumeSpecName: "kube-api-access-6fgrc") pod "bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" (UID: "bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8"). InnerVolumeSpecName "kube-api-access-6fgrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.100881 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3a2631ab-861d-40eb-9b52-b4fd269eb411" (UID: "3a2631ab-861d-40eb-9b52-b4fd269eb411"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.085222 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-1b7b-account-create-update-qlxzw"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.102565 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-1b7b-account-create-update-qlxzw"] Feb 25 07:10:36 crc kubenswrapper[4978]: W0225 07:10:36.102653 4978 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/58c891d9-fbca-406c-bd1a-ecb13f9d7444/volumes/kubernetes.io~secret/combined-ca-bundle Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.102666 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58c891d9-fbca-406c-bd1a-ecb13f9d7444" (UID: "58c891d9-fbca-406c-bd1a-ecb13f9d7444"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.102921 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a2631ab-861d-40eb-9b52-b4fd269eb411-logs" (OuterVolumeSpecName: "logs") pod "3a2631ab-861d-40eb-9b52-b4fd269eb411" (UID: "3a2631ab-861d-40eb-9b52-b4fd269eb411"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: W0225 07:10:36.103058 4978 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/58c891d9-fbca-406c-bd1a-ecb13f9d7444/volumes/kubernetes.io~secret/internal-tls-certs Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.103078 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "58c891d9-fbca-406c-bd1a-ecb13f9d7444" (UID: "58c891d9-fbca-406c-bd1a-ecb13f9d7444"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.104205 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-logs" (OuterVolumeSpecName: "logs") pod "bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" (UID: "bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: E0225 07:10:36.104306 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6\": container with ID starting with 03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6 not found: ID does not exist" containerID="03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.104332 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6"} err="failed to get container status \"03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6\": rpc error: code = NotFound desc = could not find container \"03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6\": container with ID starting with 03496bbb07e2183020b89f48955d7905971b9df0fc1c229d8598bf4229772cf6 not found: ID does not exist" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.104356 4978 scope.go:117] "RemoveContainer" containerID="9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.118981 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a2631ab-861d-40eb-9b52-b4fd269eb411-kube-api-access-5vsbp" (OuterVolumeSpecName: "kube-api-access-5vsbp") pod "3a2631ab-861d-40eb-9b52-b4fd269eb411" (UID: "3a2631ab-861d-40eb-9b52-b4fd269eb411"). InnerVolumeSpecName "kube-api-access-5vsbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: E0225 07:10:36.119106 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af\": container with ID starting with 9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af not found: ID does not exist" containerID="9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.119144 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af"} err="failed to get container status \"9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af\": rpc error: code = NotFound desc = could not find container \"9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af\": container with ID starting with 9c98f0be62902c84adea45223547d430f32cd021e3ce10adbf9ebff6ea2217af not found: ID does not exist" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.121812 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-api-access-9jh5v" (OuterVolumeSpecName: "kube-api-access-9jh5v") pod "3be5007a-36e1-4b55-a8d8-5fd560efb8fe" (UID: "3be5007a-36e1-4b55-a8d8-5fd560efb8fe"). InnerVolumeSpecName "kube-api-access-9jh5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.126333 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.165656 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" (UID: "bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.180856 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9373-account-create-update-lqszb" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198288 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vsbp\" (UniqueName: \"kubernetes.io/projected/3a2631ab-861d-40eb-9b52-b4fd269eb411-kube-api-access-5vsbp\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198412 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jh5v\" (UniqueName: \"kubernetes.io/projected/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-api-access-9jh5v\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198485 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198548 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d28d9f8-de4f-4223-bfa6-fdf398f133bb-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198622 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198678 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198729 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198780 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198837 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198888 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fgrc\" (UniqueName: \"kubernetes.io/projected/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-kube-api-access-6fgrc\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198938 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58c891d9-fbca-406c-bd1a-ecb13f9d7444-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.198987 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a2631ab-861d-40eb-9b52-b4fd269eb411-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.267440 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.278061 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.299346 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "3be5007a-36e1-4b55-a8d8-5fd560efb8fe" (UID: "3be5007a-36e1-4b55-a8d8-5fd560efb8fe"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.299825 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-operator-scripts\") pod \"cbba37d2-4c5c-467c-9e1c-90c2c3aea110\" (UID: \"cbba37d2-4c5c-467c-9e1c-90c2c3aea110\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.299944 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whfrj\" (UniqueName: \"kubernetes.io/projected/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-kube-api-access-whfrj\") pod \"cbba37d2-4c5c-467c-9e1c-90c2c3aea110\" (UID: \"cbba37d2-4c5c-467c-9e1c-90c2c3aea110\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.301409 4978 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.301923 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cbba37d2-4c5c-467c-9e1c-90c2c3aea110" (UID: "cbba37d2-4c5c-467c-9e1c-90c2c3aea110"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.327318 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-kube-api-access-whfrj" (OuterVolumeSpecName: "kube-api-access-whfrj") pod "cbba37d2-4c5c-467c-9e1c-90c2c3aea110" (UID: "cbba37d2-4c5c-467c-9e1c-90c2c3aea110"). InnerVolumeSpecName "kube-api-access-whfrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.361190 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a2631ab-861d-40eb-9b52-b4fd269eb411" (UID: "3a2631ab-861d-40eb-9b52-b4fd269eb411"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.380175 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3a2631ab-861d-40eb-9b52-b4fd269eb411" (UID: "3a2631ab-861d-40eb-9b52-b4fd269eb411"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.380556 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-config-data" (OuterVolumeSpecName: "config-data") pod "bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" (UID: "bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.397842 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "3be5007a-36e1-4b55-a8d8-5fd560efb8fe" (UID: "3be5007a-36e1-4b55-a8d8-5fd560efb8fe"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.399743 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3be5007a-36e1-4b55-a8d8-5fd560efb8fe" (UID: "3be5007a-36e1-4b55-a8d8-5fd560efb8fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.406958 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-combined-ca-bundle\") pod \"50c62987-be2b-4deb-a88c-107db5d00415\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.407043 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75l5j\" (UniqueName: \"kubernetes.io/projected/1ddb624b-1c38-44b8-8f02-bd177937ef4d-kube-api-access-75l5j\") pod \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.407067 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-combined-ca-bundle\") pod \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.407111 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-config-data\") pod \"50c62987-be2b-4deb-a88c-107db5d00415\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.407406 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-scripts\") pod \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.407437 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-config-data\") pod \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.407536 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-run-httpd\") pod \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.407665 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt5fk\" (UniqueName: \"kubernetes.io/projected/50c62987-be2b-4deb-a88c-107db5d00415-kube-api-access-tt5fk\") pod \"50c62987-be2b-4deb-a88c-107db5d00415\" (UID: \"50c62987-be2b-4deb-a88c-107db5d00415\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.407750 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-sg-core-conf-yaml\") pod \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.407790 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-log-httpd\") pod \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.407880 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-ceilometer-tls-certs\") pod \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\" (UID: \"1ddb624b-1c38-44b8-8f02-bd177937ef4d\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.408355 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whfrj\" (UniqueName: \"kubernetes.io/projected/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-kube-api-access-whfrj\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.408407 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.408419 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.408427 4978 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.408437 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3be5007a-36e1-4b55-a8d8-5fd560efb8fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.408446 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbba37d2-4c5c-467c-9e1c-90c2c3aea110-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.408548 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.410560 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3a2631ab-861d-40eb-9b52-b4fd269eb411" (UID: "3a2631ab-861d-40eb-9b52-b4fd269eb411"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.412997 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1ddb624b-1c38-44b8-8f02-bd177937ef4d" (UID: "1ddb624b-1c38-44b8-8f02-bd177937ef4d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.415317 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1ddb624b-1c38-44b8-8f02-bd177937ef4d" (UID: "1ddb624b-1c38-44b8-8f02-bd177937ef4d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.417119 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ddb624b-1c38-44b8-8f02-bd177937ef4d-kube-api-access-75l5j" (OuterVolumeSpecName: "kube-api-access-75l5j") pod "1ddb624b-1c38-44b8-8f02-bd177937ef4d" (UID: "1ddb624b-1c38-44b8-8f02-bd177937ef4d"). InnerVolumeSpecName "kube-api-access-75l5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.424024 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50c62987-be2b-4deb-a88c-107db5d00415-kube-api-access-tt5fk" (OuterVolumeSpecName: "kube-api-access-tt5fk") pod "50c62987-be2b-4deb-a88c-107db5d00415" (UID: "50c62987-be2b-4deb-a88c-107db5d00415"). InnerVolumeSpecName "kube-api-access-tt5fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.426493 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-scripts" (OuterVolumeSpecName: "scripts") pod "1ddb624b-1c38-44b8-8f02-bd177937ef4d" (UID: "1ddb624b-1c38-44b8-8f02-bd177937ef4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.444910 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data" (OuterVolumeSpecName: "config-data") pod "3a2631ab-861d-40eb-9b52-b4fd269eb411" (UID: "3a2631ab-861d-40eb-9b52-b4fd269eb411"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.471569 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50c62987-be2b-4deb-a88c-107db5d00415" (UID: "50c62987-be2b-4deb-a88c-107db5d00415"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.472097 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-config-data" (OuterVolumeSpecName: "config-data") pod "50c62987-be2b-4deb-a88c-107db5d00415" (UID: "50c62987-be2b-4deb-a88c-107db5d00415"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.474468 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1ddb624b-1c38-44b8-8f02-bd177937ef4d" (UID: "1ddb624b-1c38-44b8-8f02-bd177937ef4d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.519569 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.519605 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.519628 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a2631ab-861d-40eb-9b52-b4fd269eb411-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.519639 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.519649 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt5fk\" (UniqueName: \"kubernetes.io/projected/50c62987-be2b-4deb-a88c-107db5d00415-kube-api-access-tt5fk\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.519658 4978 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.519670 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1ddb624b-1c38-44b8-8f02-bd177937ef4d-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.519679 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.519689 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75l5j\" (UniqueName: \"kubernetes.io/projected/1ddb624b-1c38-44b8-8f02-bd177937ef4d-kube-api-access-75l5j\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.519701 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50c62987-be2b-4deb-a88c-107db5d00415-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.524694 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8","Type":"ContainerDied","Data":"4f341182a2e529d3fb3a7105185d2c313bf7306e97ceb3b01d58a2bf81fcb51a"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.524753 4978 scope.go:117] "RemoveContainer" containerID="33f35943890b2d404b619750739644e060d0152191e4ffd0d1bb79cdebfb8e65" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.524943 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.542349 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9373-account-create-update-lqszb" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.542680 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9373-account-create-update-lqszb" event={"ID":"cbba37d2-4c5c-467c-9e1c-90c2c3aea110","Type":"ContainerDied","Data":"a1d022e266802e5a68762d39becfd60239b0242f1649c9c4be4e18ec3faefe60"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.559866 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-77bc5d768-qx88q" event={"ID":"3a2631ab-861d-40eb-9b52-b4fd269eb411","Type":"ContainerDied","Data":"3eda4c956f7bfa3516a7c81625512fe074672ee19730bb4f83f588aa74527021"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.560432 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-77bc5d768-qx88q" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.560848 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1ddb624b-1c38-44b8-8f02-bd177937ef4d" (UID: "1ddb624b-1c38-44b8-8f02-bd177937ef4d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.561743 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rzj6b" event={"ID":"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c","Type":"ContainerStarted","Data":"b74a748b253680333605d35b10a423526a41b81f2dcdc68c37a40527bc86cb3e"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.610932 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ddb624b-1c38-44b8-8f02-bd177937ef4d" (UID: "1ddb624b-1c38-44b8-8f02-bd177937ef4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.615856 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"7d28d9f8-de4f-4223-bfa6-fdf398f133bb","Type":"ContainerDied","Data":"29b79f431130815c4fa90c1335e3ead67afd184fefc3693c8996018550e93ddb"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.616013 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.620896 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.620929 4978 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.632494 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.632564 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3be5007a-36e1-4b55-a8d8-5fd560efb8fe","Type":"ContainerDied","Data":"9a69be620040b4f2389b269e72f36725d52d5fe065d63127c807f79176956fa0"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.644311 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.644327 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"58c891d9-fbca-406c-bd1a-ecb13f9d7444","Type":"ContainerDied","Data":"591db7360b851f95617b4ff54b6a72b81df0b8fd45afff36b1f06a0c114de670"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.649193 4978 generic.go:334] "Generic (PLEG): container finished" podID="2e39fd28-0520-44bc-9486-a118cec688c0" containerID="ed748fe64533cd208cfaa954b8175f2c179485a3b36021fb250466d83a45c91e" exitCode=0 Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.649245 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e39fd28-0520-44bc-9486-a118cec688c0","Type":"ContainerDied","Data":"ed748fe64533cd208cfaa954b8175f2c179485a3b36021fb250466d83a45c91e"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.650677 4978 generic.go:334] "Generic (PLEG): container finished" podID="50c62987-be2b-4deb-a88c-107db5d00415" containerID="91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f" exitCode=0 Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.650716 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"50c62987-be2b-4deb-a88c-107db5d00415","Type":"ContainerDied","Data":"91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.650734 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"50c62987-be2b-4deb-a88c-107db5d00415","Type":"ContainerDied","Data":"dab00b558315dd670ef2dc1afe389f10485644ac083dd18bde12f210ae9add55"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.650785 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.662629 4978 generic.go:334] "Generic (PLEG): container finished" podID="85624b7d-192e-42c3-9c5b-18cfd91d37e9" containerID="22db8e67ef3c374833a5a7d3778dc19f3cb69c3620c3009d7badbe8c31ef7006" exitCode=0 Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.662691 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"85624b7d-192e-42c3-9c5b-18cfd91d37e9","Type":"ContainerDied","Data":"22db8e67ef3c374833a5a7d3778dc19f3cb69c3620c3009d7badbe8c31ef7006"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.662708 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"85624b7d-192e-42c3-9c5b-18cfd91d37e9","Type":"ContainerDied","Data":"17df30ac7bbc09d7f30340ec5bcdd268698d7d872757d6efefa7590e58e98dd9"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.662719 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17df30ac7bbc09d7f30340ec5bcdd268698d7d872757d6efefa7590e58e98dd9" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.664974 4978 generic.go:334] "Generic (PLEG): container finished" podID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerID="48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae" exitCode=0 Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.664988 4978 generic.go:334] "Generic (PLEG): container finished" podID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerID="a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a" exitCode=0 Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.665012 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ddb624b-1c38-44b8-8f02-bd177937ef4d","Type":"ContainerDied","Data":"48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.665027 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ddb624b-1c38-44b8-8f02-bd177937ef4d","Type":"ContainerDied","Data":"a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.665036 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1ddb624b-1c38-44b8-8f02-bd177937ef4d","Type":"ContainerDied","Data":"48b11d3570e4f26f90f450bca7435c16a6f6a7b6b849ce2e5353a33ace7745e9"} Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.665111 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.672562 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.706658 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.722672 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts\") pod \"keystone-56f1-account-create-update-dvwqn\" (UID: \"4e53c877-f109-4d98-9388-25dccf9bd229\") " pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.722763 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcnqg\" (UniqueName: \"kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg\") pod \"keystone-56f1-account-create-update-dvwqn\" (UID: \"4e53c877-f109-4d98-9388-25dccf9bd229\") " pod="openstack/keystone-56f1-account-create-update-dvwqn" Feb 25 07:10:36 crc kubenswrapper[4978]: E0225 07:10:36.723493 4978 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Feb 25 07:10:36 crc kubenswrapper[4978]: E0225 07:10:36.723568 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts podName:4e53c877-f109-4d98-9388-25dccf9bd229 nodeName:}" failed. No retries permitted until 2026-02-25 07:10:38.723548251 +0000 UTC m=+1532.162804710 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts") pod "keystone-56f1-account-create-update-dvwqn" (UID: "4e53c877-f109-4d98-9388-25dccf9bd229") : configmap "openstack-scripts" not found Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.728154 4978 scope.go:117] "RemoveContainer" containerID="2881499eed920adb23f199015c996a12e05fc38be2beb2a1bf0cc87bbbb6b34a" Feb 25 07:10:36 crc kubenswrapper[4978]: E0225 07:10:36.731228 4978 projected.go:194] Error preparing data for projected volume kube-api-access-rcnqg for pod openstack/keystone-56f1-account-create-update-dvwqn: failed to fetch token: serviceaccounts "galera-openstack" not found Feb 25 07:10:36 crc kubenswrapper[4978]: E0225 07:10:36.731294 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg podName:4e53c877-f109-4d98-9388-25dccf9bd229 nodeName:}" failed. No retries permitted until 2026-02-25 07:10:38.731274659 +0000 UTC m=+1532.170531118 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-rcnqg" (UniqueName: "kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg") pod "keystone-56f1-account-create-update-dvwqn" (UID: "4e53c877-f109-4d98-9388-25dccf9bd229") : failed to fetch token: serviceaccounts "galera-openstack" not found Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.734519 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-config-data" (OuterVolumeSpecName: "config-data") pod "1ddb624b-1c38-44b8-8f02-bd177937ef4d" (UID: "1ddb624b-1c38-44b8-8f02-bd177937ef4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.746237 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" (UID: "bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: E0225 07:10:36.747137 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Feb 25 07:10:36 crc kubenswrapper[4978]: E0225 07:10:36.753226 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Feb 25 07:10:36 crc kubenswrapper[4978]: E0225 07:10:36.756376 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Feb 25 07:10:36 crc kubenswrapper[4978]: E0225 07:10:36.756422 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerName="ovn-northd" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.767924 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.770720 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.786495 4978 scope.go:117] "RemoveContainer" containerID="e3bb5987c1f9fda6b9f36d86116e1f3a943d9693a3917d0c5c728a4dd3429744" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.802513 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.811981 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-77bc5d768-qx88q"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.821488 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-77bc5d768-qx88q"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.823407 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-memcached-tls-certs\") pod \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.823439 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-combined-ca-bundle\") pod \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.823490 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kolla-config\") pod \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.823527 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwgww\" (UniqueName: \"kubernetes.io/projected/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kube-api-access-fwgww\") pod \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.823612 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-config-data\") pod \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\" (UID: \"85624b7d-192e-42c3-9c5b-18cfd91d37e9\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.823959 4978 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.823978 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ddb624b-1c38-44b8-8f02-bd177937ef4d-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.824320 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-config-data" (OuterVolumeSpecName: "config-data") pod "85624b7d-192e-42c3-9c5b-18cfd91d37e9" (UID: "85624b7d-192e-42c3-9c5b-18cfd91d37e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.824792 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "85624b7d-192e-42c3-9c5b-18cfd91d37e9" (UID: "85624b7d-192e-42c3-9c5b-18cfd91d37e9"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.827610 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kube-api-access-fwgww" (OuterVolumeSpecName: "kube-api-access-fwgww") pod "85624b7d-192e-42c3-9c5b-18cfd91d37e9" (UID: "85624b7d-192e-42c3-9c5b-18cfd91d37e9"). InnerVolumeSpecName "kube-api-access-fwgww". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.853551 4978 scope.go:117] "RemoveContainer" containerID="3d765602f0864d6d998940f94e6b40eaa60dbf1a6c6bdb9427a47ca7993400a3" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.858460 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-56f1-account-create-update-dvwqn"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.867520 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85624b7d-192e-42c3-9c5b-18cfd91d37e9" (UID: "85624b7d-192e-42c3-9c5b-18cfd91d37e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.876893 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "85624b7d-192e-42c3-9c5b-18cfd91d37e9" (UID: "85624b7d-192e-42c3-9c5b-18cfd91d37e9"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.888429 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-56f1-account-create-update-dvwqn"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.890508 4978 scope.go:117] "RemoveContainer" containerID="231167ca97b4be8ade22723cde0ed17906f6163a3cedd000bd07b8e2dddaaa83" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.917817 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.922608 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.925504 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-combined-ca-bundle\") pod \"2e39fd28-0520-44bc-9486-a118cec688c0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.925577 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9fp8\" (UniqueName: \"kubernetes.io/projected/2e39fd28-0520-44bc-9486-a118cec688c0-kube-api-access-s9fp8\") pod \"2e39fd28-0520-44bc-9486-a118cec688c0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.932640 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-config-data\") pod \"2e39fd28-0520-44bc-9486-a118cec688c0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.932670 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-internal-tls-certs\") pod \"2e39fd28-0520-44bc-9486-a118cec688c0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.932754 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e39fd28-0520-44bc-9486-a118cec688c0-logs\") pod \"2e39fd28-0520-44bc-9486-a118cec688c0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.932822 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-public-tls-certs\") pod \"2e39fd28-0520-44bc-9486-a118cec688c0\" (UID: \"2e39fd28-0520-44bc-9486-a118cec688c0\") " Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.933328 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwgww\" (UniqueName: \"kubernetes.io/projected/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kube-api-access-fwgww\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.933345 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.933356 4978 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.933387 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85624b7d-192e-42c3-9c5b-18cfd91d37e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.933399 4978 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/85624b7d-192e-42c3-9c5b-18cfd91d37e9-kolla-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.934671 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e39fd28-0520-44bc-9486-a118cec688c0-logs" (OuterVolumeSpecName: "logs") pod "2e39fd28-0520-44bc-9486-a118cec688c0" (UID: "2e39fd28-0520-44bc-9486-a118cec688c0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.938474 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e39fd28-0520-44bc-9486-a118cec688c0-kube-api-access-s9fp8" (OuterVolumeSpecName: "kube-api-access-s9fp8") pod "2e39fd28-0520-44bc-9486-a118cec688c0" (UID: "2e39fd28-0520-44bc-9486-a118cec688c0"). InnerVolumeSpecName "kube-api-access-s9fp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.941713 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.952923 4978 scope.go:117] "RemoveContainer" containerID="251a79da9fba47a65d4db2f16a1ee10da3ad7f02ffcf859eb40da544a9209781" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.953049 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.962344 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.979657 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.982727 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e39fd28-0520-44bc-9486-a118cec688c0" (UID: "2e39fd28-0520-44bc-9486-a118cec688c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.982756 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-config-data" (OuterVolumeSpecName: "config-data") pod "2e39fd28-0520-44bc-9486-a118cec688c0" (UID: "2e39fd28-0520-44bc-9486-a118cec688c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.984650 4978 scope.go:117] "RemoveContainer" containerID="78414f7a4e1d623e9a11a84d9c8ca28b022bcdb53fed8a932a715712c16abf12" Feb 25 07:10:36 crc kubenswrapper[4978]: I0225 07:10:36.989192 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2e39fd28-0520-44bc-9486-a118cec688c0" (UID: "2e39fd28-0520-44bc-9486-a118cec688c0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.009158 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-9373-account-create-update-lqszb"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.014546 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2e39fd28-0520-44bc-9486-a118cec688c0" (UID: "2e39fd28-0520-44bc-9486-a118cec688c0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.016475 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-9373-account-create-update-lqszb"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.029794 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.036766 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9fp8\" (UniqueName: \"kubernetes.io/projected/2e39fd28-0520-44bc-9486-a118cec688c0-kube-api-access-s9fp8\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.036812 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.036822 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.036832 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e53c877-f109-4d98-9388-25dccf9bd229-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.036840 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e39fd28-0520-44bc-9486-a118cec688c0-logs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.036848 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcnqg\" (UniqueName: \"kubernetes.io/projected/4e53c877-f109-4d98-9388-25dccf9bd229-kube-api-access-rcnqg\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.036856 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.036864 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e39fd28-0520-44bc-9486-a118cec688c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.037070 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.043130 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.048429 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rzj6b" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.050462 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.071331 4978 scope.go:117] "RemoveContainer" containerID="4b497e57dce20bdb93e7a82daed8baa0e9acd2b6048c9e18954a4341399c8f5b" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.108596 4978 scope.go:117] "RemoveContainer" containerID="23b908a2d2c8c5b19239c4d15a74e513b855d7ec6155c2aab43b358150b36fe5" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.137873 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9bx\" (UniqueName: \"kubernetes.io/projected/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-kube-api-access-2w9bx\") pod \"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c\" (UID: \"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c\") " Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.138101 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-operator-scripts\") pod \"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c\" (UID: \"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c\") " Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.138601 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c" (UID: "6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.138757 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.141278 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-kube-api-access-2w9bx" (OuterVolumeSpecName: "kube-api-access-2w9bx") pod "6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c" (UID: "6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c"). InnerVolumeSpecName "kube-api-access-2w9bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.144862 4978 scope.go:117] "RemoveContainer" containerID="91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.166209 4978 scope.go:117] "RemoveContainer" containerID="91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.167652 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f\": container with ID starting with 91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f not found: ID does not exist" containerID="91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.167684 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f"} err="failed to get container status \"91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f\": rpc error: code = NotFound desc = could not find container \"91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f\": container with ID starting with 91d10be650c667174d1fc7cc21dea01e4348d1bf4f3471538d418268532c651f not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.167710 4978 scope.go:117] "RemoveContainer" containerID="f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.193766 4978 scope.go:117] "RemoveContainer" containerID="887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.212807 4978 scope.go:117] "RemoveContainer" containerID="48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.240195 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9bx\" (UniqueName: \"kubernetes.io/projected/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c-kube-api-access-2w9bx\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.247552 4978 scope.go:117] "RemoveContainer" containerID="a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.263200 4978 scope.go:117] "RemoveContainer" containerID="f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.263509 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b\": container with ID starting with f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b not found: ID does not exist" containerID="f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.263542 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b"} err="failed to get container status \"f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b\": rpc error: code = NotFound desc = could not find container \"f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b\": container with ID starting with f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.263570 4978 scope.go:117] "RemoveContainer" containerID="887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.263802 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d\": container with ID starting with 887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d not found: ID does not exist" containerID="887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.263832 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d"} err="failed to get container status \"887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d\": rpc error: code = NotFound desc = could not find container \"887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d\": container with ID starting with 887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.263853 4978 scope.go:117] "RemoveContainer" containerID="48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.264095 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae\": container with ID starting with 48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae not found: ID does not exist" containerID="48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.264134 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae"} err="failed to get container status \"48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae\": rpc error: code = NotFound desc = could not find container \"48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae\": container with ID starting with 48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.264165 4978 scope.go:117] "RemoveContainer" containerID="a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.264404 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a\": container with ID starting with a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a not found: ID does not exist" containerID="a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.264430 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a"} err="failed to get container status \"a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a\": rpc error: code = NotFound desc = could not find container \"a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a\": container with ID starting with a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.264444 4978 scope.go:117] "RemoveContainer" containerID="f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.264630 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b"} err="failed to get container status \"f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b\": rpc error: code = NotFound desc = could not find container \"f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b\": container with ID starting with f5cc0e670855e645135755c17b160296b17619f7ad70bce775af93e0b629f54b not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.264645 4978 scope.go:117] "RemoveContainer" containerID="887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.265265 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d"} err="failed to get container status \"887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d\": rpc error: code = NotFound desc = could not find container \"887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d\": container with ID starting with 887e84da046c26974a6dd0fb06712f14668cbb12f15a48f8c7d10e332e82c84d not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.265286 4978 scope.go:117] "RemoveContainer" containerID="48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.265502 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae"} err="failed to get container status \"48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae\": rpc error: code = NotFound desc = could not find container \"48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae\": container with ID starting with 48cd8433c32caf1a4ad8a097b161d38d2ad9b52f269336e1caebc894b4be47ae not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.265522 4978 scope.go:117] "RemoveContainer" containerID="a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.265753 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a"} err="failed to get container status \"a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a\": rpc error: code = NotFound desc = could not find container \"a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a\": container with ID starting with a2f5c23a5fbdc598ad8cf280ad0b75ac7bf0d809b35010e4ec9f4125adbc170a not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.316617 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_27864863-0a60-43ec-ac3e-f86d9f68c7d2/ovn-northd/0.log" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.316687 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.351203 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.355230 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.360326 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.360560 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="8a6b84e2-353d-47ef-82be-c7e182533a57" containerName="nova-cell1-conductor-conductor" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.378578 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" path="/var/lib/kubelet/pods/1ddb624b-1c38-44b8-8f02-bd177937ef4d/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.379359 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" path="/var/lib/kubelet/pods/3a2631ab-861d-40eb-9b52-b4fd269eb411/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.379946 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3be5007a-36e1-4b55-a8d8-5fd560efb8fe" path="/var/lib/kubelet/pods/3be5007a-36e1-4b55-a8d8-5fd560efb8fe/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.381254 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e53c877-f109-4d98-9388-25dccf9bd229" path="/var/lib/kubelet/pods/4e53c877-f109-4d98-9388-25dccf9bd229/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.381773 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50c62987-be2b-4deb-a88c-107db5d00415" path="/var/lib/kubelet/pods/50c62987-be2b-4deb-a88c-107db5d00415/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.382849 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" path="/var/lib/kubelet/pods/58c891d9-fbca-406c-bd1a-ecb13f9d7444/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.384758 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" path="/var/lib/kubelet/pods/7d28d9f8-de4f-4223-bfa6-fdf398f133bb/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.385835 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83bd2445-af69-432a-bc55-d5cee299162d" path="/var/lib/kubelet/pods/83bd2445-af69-432a-bc55-d5cee299162d/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.386865 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84e32f1f-5ad1-40a8-8f61-21025d84d003" path="/var/lib/kubelet/pods/84e32f1f-5ad1-40a8-8f61-21025d84d003/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.387297 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc147cb6-a5d8-47f7-8ce6-168455f18512" path="/var/lib/kubelet/pods/bc147cb6-a5d8-47f7-8ce6-168455f18512/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.388449 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" path="/var/lib/kubelet/pods/bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.389314 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6a16d36-c629-483e-9b15-fbd7b8913883" path="/var/lib/kubelet/pods/c6a16d36-c629-483e-9b15-fbd7b8913883/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.390028 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7bd49f1-22d2-4ef1-995a-b98bf91a60ce" path="/var/lib/kubelet/pods/c7bd49f1-22d2-4ef1-995a-b98bf91a60ce/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.390731 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbba37d2-4c5c-467c-9e1c-90c2c3aea110" path="/var/lib/kubelet/pods/cbba37d2-4c5c-467c-9e1c-90c2c3aea110/volumes" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.457607 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-metrics-certs-tls-certs\") pod \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.457885 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-northd-tls-certs\") pod \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.458044 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-scripts\") pod \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.458139 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-config\") pod \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.458246 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-combined-ca-bundle\") pod \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.458423 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg4nd\" (UniqueName: \"kubernetes.io/projected/27864863-0a60-43ec-ac3e-f86d9f68c7d2-kube-api-access-fg4nd\") pod \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.458577 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-rundir\") pod \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\" (UID: \"27864863-0a60-43ec-ac3e-f86d9f68c7d2\") " Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.459406 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "27864863-0a60-43ec-ac3e-f86d9f68c7d2" (UID: "27864863-0a60-43ec-ac3e-f86d9f68c7d2"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.460608 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-config" (OuterVolumeSpecName: "config") pod "27864863-0a60-43ec-ac3e-f86d9f68c7d2" (UID: "27864863-0a60-43ec-ac3e-f86d9f68c7d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.460939 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-scripts" (OuterVolumeSpecName: "scripts") pod "27864863-0a60-43ec-ac3e-f86d9f68c7d2" (UID: "27864863-0a60-43ec-ac3e-f86d9f68c7d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.479917 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27864863-0a60-43ec-ac3e-f86d9f68c7d2-kube-api-access-fg4nd" (OuterVolumeSpecName: "kube-api-access-fg4nd") pod "27864863-0a60-43ec-ac3e-f86d9f68c7d2" (UID: "27864863-0a60-43ec-ac3e-f86d9f68c7d2"). InnerVolumeSpecName "kube-api-access-fg4nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.489300 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27864863-0a60-43ec-ac3e-f86d9f68c7d2" (UID: "27864863-0a60-43ec-ac3e-f86d9f68c7d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.543486 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "27864863-0a60-43ec-ac3e-f86d9f68c7d2" (UID: "27864863-0a60-43ec-ac3e-f86d9f68c7d2"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.568652 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-rundir\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.568786 4978 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.568860 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.568935 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27864863-0a60-43ec-ac3e-f86d9f68c7d2-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.569007 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.569086 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg4nd\" (UniqueName: \"kubernetes.io/projected/27864863-0a60-43ec-ac3e-f86d9f68c7d2-kube-api-access-fg4nd\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.620450 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "27864863-0a60-43ec-ac3e-f86d9f68c7d2" (UID: "27864863-0a60-43ec-ac3e-f86d9f68c7d2"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.670292 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/27864863-0a60-43ec-ac3e-f86d9f68c7d2-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.681479 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-rzj6b" event={"ID":"6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c","Type":"ContainerDied","Data":"b74a748b253680333605d35b10a423526a41b81f2dcdc68c37a40527bc86cb3e"} Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.682023 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-rzj6b" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.686507 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e39fd28-0520-44bc-9486-a118cec688c0","Type":"ContainerDied","Data":"580286c6c360cd5f8d7a3305e3d16a1600058344e7c9c9a4bcfa71811e035a60"} Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.686538 4978 scope.go:117] "RemoveContainer" containerID="ed748fe64533cd208cfaa954b8175f2c179485a3b36021fb250466d83a45c91e" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.686864 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.701137 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_27864863-0a60-43ec-ac3e-f86d9f68c7d2/ovn-northd/0.log" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.701182 4978 generic.go:334] "Generic (PLEG): container finished" podID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerID="6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98" exitCode=139 Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.701262 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.703273 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.705390 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"27864863-0a60-43ec-ac3e-f86d9f68c7d2","Type":"ContainerDied","Data":"6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98"} Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.705424 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"27864863-0a60-43ec-ac3e-f86d9f68c7d2","Type":"ContainerDied","Data":"b2bcc3e0dfa7ab325a01e93711b4a10efbcd7ee43e7e9c52abf8b0b46867fb7e"} Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.715426 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.726060 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.728340 4978 scope.go:117] "RemoveContainer" containerID="d48b870ae7f0075341f3f007177eb69704874efaa7faac91545002adf5db1de3" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.739143 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod315eefbb_24fb_4e45_97ed_03caf894bc08.slice/crio-4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f.scope\": RecentStats: unable to find data in memory cache]" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.752090 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-rzj6b"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.757347 4978 scope.go:117] "RemoveContainer" containerID="c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.758238 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-rzj6b"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.769036 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.775410 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.779549 4978 scope.go:117] "RemoveContainer" containerID="6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.783408 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.784261 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.820667 4978 scope.go:117] "RemoveContainer" containerID="c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.822875 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d\": container with ID starting with c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d not found: ID does not exist" containerID="c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.822918 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d"} err="failed to get container status \"c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d\": rpc error: code = NotFound desc = could not find container \"c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d\": container with ID starting with c44602afd36c84258e6c3ac446c3bb9f0c98fe65f51014a3bb2f3dda47babf2d not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.822945 4978 scope.go:117] "RemoveContainer" containerID="6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.823510 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98\": container with ID starting with 6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98 not found: ID does not exist" containerID="6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98" Feb 25 07:10:37 crc kubenswrapper[4978]: I0225 07:10:37.823567 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98"} err="failed to get container status \"6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98\": rpc error: code = NotFound desc = could not find container \"6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98\": container with ID starting with 6395f9fb745820bb943d05c805bc9e2984445470da7c636faf395a644f4cac98 not found: ID does not exist" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.954830 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.955158 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.955409 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.955441 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.958603 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.959984 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.961150 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.961213 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovs-vswitchd" Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.973761 4978 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Feb 25 07:10:37 crc kubenswrapper[4978]: E0225 07:10:37.973824 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data podName:09112202-a3ef-4742-bdbe-b46c1c9a55ba nodeName:}" failed. No retries permitted until 2026-02-25 07:10:45.973808947 +0000 UTC m=+1539.413065406 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data") pod "rabbitmq-server-0" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba") : configmap "rabbitmq-config-data" not found Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.052627 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 25 07:10:38 crc kubenswrapper[4978]: E0225 07:10:38.094116 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:10:38 crc kubenswrapper[4978]: E0225 07:10:38.095244 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:10:38 crc kubenswrapper[4978]: E0225 07:10:38.096395 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 07:10:38 crc kubenswrapper[4978]: E0225 07:10:38.096421 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7456f008-69db-4a20-86c4-2ab52eb93aa7" containerName="nova-scheduler-scheduler" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.180803 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-default\") pod \"315eefbb-24fb-4e45-97ed-03caf894bc08\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.180884 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-galera-tls-certs\") pod \"315eefbb-24fb-4e45-97ed-03caf894bc08\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.180910 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7gmq\" (UniqueName: \"kubernetes.io/projected/315eefbb-24fb-4e45-97ed-03caf894bc08-kube-api-access-t7gmq\") pod \"315eefbb-24fb-4e45-97ed-03caf894bc08\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.180945 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-operator-scripts\") pod \"315eefbb-24fb-4e45-97ed-03caf894bc08\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.181016 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"315eefbb-24fb-4e45-97ed-03caf894bc08\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.181051 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-kolla-config\") pod \"315eefbb-24fb-4e45-97ed-03caf894bc08\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.181070 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-combined-ca-bundle\") pod \"315eefbb-24fb-4e45-97ed-03caf894bc08\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.181090 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-generated\") pod \"315eefbb-24fb-4e45-97ed-03caf894bc08\" (UID: \"315eefbb-24fb-4e45-97ed-03caf894bc08\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.182407 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "315eefbb-24fb-4e45-97ed-03caf894bc08" (UID: "315eefbb-24fb-4e45-97ed-03caf894bc08"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.182572 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "315eefbb-24fb-4e45-97ed-03caf894bc08" (UID: "315eefbb-24fb-4e45-97ed-03caf894bc08"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.182773 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "315eefbb-24fb-4e45-97ed-03caf894bc08" (UID: "315eefbb-24fb-4e45-97ed-03caf894bc08"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.183478 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "315eefbb-24fb-4e45-97ed-03caf894bc08" (UID: "315eefbb-24fb-4e45-97ed-03caf894bc08"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.185824 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/315eefbb-24fb-4e45-97ed-03caf894bc08-kube-api-access-t7gmq" (OuterVolumeSpecName: "kube-api-access-t7gmq") pod "315eefbb-24fb-4e45-97ed-03caf894bc08" (UID: "315eefbb-24fb-4e45-97ed-03caf894bc08"). InnerVolumeSpecName "kube-api-access-t7gmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.190436 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "mysql-db") pod "315eefbb-24fb-4e45-97ed-03caf894bc08" (UID: "315eefbb-24fb-4e45-97ed-03caf894bc08"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.204691 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "315eefbb-24fb-4e45-97ed-03caf894bc08" (UID: "315eefbb-24fb-4e45-97ed-03caf894bc08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.217651 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "315eefbb-24fb-4e45-97ed-03caf894bc08" (UID: "315eefbb-24fb-4e45-97ed-03caf894bc08"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.279520 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.282546 4978 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-kolla-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.282577 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.282593 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-generated\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.282607 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-config-data-default\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.282619 4978 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/315eefbb-24fb-4e45-97ed-03caf894bc08-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.282632 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7gmq\" (UniqueName: \"kubernetes.io/projected/315eefbb-24fb-4e45-97ed-03caf894bc08-kube-api-access-t7gmq\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.282644 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/315eefbb-24fb-4e45-97ed-03caf894bc08-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.282677 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.297942 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.355800 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.378479 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.378526 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.383832 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43bd8325-9408-4a32-9285-d522d497205e-erlang-cookie-secret\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.383902 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-plugins\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.383942 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-plugins-conf\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.383978 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-erlang-cookie\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.384020 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-server-conf\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.384042 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-confd\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.384069 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.384123 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-tls\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.384144 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.384175 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43bd8325-9408-4a32-9285-d522d497205e-pod-info\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.384210 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnl5k\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-kube-api-access-dnl5k\") pod \"43bd8325-9408-4a32-9285-d522d497205e\" (UID: \"43bd8325-9408-4a32-9285-d522d497205e\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.384513 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.386753 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.387490 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-kube-api-access-dnl5k" (OuterVolumeSpecName: "kube-api-access-dnl5k") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "kube-api-access-dnl5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.390045 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.390183 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.392444 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43bd8325-9408-4a32-9285-d522d497205e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.397572 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.399930 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/43bd8325-9408-4a32-9285-d522d497205e-pod-info" (OuterVolumeSpecName: "pod-info") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.399942 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.424770 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data" (OuterVolumeSpecName: "config-data") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.427749 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.438889 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-server-conf" (OuterVolumeSpecName: "server-conf") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.485561 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-server-conf\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.485633 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.485667 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvzgx\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-kube-api-access-cvzgx\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.485704 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-plugins\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.485735 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/09112202-a3ef-4742-bdbe-b46c1c9a55ba-erlang-cookie-secret\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.485777 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-erlang-cookie\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.485811 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-plugins-conf\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.485889 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.485934 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-tls\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.485991 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/09112202-a3ef-4742-bdbe-b46c1c9a55ba-pod-info\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486043 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-confd\") pod \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\" (UID: \"09112202-a3ef-4742-bdbe-b46c1c9a55ba\") " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486515 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486540 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486565 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486578 4978 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43bd8325-9408-4a32-9285-d522d497205e-pod-info\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486593 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnl5k\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-kube-api-access-dnl5k\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486605 4978 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43bd8325-9408-4a32-9285-d522d497205e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486617 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486648 4978 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486659 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.486670 4978 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43bd8325-9408-4a32-9285-d522d497205e-server-conf\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.487278 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.487738 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.488891 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.489493 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.491273 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09112202-a3ef-4742-bdbe-b46c1c9a55ba-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.491795 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-kube-api-access-cvzgx" (OuterVolumeSpecName: "kube-api-access-cvzgx") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "kube-api-access-cvzgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.491900 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.492386 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/09112202-a3ef-4742-bdbe-b46c1c9a55ba-pod-info" (OuterVolumeSpecName: "pod-info") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.508178 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.521212 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "43bd8325-9408-4a32-9285-d522d497205e" (UID: "43bd8325-9408-4a32-9285-d522d497205e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.523496 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-server-conf" (OuterVolumeSpecName: "server-conf") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.523826 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data" (OuterVolumeSpecName: "config-data") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.572824 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "09112202-a3ef-4742-bdbe-b46c1c9a55ba" (UID: "09112202-a3ef-4742-bdbe-b46c1c9a55ba"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587724 4978 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587758 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587768 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43bd8325-9408-4a32-9285-d522d497205e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587801 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587809 4978 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/09112202-a3ef-4742-bdbe-b46c1c9a55ba-pod-info\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587818 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587827 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587835 4978 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/09112202-a3ef-4742-bdbe-b46c1c9a55ba-server-conf\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587890 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587903 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvzgx\" (UniqueName: \"kubernetes.io/projected/09112202-a3ef-4742-bdbe-b46c1c9a55ba-kube-api-access-cvzgx\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587913 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.587924 4978 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/09112202-a3ef-4742-bdbe-b46c1c9a55ba-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.588296 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/09112202-a3ef-4742-bdbe-b46c1c9a55ba-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.602916 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.689721 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.713146 4978 generic.go:334] "Generic (PLEG): container finished" podID="3ee398cf-5bd3-4905-80f2-b66d67c57f3d" containerID="3b71dd25e21037082b24f81353af12fb7fe11432d9ea70560778f8494ba5ae98" exitCode=0 Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.713203 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7dd4fddc6c-nv49c" event={"ID":"3ee398cf-5bd3-4905-80f2-b66d67c57f3d","Type":"ContainerDied","Data":"3b71dd25e21037082b24f81353af12fb7fe11432d9ea70560778f8494ba5ae98"} Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.715040 4978 generic.go:334] "Generic (PLEG): container finished" podID="09112202-a3ef-4742-bdbe-b46c1c9a55ba" containerID="0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33" exitCode=0 Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.715151 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.715792 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"09112202-a3ef-4742-bdbe-b46c1c9a55ba","Type":"ContainerDied","Data":"0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33"} Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.715840 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"09112202-a3ef-4742-bdbe-b46c1c9a55ba","Type":"ContainerDied","Data":"b5b21db6b8c3246cc8774533366c1a40f3d7d9ddd47ef7ebb01557d48baaa599"} Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.715859 4978 scope.go:117] "RemoveContainer" containerID="0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.733817 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.733986 4978 generic.go:334] "Generic (PLEG): container finished" podID="43bd8325-9408-4a32-9285-d522d497205e" containerID="2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5" exitCode=0 Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.734046 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43bd8325-9408-4a32-9285-d522d497205e","Type":"ContainerDied","Data":"2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5"} Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.734426 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"43bd8325-9408-4a32-9285-d522d497205e","Type":"ContainerDied","Data":"594b702495e86182b648ba4b63933e81dc4baf29486b60b748ddee3ac22df6be"} Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.737901 4978 generic.go:334] "Generic (PLEG): container finished" podID="315eefbb-24fb-4e45-97ed-03caf894bc08" containerID="4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f" exitCode=0 Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.737972 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.737975 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"315eefbb-24fb-4e45-97ed-03caf894bc08","Type":"ContainerDied","Data":"4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f"} Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.738026 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"315eefbb-24fb-4e45-97ed-03caf894bc08","Type":"ContainerDied","Data":"3a840466a68665264ab5537ac4787284f8dc1e57ebaeb06a96ec0e50cca2b4eb"} Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.760987 4978 scope.go:117] "RemoveContainer" containerID="62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.788398 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.796314 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.802815 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.811247 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.818158 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.836520 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.867438 4978 scope.go:117] "RemoveContainer" containerID="0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33" Feb 25 07:10:38 crc kubenswrapper[4978]: E0225 07:10:38.867988 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33\": container with ID starting with 0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33 not found: ID does not exist" containerID="0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.868078 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33"} err="failed to get container status \"0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33\": rpc error: code = NotFound desc = could not find container \"0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33\": container with ID starting with 0b191c99c8191d9633f9bd73dc5c39eec8af94866bbdde3438938142a9570d33 not found: ID does not exist" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.868162 4978 scope.go:117] "RemoveContainer" containerID="62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732" Feb 25 07:10:38 crc kubenswrapper[4978]: E0225 07:10:38.868653 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732\": container with ID starting with 62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732 not found: ID does not exist" containerID="62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.868704 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732"} err="failed to get container status \"62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732\": rpc error: code = NotFound desc = could not find container \"62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732\": container with ID starting with 62e00ede8c608a138b686bf74e9e4bec3351f038d6a2282709538c21d8138732 not found: ID does not exist" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.868729 4978 scope.go:117] "RemoveContainer" containerID="2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.900549 4978 scope.go:117] "RemoveContainer" containerID="765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.934184 4978 scope.go:117] "RemoveContainer" containerID="2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5" Feb 25 07:10:38 crc kubenswrapper[4978]: E0225 07:10:38.947584 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5\": container with ID starting with 2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5 not found: ID does not exist" containerID="2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.947622 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5"} err="failed to get container status \"2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5\": rpc error: code = NotFound desc = could not find container \"2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5\": container with ID starting with 2fcd10e116db1a7e5fe671ab86723c5ef507e24d5512c40af6e8b5acf92c69f5 not found: ID does not exist" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.947647 4978 scope.go:117] "RemoveContainer" containerID="765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8" Feb 25 07:10:38 crc kubenswrapper[4978]: E0225 07:10:38.947981 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8\": container with ID starting with 765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8 not found: ID does not exist" containerID="765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.948023 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8"} err="failed to get container status \"765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8\": rpc error: code = NotFound desc = could not find container \"765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8\": container with ID starting with 765401d35bb18a8334b8cde66edaa1bb6a2bc539d517c8bff938581372b34ec8 not found: ID does not exist" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.948048 4978 scope.go:117] "RemoveContainer" containerID="4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f" Feb 25 07:10:38 crc kubenswrapper[4978]: I0225 07:10:38.995327 4978 scope.go:117] "RemoveContainer" containerID="414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.008920 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.022146 4978 scope.go:117] "RemoveContainer" containerID="4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f" Feb 25 07:10:39 crc kubenswrapper[4978]: E0225 07:10:39.022751 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f\": container with ID starting with 4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f not found: ID does not exist" containerID="4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.022790 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f"} err="failed to get container status \"4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f\": rpc error: code = NotFound desc = could not find container \"4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f\": container with ID starting with 4366b35430d6870aeea91abc4b1e367646859bbde561a3014a25676689caf90f not found: ID does not exist" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.022818 4978 scope.go:117] "RemoveContainer" containerID="414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b" Feb 25 07:10:39 crc kubenswrapper[4978]: E0225 07:10:39.034757 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b\": container with ID starting with 414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b not found: ID does not exist" containerID="414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.034963 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b"} err="failed to get container status \"414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b\": rpc error: code = NotFound desc = could not find container \"414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b\": container with ID starting with 414005aa2e638e1825d4c7258efe5a234564d2840b54fcd1f42169cbe588f96b not found: ID does not exist" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.205136 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-fernet-keys\") pod \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.205307 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-combined-ca-bundle\") pod \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.205393 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-scripts\") pod \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.205435 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzhn8\" (UniqueName: \"kubernetes.io/projected/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-kube-api-access-tzhn8\") pod \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.206126 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-public-tls-certs\") pod \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.206175 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-credential-keys\") pod \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.206247 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-config-data\") pod \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.206303 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-internal-tls-certs\") pod \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\" (UID: \"3ee398cf-5bd3-4905-80f2-b66d67c57f3d\") " Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.209966 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-kube-api-access-tzhn8" (OuterVolumeSpecName: "kube-api-access-tzhn8") pod "3ee398cf-5bd3-4905-80f2-b66d67c57f3d" (UID: "3ee398cf-5bd3-4905-80f2-b66d67c57f3d"). InnerVolumeSpecName "kube-api-access-tzhn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.211598 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-scripts" (OuterVolumeSpecName: "scripts") pod "3ee398cf-5bd3-4905-80f2-b66d67c57f3d" (UID: "3ee398cf-5bd3-4905-80f2-b66d67c57f3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.212464 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3ee398cf-5bd3-4905-80f2-b66d67c57f3d" (UID: "3ee398cf-5bd3-4905-80f2-b66d67c57f3d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.212606 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3ee398cf-5bd3-4905-80f2-b66d67c57f3d" (UID: "3ee398cf-5bd3-4905-80f2-b66d67c57f3d"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.225946 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ee398cf-5bd3-4905-80f2-b66d67c57f3d" (UID: "3ee398cf-5bd3-4905-80f2-b66d67c57f3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.229771 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-config-data" (OuterVolumeSpecName: "config-data") pod "3ee398cf-5bd3-4905-80f2-b66d67c57f3d" (UID: "3ee398cf-5bd3-4905-80f2-b66d67c57f3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.240996 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3ee398cf-5bd3-4905-80f2-b66d67c57f3d" (UID: "3ee398cf-5bd3-4905-80f2-b66d67c57f3d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.280499 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3ee398cf-5bd3-4905-80f2-b66d67c57f3d" (UID: "3ee398cf-5bd3-4905-80f2-b66d67c57f3d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.308612 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.308644 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzhn8\" (UniqueName: \"kubernetes.io/projected/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-kube-api-access-tzhn8\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.308657 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.308667 4978 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.308674 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.308683 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.308691 4978 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.308698 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ee398cf-5bd3-4905-80f2-b66d67c57f3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.357143 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09112202-a3ef-4742-bdbe-b46c1c9a55ba" path="/var/lib/kubelet/pods/09112202-a3ef-4742-bdbe-b46c1c9a55ba/volumes" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.357776 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" path="/var/lib/kubelet/pods/27864863-0a60-43ec-ac3e-f86d9f68c7d2/volumes" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.358740 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" path="/var/lib/kubelet/pods/2e39fd28-0520-44bc-9486-a118cec688c0/volumes" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.359386 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="315eefbb-24fb-4e45-97ed-03caf894bc08" path="/var/lib/kubelet/pods/315eefbb-24fb-4e45-97ed-03caf894bc08/volumes" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.360049 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43bd8325-9408-4a32-9285-d522d497205e" path="/var/lib/kubelet/pods/43bd8325-9408-4a32-9285-d522d497205e/volumes" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.360947 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c" path="/var/lib/kubelet/pods/6ccfe3b6-2fd4-432e-b6d4-71869edbfe4c/volumes" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.361281 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85624b7d-192e-42c3-9c5b-18cfd91d37e9" path="/var/lib/kubelet/pods/85624b7d-192e-42c3-9c5b-18cfd91d37e9/volumes" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.756018 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7dd4fddc6c-nv49c" event={"ID":"3ee398cf-5bd3-4905-80f2-b66d67c57f3d","Type":"ContainerDied","Data":"b6bce4a4fde76c203ae95aa89120619243abd412f96cdabdc6c8cd947deefeee"} Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.756065 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7dd4fddc6c-nv49c" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.756105 4978 scope.go:117] "RemoveContainer" containerID="3b71dd25e21037082b24f81353af12fb7fe11432d9ea70560778f8494ba5ae98" Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.760942 4978 generic.go:334] "Generic (PLEG): container finished" podID="8a6b84e2-353d-47ef-82be-c7e182533a57" containerID="5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc" exitCode=0 Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.761041 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8a6b84e2-353d-47ef-82be-c7e182533a57","Type":"ContainerDied","Data":"5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc"} Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.764814 4978 generic.go:334] "Generic (PLEG): container finished" podID="7456f008-69db-4a20-86c4-2ab52eb93aa7" containerID="b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3" exitCode=0 Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.764882 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7456f008-69db-4a20-86c4-2ab52eb93aa7","Type":"ContainerDied","Data":"b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3"} Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.787095 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7dd4fddc6c-nv49c"] Feb 25 07:10:39 crc kubenswrapper[4978]: I0225 07:10:39.797385 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7dd4fddc6c-nv49c"] Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.229698 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.283598 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.425197 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-config-data\") pod \"8a6b84e2-353d-47ef-82be-c7e182533a57\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.425263 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-combined-ca-bundle\") pod \"7456f008-69db-4a20-86c4-2ab52eb93aa7\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.425461 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-config-data\") pod \"7456f008-69db-4a20-86c4-2ab52eb93aa7\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.425532 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-combined-ca-bundle\") pod \"8a6b84e2-353d-47ef-82be-c7e182533a57\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.425594 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krlk2\" (UniqueName: \"kubernetes.io/projected/8a6b84e2-353d-47ef-82be-c7e182533a57-kube-api-access-krlk2\") pod \"8a6b84e2-353d-47ef-82be-c7e182533a57\" (UID: \"8a6b84e2-353d-47ef-82be-c7e182533a57\") " Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.425657 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvbkl\" (UniqueName: \"kubernetes.io/projected/7456f008-69db-4a20-86c4-2ab52eb93aa7-kube-api-access-lvbkl\") pod \"7456f008-69db-4a20-86c4-2ab52eb93aa7\" (UID: \"7456f008-69db-4a20-86c4-2ab52eb93aa7\") " Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.429214 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7456f008-69db-4a20-86c4-2ab52eb93aa7-kube-api-access-lvbkl" (OuterVolumeSpecName: "kube-api-access-lvbkl") pod "7456f008-69db-4a20-86c4-2ab52eb93aa7" (UID: "7456f008-69db-4a20-86c4-2ab52eb93aa7"). InnerVolumeSpecName "kube-api-access-lvbkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.429888 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6b84e2-353d-47ef-82be-c7e182533a57-kube-api-access-krlk2" (OuterVolumeSpecName: "kube-api-access-krlk2") pod "8a6b84e2-353d-47ef-82be-c7e182533a57" (UID: "8a6b84e2-353d-47ef-82be-c7e182533a57"). InnerVolumeSpecName "kube-api-access-krlk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.456578 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7456f008-69db-4a20-86c4-2ab52eb93aa7" (UID: "7456f008-69db-4a20-86c4-2ab52eb93aa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.458925 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a6b84e2-353d-47ef-82be-c7e182533a57" (UID: "8a6b84e2-353d-47ef-82be-c7e182533a57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.459353 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-config-data" (OuterVolumeSpecName: "config-data") pod "7456f008-69db-4a20-86c4-2ab52eb93aa7" (UID: "7456f008-69db-4a20-86c4-2ab52eb93aa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.463290 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-config-data" (OuterVolumeSpecName: "config-data") pod "8a6b84e2-353d-47ef-82be-c7e182533a57" (UID: "8a6b84e2-353d-47ef-82be-c7e182533a57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.534490 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.534538 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.534554 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krlk2\" (UniqueName: \"kubernetes.io/projected/8a6b84e2-353d-47ef-82be-c7e182533a57-kube-api-access-krlk2\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.534569 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvbkl\" (UniqueName: \"kubernetes.io/projected/7456f008-69db-4a20-86c4-2ab52eb93aa7-kube-api-access-lvbkl\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.534585 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6b84e2-353d-47ef-82be-c7e182533a57-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.534598 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7456f008-69db-4a20-86c4-2ab52eb93aa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.782015 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8a6b84e2-353d-47ef-82be-c7e182533a57","Type":"ContainerDied","Data":"e9b14ecbd2a5d502ac04dc0d81c1735e4284d9d51f459cf8d68919a3d771888c"} Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.782079 4978 scope.go:117] "RemoveContainer" containerID="5c87c2154f393da021e7050b27d484005d3a5b762affcd9fd25022596c6c8abc" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.782029 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.784622 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7456f008-69db-4a20-86c4-2ab52eb93aa7","Type":"ContainerDied","Data":"21b8107a73d909dcb7058b275e04afcd4a8f5bcbbdbed6e5e7901b5a9d1bdfc9"} Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.784685 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.811767 4978 scope.go:117] "RemoveContainer" containerID="b7801435f02b3304c8061f4af59c9df1534730abeda7f0849774d515f964d0a3" Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.823536 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.827862 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.841605 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:10:40 crc kubenswrapper[4978]: I0225 07:10:40.847333 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 07:10:41 crc kubenswrapper[4978]: I0225 07:10:41.346597 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ee398cf-5bd3-4905-80f2-b66d67c57f3d" path="/var/lib/kubelet/pods/3ee398cf-5bd3-4905-80f2-b66d67c57f3d/volumes" Feb 25 07:10:41 crc kubenswrapper[4978]: I0225 07:10:41.347984 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7456f008-69db-4a20-86c4-2ab52eb93aa7" path="/var/lib/kubelet/pods/7456f008-69db-4a20-86c4-2ab52eb93aa7/volumes" Feb 25 07:10:41 crc kubenswrapper[4978]: I0225 07:10:41.349588 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a6b84e2-353d-47ef-82be-c7e182533a57" path="/var/lib/kubelet/pods/8a6b84e2-353d-47ef-82be-c7e182533a57/volumes" Feb 25 07:10:42 crc kubenswrapper[4978]: E0225 07:10:42.958960 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:42 crc kubenswrapper[4978]: E0225 07:10:42.973600 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:42 crc kubenswrapper[4978]: E0225 07:10:42.974019 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:42 crc kubenswrapper[4978]: E0225 07:10:42.974198 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:42 crc kubenswrapper[4978]: E0225 07:10:42.974276 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" Feb 25 07:10:42 crc kubenswrapper[4978]: E0225 07:10:42.984959 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:42 crc kubenswrapper[4978]: E0225 07:10:42.986844 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:42 crc kubenswrapper[4978]: E0225 07:10:42.986934 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovs-vswitchd" Feb 25 07:10:45 crc kubenswrapper[4978]: I0225 07:10:45.267427 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bwjnj" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerName="registry-server" probeResult="failure" output=< Feb 25 07:10:45 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 07:10:45 crc kubenswrapper[4978]: > Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.540810 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.541123 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.687848 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.831923 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-public-tls-certs\") pod \"d1ba0ae6-d125-48d2-a866-615d927dc525\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.832019 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-ovndb-tls-certs\") pod \"d1ba0ae6-d125-48d2-a866-615d927dc525\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.832063 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-combined-ca-bundle\") pod \"d1ba0ae6-d125-48d2-a866-615d927dc525\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.832098 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2fdk\" (UniqueName: \"kubernetes.io/projected/d1ba0ae6-d125-48d2-a866-615d927dc525-kube-api-access-m2fdk\") pod \"d1ba0ae6-d125-48d2-a866-615d927dc525\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.832184 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-internal-tls-certs\") pod \"d1ba0ae6-d125-48d2-a866-615d927dc525\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.832219 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-config\") pod \"d1ba0ae6-d125-48d2-a866-615d927dc525\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.832266 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-httpd-config\") pod \"d1ba0ae6-d125-48d2-a866-615d927dc525\" (UID: \"d1ba0ae6-d125-48d2-a866-615d927dc525\") " Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.840921 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d1ba0ae6-d125-48d2-a866-615d927dc525" (UID: "d1ba0ae6-d125-48d2-a866-615d927dc525"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.841336 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1ba0ae6-d125-48d2-a866-615d927dc525-kube-api-access-m2fdk" (OuterVolumeSpecName: "kube-api-access-m2fdk") pod "d1ba0ae6-d125-48d2-a866-615d927dc525" (UID: "d1ba0ae6-d125-48d2-a866-615d927dc525"). InnerVolumeSpecName "kube-api-access-m2fdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.864183 4978 generic.go:334] "Generic (PLEG): container finished" podID="d1ba0ae6-d125-48d2-a866-615d927dc525" containerID="35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e" exitCode=0 Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.864231 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cdb588b97-7jkd8" event={"ID":"d1ba0ae6-d125-48d2-a866-615d927dc525","Type":"ContainerDied","Data":"35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e"} Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.864260 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cdb588b97-7jkd8" event={"ID":"d1ba0ae6-d125-48d2-a866-615d927dc525","Type":"ContainerDied","Data":"7f005098f780e1f9032ee5a1662dbc43eb64e66c0c2fa9b3c638c5e14258ff69"} Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.864279 4978 scope.go:117] "RemoveContainer" containerID="6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.864458 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cdb588b97-7jkd8" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.878816 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d1ba0ae6-d125-48d2-a866-615d927dc525" (UID: "d1ba0ae6-d125-48d2-a866-615d927dc525"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.885006 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1ba0ae6-d125-48d2-a866-615d927dc525" (UID: "d1ba0ae6-d125-48d2-a866-615d927dc525"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.903349 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d1ba0ae6-d125-48d2-a866-615d927dc525" (UID: "d1ba0ae6-d125-48d2-a866-615d927dc525"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.908743 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-config" (OuterVolumeSpecName: "config") pod "d1ba0ae6-d125-48d2-a866-615d927dc525" (UID: "d1ba0ae6-d125-48d2-a866-615d927dc525"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.911471 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d1ba0ae6-d125-48d2-a866-615d927dc525" (UID: "d1ba0ae6-d125-48d2-a866-615d927dc525"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.926941 4978 scope.go:117] "RemoveContainer" containerID="35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.933683 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.933751 4978 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.933761 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.933771 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2fdk\" (UniqueName: \"kubernetes.io/projected/d1ba0ae6-d125-48d2-a866-615d927dc525-kube-api-access-m2fdk\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.933782 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.933793 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.933813 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d1ba0ae6-d125-48d2-a866-615d927dc525-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.948599 4978 scope.go:117] "RemoveContainer" containerID="6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba" Feb 25 07:10:46 crc kubenswrapper[4978]: E0225 07:10:46.949096 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba\": container with ID starting with 6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba not found: ID does not exist" containerID="6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.949132 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba"} err="failed to get container status \"6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba\": rpc error: code = NotFound desc = could not find container \"6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba\": container with ID starting with 6b0a2f6113dfa8c6e060f2368800158466dd851f68261f919c5b82b02ade85ba not found: ID does not exist" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.949159 4978 scope.go:117] "RemoveContainer" containerID="35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e" Feb 25 07:10:46 crc kubenswrapper[4978]: E0225 07:10:46.949404 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e\": container with ID starting with 35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e not found: ID does not exist" containerID="35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e" Feb 25 07:10:46 crc kubenswrapper[4978]: I0225 07:10:46.949433 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e"} err="failed to get container status \"35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e\": rpc error: code = NotFound desc = could not find container \"35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e\": container with ID starting with 35c0f1039df95c6454afb33a5bdc97bd3903180d47058bcac7f1bb2ac2abbc9e not found: ID does not exist" Feb 25 07:10:47 crc kubenswrapper[4978]: I0225 07:10:47.212950 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5cdb588b97-7jkd8"] Feb 25 07:10:47 crc kubenswrapper[4978]: I0225 07:10:47.223294 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5cdb588b97-7jkd8"] Feb 25 07:10:47 crc kubenswrapper[4978]: I0225 07:10:47.343933 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1ba0ae6-d125-48d2-a866-615d927dc525" path="/var/lib/kubelet/pods/d1ba0ae6-d125-48d2-a866-615d927dc525/volumes" Feb 25 07:10:47 crc kubenswrapper[4978]: E0225 07:10:47.955431 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:47 crc kubenswrapper[4978]: E0225 07:10:47.955876 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:47 crc kubenswrapper[4978]: E0225 07:10:47.956549 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:47 crc kubenswrapper[4978]: E0225 07:10:47.956605 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" Feb 25 07:10:47 crc kubenswrapper[4978]: E0225 07:10:47.956923 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:47 crc kubenswrapper[4978]: E0225 07:10:47.959192 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:47 crc kubenswrapper[4978]: E0225 07:10:47.960766 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:47 crc kubenswrapper[4978]: E0225 07:10:47.960826 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovs-vswitchd" Feb 25 07:10:48 crc kubenswrapper[4978]: I0225 07:10:48.473394 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:48 crc kubenswrapper[4978]: I0225 07:10:48.533799 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwmn"] Feb 25 07:10:48 crc kubenswrapper[4978]: I0225 07:10:48.892160 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nxwmn" podUID="34752acd-4270-408b-bc82-af9fc515da2b" containerName="registry-server" containerID="cri-o://9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f" gracePeriod=2 Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.456432 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.592164 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-utilities\") pod \"34752acd-4270-408b-bc82-af9fc515da2b\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.592494 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj5qw\" (UniqueName: \"kubernetes.io/projected/34752acd-4270-408b-bc82-af9fc515da2b-kube-api-access-sj5qw\") pod \"34752acd-4270-408b-bc82-af9fc515da2b\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.592622 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-catalog-content\") pod \"34752acd-4270-408b-bc82-af9fc515da2b\" (UID: \"34752acd-4270-408b-bc82-af9fc515da2b\") " Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.593050 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-utilities" (OuterVolumeSpecName: "utilities") pod "34752acd-4270-408b-bc82-af9fc515da2b" (UID: "34752acd-4270-408b-bc82-af9fc515da2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.598534 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34752acd-4270-408b-bc82-af9fc515da2b-kube-api-access-sj5qw" (OuterVolumeSpecName: "kube-api-access-sj5qw") pod "34752acd-4270-408b-bc82-af9fc515da2b" (UID: "34752acd-4270-408b-bc82-af9fc515da2b"). InnerVolumeSpecName "kube-api-access-sj5qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.643809 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34752acd-4270-408b-bc82-af9fc515da2b" (UID: "34752acd-4270-408b-bc82-af9fc515da2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.694424 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.694473 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj5qw\" (UniqueName: \"kubernetes.io/projected/34752acd-4270-408b-bc82-af9fc515da2b-kube-api-access-sj5qw\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.694496 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34752acd-4270-408b-bc82-af9fc515da2b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.905550 4978 generic.go:334] "Generic (PLEG): container finished" podID="34752acd-4270-408b-bc82-af9fc515da2b" containerID="9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f" exitCode=0 Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.905609 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwmn" event={"ID":"34752acd-4270-408b-bc82-af9fc515da2b","Type":"ContainerDied","Data":"9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f"} Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.905672 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxwmn" event={"ID":"34752acd-4270-408b-bc82-af9fc515da2b","Type":"ContainerDied","Data":"840671ede6def3b430ea5faed42cf1a9b295e209141efd979450b48941cc4fe0"} Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.905702 4978 scope.go:117] "RemoveContainer" containerID="9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f" Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.905878 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxwmn" Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.946161 4978 scope.go:117] "RemoveContainer" containerID="62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd" Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.960717 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwmn"] Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.966915 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxwmn"] Feb 25 07:10:49 crc kubenswrapper[4978]: I0225 07:10:49.985475 4978 scope.go:117] "RemoveContainer" containerID="0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74" Feb 25 07:10:50 crc kubenswrapper[4978]: I0225 07:10:50.027742 4978 scope.go:117] "RemoveContainer" containerID="9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f" Feb 25 07:10:50 crc kubenswrapper[4978]: E0225 07:10:50.028177 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f\": container with ID starting with 9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f not found: ID does not exist" containerID="9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f" Feb 25 07:10:50 crc kubenswrapper[4978]: I0225 07:10:50.028212 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f"} err="failed to get container status \"9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f\": rpc error: code = NotFound desc = could not find container \"9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f\": container with ID starting with 9889e905b5edfb7164b864921a16d05716b77f5ec7a03ac76a047c2ea156513f not found: ID does not exist" Feb 25 07:10:50 crc kubenswrapper[4978]: I0225 07:10:50.028235 4978 scope.go:117] "RemoveContainer" containerID="62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd" Feb 25 07:10:50 crc kubenswrapper[4978]: E0225 07:10:50.028744 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd\": container with ID starting with 62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd not found: ID does not exist" containerID="62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd" Feb 25 07:10:50 crc kubenswrapper[4978]: I0225 07:10:50.028786 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd"} err="failed to get container status \"62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd\": rpc error: code = NotFound desc = could not find container \"62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd\": container with ID starting with 62b4510a873bc49d757e87f3fb6a9c04200e39351fbd59e6355707fb06ebb0dd not found: ID does not exist" Feb 25 07:10:50 crc kubenswrapper[4978]: I0225 07:10:50.028815 4978 scope.go:117] "RemoveContainer" containerID="0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74" Feb 25 07:10:50 crc kubenswrapper[4978]: E0225 07:10:50.029282 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74\": container with ID starting with 0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74 not found: ID does not exist" containerID="0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74" Feb 25 07:10:50 crc kubenswrapper[4978]: I0225 07:10:50.029408 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74"} err="failed to get container status \"0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74\": rpc error: code = NotFound desc = could not find container \"0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74\": container with ID starting with 0eb40c100fc2df760227f2883c82ba0f7e84cfc3f84a5d79560e41f92609da74 not found: ID does not exist" Feb 25 07:10:51 crc kubenswrapper[4978]: I0225 07:10:51.344485 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34752acd-4270-408b-bc82-af9fc515da2b" path="/var/lib/kubelet/pods/34752acd-4270-408b-bc82-af9fc515da2b/volumes" Feb 25 07:10:52 crc kubenswrapper[4978]: E0225 07:10:52.955923 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:52 crc kubenswrapper[4978]: E0225 07:10:52.957214 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:52 crc kubenswrapper[4978]: E0225 07:10:52.957398 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:52 crc kubenswrapper[4978]: E0225 07:10:52.958065 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:52 crc kubenswrapper[4978]: E0225 07:10:52.958125 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" Feb 25 07:10:52 crc kubenswrapper[4978]: E0225 07:10:52.959278 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:52 crc kubenswrapper[4978]: E0225 07:10:52.960660 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:52 crc kubenswrapper[4978]: E0225 07:10:52.960731 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovs-vswitchd" Feb 25 07:10:54 crc kubenswrapper[4978]: I0225 07:10:54.267785 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:54 crc kubenswrapper[4978]: I0225 07:10:54.347158 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 07:10:54 crc kubenswrapper[4978]: I0225 07:10:54.433110 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bwjnj"] Feb 25 07:10:54 crc kubenswrapper[4978]: I0225 07:10:54.513318 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krctw"] Feb 25 07:10:54 crc kubenswrapper[4978]: I0225 07:10:54.513610 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-krctw" podUID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerName="registry-server" containerID="cri-o://b177e600244116e90d5fab43280a48653d1fb3d11c36074d427767becec43ed2" gracePeriod=2 Feb 25 07:10:54 crc kubenswrapper[4978]: I0225 07:10:54.961281 4978 generic.go:334] "Generic (PLEG): container finished" podID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerID="b177e600244116e90d5fab43280a48653d1fb3d11c36074d427767becec43ed2" exitCode=0 Feb 25 07:10:54 crc kubenswrapper[4978]: I0225 07:10:54.961402 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krctw" event={"ID":"8520c8a9-3af3-465a-8c1d-aa9d10c36c40","Type":"ContainerDied","Data":"b177e600244116e90d5fab43280a48653d1fb3d11c36074d427767becec43ed2"} Feb 25 07:10:54 crc kubenswrapper[4978]: I0225 07:10:54.961624 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krctw" event={"ID":"8520c8a9-3af3-465a-8c1d-aa9d10c36c40","Type":"ContainerDied","Data":"0976048df8fa87410d01567784b369b848bf3e8022d1889e49a008d23f0235ee"} Feb 25 07:10:54 crc kubenswrapper[4978]: I0225 07:10:54.961658 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0976048df8fa87410d01567784b369b848bf3e8022d1889e49a008d23f0235ee" Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.011020 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krctw" Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.097959 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-catalog-content\") pod \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.098092 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg6kp\" (UniqueName: \"kubernetes.io/projected/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-kube-api-access-qg6kp\") pod \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.098153 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-utilities\") pod \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\" (UID: \"8520c8a9-3af3-465a-8c1d-aa9d10c36c40\") " Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.098847 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-utilities" (OuterVolumeSpecName: "utilities") pod "8520c8a9-3af3-465a-8c1d-aa9d10c36c40" (UID: "8520c8a9-3af3-465a-8c1d-aa9d10c36c40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.114438 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-kube-api-access-qg6kp" (OuterVolumeSpecName: "kube-api-access-qg6kp") pod "8520c8a9-3af3-465a-8c1d-aa9d10c36c40" (UID: "8520c8a9-3af3-465a-8c1d-aa9d10c36c40"). InnerVolumeSpecName "kube-api-access-qg6kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.199414 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.199440 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg6kp\" (UniqueName: \"kubernetes.io/projected/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-kube-api-access-qg6kp\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.210118 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8520c8a9-3af3-465a-8c1d-aa9d10c36c40" (UID: "8520c8a9-3af3-465a-8c1d-aa9d10c36c40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.301171 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8520c8a9-3af3-465a-8c1d-aa9d10c36c40-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.969388 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krctw" Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.988746 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krctw"] Feb 25 07:10:55 crc kubenswrapper[4978]: I0225 07:10:55.993659 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-krctw"] Feb 25 07:10:57 crc kubenswrapper[4978]: I0225 07:10:57.341196 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" path="/var/lib/kubelet/pods/8520c8a9-3af3-465a-8c1d-aa9d10c36c40/volumes" Feb 25 07:10:57 crc kubenswrapper[4978]: E0225 07:10:57.956032 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:57 crc kubenswrapper[4978]: E0225 07:10:57.956751 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:57 crc kubenswrapper[4978]: E0225 07:10:57.957164 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Feb 25 07:10:57 crc kubenswrapper[4978]: E0225 07:10:57.957205 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" Feb 25 07:10:57 crc kubenswrapper[4978]: E0225 07:10:57.958019 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:57 crc kubenswrapper[4978]: E0225 07:10:57.960238 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:57 crc kubenswrapper[4978]: E0225 07:10:57.962147 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Feb 25 07:10:57 crc kubenswrapper[4978]: E0225 07:10:57.962227 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-njndh" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovs-vswitchd" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.051204 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerID="ce64ee6e29ee6ea1ff2d376f27c1117c98cf4060372883033763a60e602660f9" exitCode=137 Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.051453 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-njndh_01eac805-ca08-46de-b79b-0ca5645f6955/ovs-vswitchd/0.log" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.051475 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"ce64ee6e29ee6ea1ff2d376f27c1117c98cf4060372883033763a60e602660f9"} Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.054646 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.061942 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-njndh_01eac805-ca08-46de-b79b-0ca5645f6955/ovs-vswitchd/0.log" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.076874 4978 generic.go:334] "Generic (PLEG): container finished" podID="01eac805-ca08-46de-b79b-0ca5645f6955" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" exitCode=137 Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.076926 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-njndh" event={"ID":"01eac805-ca08-46de-b79b-0ca5645f6955","Type":"ContainerDied","Data":"02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b"} Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.076960 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-njndh" event={"ID":"01eac805-ca08-46de-b79b-0ca5645f6955","Type":"ContainerDied","Data":"3e44376df0425e7f67c053a698f5548c2e2d3c3e261fe12a08955000080a6fbc"} Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.077006 4978 scope.go:117] "RemoveContainer" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.114691 4978 scope.go:117] "RemoveContainer" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.140813 4978 scope.go:117] "RemoveContainer" containerID="6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.185061 4978 scope.go:117] "RemoveContainer" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" Feb 25 07:11:01 crc kubenswrapper[4978]: E0225 07:11:01.187316 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b\": container with ID starting with 02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b not found: ID does not exist" containerID="02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.187384 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b"} err="failed to get container status \"02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b\": rpc error: code = NotFound desc = could not find container \"02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b\": container with ID starting with 02fc1ca696b84b4550ec67e0148b91e74cbf8b19c3e649b9bc74b604750f4c4b not found: ID does not exist" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.187446 4978 scope.go:117] "RemoveContainer" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.187731 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01eac805-ca08-46de-b79b-0ca5645f6955-scripts\") pod \"01eac805-ca08-46de-b79b-0ca5645f6955\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.187886 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-etc-ovs\") pod \"01eac805-ca08-46de-b79b-0ca5645f6955\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188077 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd7s4\" (UniqueName: \"kubernetes.io/projected/01eac805-ca08-46de-b79b-0ca5645f6955-kube-api-access-zd7s4\") pod \"01eac805-ca08-46de-b79b-0ca5645f6955\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188148 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-log\") pod \"01eac805-ca08-46de-b79b-0ca5645f6955\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " Feb 25 07:11:01 crc kubenswrapper[4978]: E0225 07:11:01.188195 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99\": container with ID starting with f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 not found: ID does not exist" containerID="f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188237 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-run\") pod \"01eac805-ca08-46de-b79b-0ca5645f6955\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188269 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-lib\") pod \"01eac805-ca08-46de-b79b-0ca5645f6955\" (UID: \"01eac805-ca08-46de-b79b-0ca5645f6955\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188324 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-log" (OuterVolumeSpecName: "var-log") pod "01eac805-ca08-46de-b79b-0ca5645f6955" (UID: "01eac805-ca08-46de-b79b-0ca5645f6955"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188415 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "01eac805-ca08-46de-b79b-0ca5645f6955" (UID: "01eac805-ca08-46de-b79b-0ca5645f6955"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188442 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01eac805-ca08-46de-b79b-0ca5645f6955-scripts" (OuterVolumeSpecName: "scripts") pod "01eac805-ca08-46de-b79b-0ca5645f6955" (UID: "01eac805-ca08-46de-b79b-0ca5645f6955"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188490 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-run" (OuterVolumeSpecName: "var-run") pod "01eac805-ca08-46de-b79b-0ca5645f6955" (UID: "01eac805-ca08-46de-b79b-0ca5645f6955"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188512 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-lib" (OuterVolumeSpecName: "var-lib") pod "01eac805-ca08-46de-b79b-0ca5645f6955" (UID: "01eac805-ca08-46de-b79b-0ca5645f6955"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188254 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99"} err="failed to get container status \"f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99\": rpc error: code = NotFound desc = could not find container \"f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99\": container with ID starting with f4a3a9050a7413c3e606b7ce731db7144879436b31bafdc280852731cec9dd99 not found: ID does not exist" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188601 4978 scope.go:117] "RemoveContainer" containerID="6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8" Feb 25 07:11:01 crc kubenswrapper[4978]: E0225 07:11:01.188954 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8\": container with ID starting with 6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8 not found: ID does not exist" containerID="6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.188995 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8"} err="failed to get container status \"6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8\": rpc error: code = NotFound desc = could not find container \"6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8\": container with ID starting with 6756999c249561cdb0fab1708b99125bb68934ea6ee17be6eeea63aa0034e3e8 not found: ID does not exist" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.189777 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01eac805-ca08-46de-b79b-0ca5645f6955-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.189806 4978 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-etc-ovs\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.189844 4978 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-log\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.189855 4978 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-run\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.189866 4978 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/01eac805-ca08-46de-b79b-0ca5645f6955-var-lib\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.197774 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01eac805-ca08-46de-b79b-0ca5645f6955-kube-api-access-zd7s4" (OuterVolumeSpecName: "kube-api-access-zd7s4") pod "01eac805-ca08-46de-b79b-0ca5645f6955" (UID: "01eac805-ca08-46de-b79b-0ca5645f6955"). InnerVolumeSpecName "kube-api-access-zd7s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.293284 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd7s4\" (UniqueName: \"kubernetes.io/projected/01eac805-ca08-46de-b79b-0ca5645f6955-kube-api-access-zd7s4\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.373966 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.495947 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.496027 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwbxf\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-kube-api-access-lwbxf\") pod \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.496060 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-combined-ca-bundle\") pod \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.496092 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift\") pod \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.496550 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-lock\") pod \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.496608 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-cache\") pod \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\" (UID: \"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887\") " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.496947 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-lock" (OuterVolumeSpecName: "lock") pod "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.497274 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-cache" (OuterVolumeSpecName: "cache") pod "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.498140 4978 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-lock\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.498180 4978 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-cache\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.499949 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.500271 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-kube-api-access-lwbxf" (OuterVolumeSpecName: "kube-api-access-lwbxf") pod "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887"). InnerVolumeSpecName "kube-api-access-lwbxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.500481 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "swift") pod "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.599983 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.600216 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwbxf\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-kube-api-access-lwbxf\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.600227 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.613023 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.702677 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.833573 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" (UID: "9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:11:01 crc kubenswrapper[4978]: I0225 07:11:01.905706 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.101040 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887","Type":"ContainerDied","Data":"d1f7e3a38ee0eaf8e5860da6b520860688f76a4907f43b92be68142e113845f5"} Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.101197 4978 scope.go:117] "RemoveContainer" containerID="ce64ee6e29ee6ea1ff2d376f27c1117c98cf4060372883033763a60e602660f9" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.101606 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.120688 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-njndh" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.160789 4978 scope.go:117] "RemoveContainer" containerID="d908490f57d2f2a43ad75a1a8576681b6638c3b6cc2217e47bfbe190b3f31f3b" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.163252 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-njndh"] Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.168733 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-njndh"] Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.189468 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.193380 4978 scope.go:117] "RemoveContainer" containerID="4dfed42feb8ac7a42412c46dcae347795ab03fb0af477b3ffbb5372f2f1e92c3" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.199163 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.217126 4978 scope.go:117] "RemoveContainer" containerID="c5c66777c2f36490e61ea071a8cde68af4d8aa4113f9db174c104aae89ea3baf" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.240418 4978 scope.go:117] "RemoveContainer" containerID="1ec7901b94aa973e76af694ac0b6f07b5c9788edfe7b888e4ae6f0ec6806b20b" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.273938 4978 scope.go:117] "RemoveContainer" containerID="6fefdf96f9ea31d2462085c35bca5e00e4f192d43b6b9dbe2f57601d1afde09c" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.318680 4978 scope.go:117] "RemoveContainer" containerID="cfdc985f7b5bb332d961a90b34a12545a7880322a9b3eaa992362b444ccba073" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.344051 4978 scope.go:117] "RemoveContainer" containerID="f892ef9c825b59066a0c023cd3b507cce63da08b00077af4f81ced0830c89f1c" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.379830 4978 scope.go:117] "RemoveContainer" containerID="c247f762435a5bf5a26de2224e27277501e932169aad0cc00bd5b31c4c134647" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.398357 4978 scope.go:117] "RemoveContainer" containerID="36f71fa276b25e88398462e3bbc6d6b87e2bee8c2aade1e2f5f0ffdc61440473" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.420282 4978 scope.go:117] "RemoveContainer" containerID="2c1a63624536475bb718a04a658404433f192ece6f6682b1e5ca98142ac9eb79" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.439743 4978 scope.go:117] "RemoveContainer" containerID="c196bca824e71c59dc1d9939514fc9f5d009c35f1cb8b40a2d2b99b8d722f60b" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.457425 4978 scope.go:117] "RemoveContainer" containerID="25a3891e332bc2d905bcc5238aea824519156ba640e57709ba7a03f1986ab36b" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.481812 4978 scope.go:117] "RemoveContainer" containerID="bb9960184024cbe585b528d58a3dd1a8b9fdeda2c6566bd7d76615473e70311e" Feb 25 07:11:02 crc kubenswrapper[4978]: I0225 07:11:02.508335 4978 scope.go:117] "RemoveContainer" containerID="fe36f33cfa027ca4f84b465a94aa8c1bbd4436f1982627dab087ee9bdbe66214" Feb 25 07:11:03 crc kubenswrapper[4978]: I0225 07:11:03.344213 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" path="/var/lib/kubelet/pods/01eac805-ca08-46de-b79b-0ca5645f6955/volumes" Feb 25 07:11:03 crc kubenswrapper[4978]: I0225 07:11:03.346013 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" path="/var/lib/kubelet/pods/9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887/volumes" Feb 25 07:11:16 crc kubenswrapper[4978]: I0225 07:11:16.540977 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:11:16 crc kubenswrapper[4978]: I0225 07:11:16.541742 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:11:16 crc kubenswrapper[4978]: I0225 07:11:16.541810 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:11:16 crc kubenswrapper[4978]: I0225 07:11:16.542750 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:11:16 crc kubenswrapper[4978]: I0225 07:11:16.542835 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" gracePeriod=600 Feb 25 07:11:16 crc kubenswrapper[4978]: E0225 07:11:16.676306 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:11:17 crc kubenswrapper[4978]: I0225 07:11:17.291727 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" exitCode=0 Feb 25 07:11:17 crc kubenswrapper[4978]: I0225 07:11:17.291803 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5"} Feb 25 07:11:17 crc kubenswrapper[4978]: I0225 07:11:17.292138 4978 scope.go:117] "RemoveContainer" containerID="b1d31d1a63c47b381a67200dff4a026477c582f264c877adf6984b366b252ba2" Feb 25 07:11:17 crc kubenswrapper[4978]: I0225 07:11:17.294113 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:11:17 crc kubenswrapper[4978]: E0225 07:11:17.294604 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:11:17 crc kubenswrapper[4978]: I0225 07:11:17.744275 4978 scope.go:117] "RemoveContainer" containerID="0f5bccda1e84ea1f67c9b0c2e1a920ebe0c9151d2e7c2daf427dacad83d535c2" Feb 25 07:11:17 crc kubenswrapper[4978]: I0225 07:11:17.779701 4978 scope.go:117] "RemoveContainer" containerID="de7ecaac1189f75bd90f33d54a8d89e205e6520a488f4046c8e09e86bbcd6e57" Feb 25 07:11:17 crc kubenswrapper[4978]: I0225 07:11:17.819865 4978 scope.go:117] "RemoveContainer" containerID="b177e600244116e90d5fab43280a48653d1fb3d11c36074d427767becec43ed2" Feb 25 07:11:30 crc kubenswrapper[4978]: I0225 07:11:30.328073 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:11:30 crc kubenswrapper[4978]: E0225 07:11:30.329078 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.638620 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hb7nx"] Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639439 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639460 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-server" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639470 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovs-vswitchd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639476 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovs-vswitchd" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639484 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315eefbb-24fb-4e45-97ed-03caf894bc08" containerName="mysql-bootstrap" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639490 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="315eefbb-24fb-4e45-97ed-03caf894bc08" containerName="mysql-bootstrap" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639500 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" containerName="glance-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639506 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" containerName="glance-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639519 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee398cf-5bd3-4905-80f2-b66d67c57f3d" containerName="keystone-api" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639525 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee398cf-5bd3-4905-80f2-b66d67c57f3d" containerName="keystone-api" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639534 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerName="barbican-api-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639541 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerName="barbican-api-log" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639551 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="swift-recon-cron" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639556 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="swift-recon-cron" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639565 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09112202-a3ef-4742-bdbe-b46c1c9a55ba" containerName="rabbitmq" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639571 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="09112202-a3ef-4742-bdbe-b46c1c9a55ba" containerName="rabbitmq" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639578 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-expirer" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639584 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-expirer" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639593 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34752acd-4270-408b-bc82-af9fc515da2b" containerName="extract-utilities" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639599 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="34752acd-4270-408b-bc82-af9fc515da2b" containerName="extract-utilities" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639609 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-replicator" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639615 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-replicator" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639623 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639629 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639639 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43bd8325-9408-4a32-9285-d522d497205e" containerName="rabbitmq" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639645 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="43bd8325-9408-4a32-9285-d522d497205e" containerName="rabbitmq" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639653 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34752acd-4270-408b-bc82-af9fc515da2b" containerName="registry-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639659 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="34752acd-4270-408b-bc82-af9fc515da2b" containerName="registry-server" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639669 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerName="barbican-api" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639675 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerName="barbican-api" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639681 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85624b7d-192e-42c3-9c5b-18cfd91d37e9" containerName="memcached" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639687 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="85624b7d-192e-42c3-9c5b-18cfd91d37e9" containerName="memcached" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639695 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c62987-be2b-4deb-a88c-107db5d00415" containerName="nova-cell0-conductor-conductor" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639702 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c62987-be2b-4deb-a88c-107db5d00415" containerName="nova-cell0-conductor-conductor" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639709 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="ceilometer-central-agent" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639716 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="ceilometer-central-agent" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639726 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-replicator" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639732 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-replicator" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639742 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerName="extract-utilities" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639747 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerName="extract-utilities" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639755 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerName="ovn-northd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639761 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerName="ovn-northd" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639771 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a16d36-c629-483e-9b15-fbd7b8913883" containerName="placement-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639777 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a16d36-c629-483e-9b15-fbd7b8913883" containerName="placement-log" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639786 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7456f008-69db-4a20-86c4-2ab52eb93aa7" containerName="nova-scheduler-scheduler" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639792 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7456f008-69db-4a20-86c4-2ab52eb93aa7" containerName="nova-scheduler-scheduler" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639798 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" containerName="glance-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639804 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" containerName="glance-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639810 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" containerName="glance-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639818 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" containerName="glance-log" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639824 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639830 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-server" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639838 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" containerName="glance-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639843 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" containerName="glance-log" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639851 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="315eefbb-24fb-4e45-97ed-03caf894bc08" containerName="galera" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639856 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="315eefbb-24fb-4e45-97ed-03caf894bc08" containerName="galera" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639863 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="proxy-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639868 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="proxy-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639876 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server-init" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639881 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server-init" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639890 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639896 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-log" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639903 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6a16d36-c629-483e-9b15-fbd7b8913883" containerName="placement-api" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639908 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6a16d36-c629-483e-9b15-fbd7b8913883" containerName="placement-api" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639915 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43bd8325-9408-4a32-9285-d522d497205e" containerName="setup-container" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639921 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="43bd8325-9408-4a32-9285-d522d497205e" containerName="setup-container" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639927 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="ceilometer-notification-agent" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639934 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="ceilometer-notification-agent" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639943 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ba0ae6-d125-48d2-a866-615d927dc525" containerName="neutron-api" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639948 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ba0ae6-d125-48d2-a866-615d927dc525" containerName="neutron-api" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639954 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ba0ae6-d125-48d2-a866-615d927dc525" containerName="neutron-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639960 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ba0ae6-d125-48d2-a866-615d927dc525" containerName="neutron-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639971 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09112202-a3ef-4742-bdbe-b46c1c9a55ba" containerName="setup-container" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639976 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="09112202-a3ef-4742-bdbe-b46c1c9a55ba" containerName="setup-container" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.639985 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3be5007a-36e1-4b55-a8d8-5fd560efb8fe" containerName="kube-state-metrics" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.639991 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3be5007a-36e1-4b55-a8d8-5fd560efb8fe" containerName="kube-state-metrics" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640001 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-reaper" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640006 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-reaper" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640015 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-updater" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640021 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-updater" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640028 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerName="registry-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640033 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerName="registry-server" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640044 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-updater" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640049 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-updater" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640055 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-metadata" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640060 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-metadata" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640071 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-auditor" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640076 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-auditor" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640083 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" containerName="nova-api-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640089 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" containerName="nova-api-log" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640097 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640102 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-server" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640111 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="sg-core" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640117 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="sg-core" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640122 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34752acd-4270-408b-bc82-af9fc515da2b" containerName="extract-content" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640127 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="34752acd-4270-408b-bc82-af9fc515da2b" containerName="extract-content" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640136 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerName="openstack-network-exporter" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640142 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerName="openstack-network-exporter" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640148 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-auditor" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640154 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-auditor" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640160 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" containerName="nova-api-api" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640166 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" containerName="nova-api-api" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640174 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="rsync" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640180 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="rsync" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640187 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6b84e2-353d-47ef-82be-c7e182533a57" containerName="nova-cell1-conductor-conductor" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640192 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6b84e2-353d-47ef-82be-c7e182533a57" containerName="nova-cell1-conductor-conductor" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640201 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerName="extract-content" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640206 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerName="extract-content" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640213 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-auditor" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640219 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-auditor" Feb 25 07:11:39 crc kubenswrapper[4978]: E0225 07:11:39.640228 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-replicator" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640233 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-replicator" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640347 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-updater" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640362 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640419 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" containerName="nova-api-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640429 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6a16d36-c629-483e-9b15-fbd7b8913883" containerName="placement-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640440 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="rsync" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640448 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ba0ae6-d125-48d2-a866-615d927dc525" containerName="neutron-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640458 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-auditor" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640468 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovs-vswitchd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640475 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7456f008-69db-4a20-86c4-2ab52eb93aa7" containerName="nova-scheduler-scheduler" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640483 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-expirer" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640493 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcd31a1f-b5cf-4fb2-abf8-b2b8b78903b8" containerName="nova-metadata-metadata" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640501 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ee398cf-5bd3-4905-80f2-b66d67c57f3d" containerName="keystone-api" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640512 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="ceilometer-notification-agent" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640520 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="43bd8325-9408-4a32-9285-d522d497205e" containerName="rabbitmq" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640527 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-auditor" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640538 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-updater" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640547 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8520c8a9-3af3-465a-8c1d-aa9d10c36c40" containerName="registry-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640558 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6a16d36-c629-483e-9b15-fbd7b8913883" containerName="placement-api" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640564 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerName="barbican-api-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640571 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="ceilometer-central-agent" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640577 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640611 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="85624b7d-192e-42c3-9c5b-18cfd91d37e9" containerName="memcached" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640622 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="swift-recon-cron" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640631 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="01eac805-ca08-46de-b79b-0ca5645f6955" containerName="ovsdb-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640639 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="proxy-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640648 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" containerName="glance-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640657 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e39fd28-0520-44bc-9486-a118cec688c0" containerName="nova-api-api" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640668 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="50c62987-be2b-4deb-a88c-107db5d00415" containerName="nova-cell0-conductor-conductor" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640677 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-reaper" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640686 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="object-replicator" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640692 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="315eefbb-24fb-4e45-97ed-03caf894bc08" containerName="galera" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640699 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640706 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="09112202-a3ef-4742-bdbe-b46c1c9a55ba" containerName="rabbitmq" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640713 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ba0ae6-d125-48d2-a866-615d927dc525" containerName="neutron-api" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640720 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="34752acd-4270-408b-bc82-af9fc515da2b" containerName="registry-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640727 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" containerName="glance-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640734 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2631ab-861d-40eb-9b52-b4fd269eb411" containerName="barbican-api" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640743 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerName="openstack-network-exporter" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640752 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d28d9f8-de4f-4223-bfa6-fdf398f133bb" containerName="glance-log" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640759 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-auditor" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640767 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-replicator" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640773 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="58c891d9-fbca-406c-bd1a-ecb13f9d7444" containerName="glance-httpd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640781 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3be5007a-36e1-4b55-a8d8-5fd560efb8fe" containerName="kube-state-metrics" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640787 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="27864863-0a60-43ec-ac3e-f86d9f68c7d2" containerName="ovn-northd" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640794 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="account-replicator" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640801 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ddb624b-1c38-44b8-8f02-bd177937ef4d" containerName="sg-core" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640808 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6b84e2-353d-47ef-82be-c7e182533a57" containerName="nova-cell1-conductor-conductor" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.640816 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a45fe8b-6fb1-4511-80dc-5b1e1ecdd887" containerName="container-server" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.641880 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.651252 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hb7nx"] Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.801158 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-catalog-content\") pod \"community-operators-hb7nx\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.801263 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-utilities\") pod \"community-operators-hb7nx\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.801392 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wpkb\" (UniqueName: \"kubernetes.io/projected/51dc9706-6870-41fb-973b-e59108559437-kube-api-access-5wpkb\") pod \"community-operators-hb7nx\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.902218 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-catalog-content\") pod \"community-operators-hb7nx\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.902288 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-utilities\") pod \"community-operators-hb7nx\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.902342 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wpkb\" (UniqueName: \"kubernetes.io/projected/51dc9706-6870-41fb-973b-e59108559437-kube-api-access-5wpkb\") pod \"community-operators-hb7nx\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.903228 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-catalog-content\") pod \"community-operators-hb7nx\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.903524 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-utilities\") pod \"community-operators-hb7nx\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.928146 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wpkb\" (UniqueName: \"kubernetes.io/projected/51dc9706-6870-41fb-973b-e59108559437-kube-api-access-5wpkb\") pod \"community-operators-hb7nx\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:39 crc kubenswrapper[4978]: I0225 07:11:39.972063 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:40 crc kubenswrapper[4978]: I0225 07:11:40.489903 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hb7nx"] Feb 25 07:11:40 crc kubenswrapper[4978]: I0225 07:11:40.533415 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb7nx" event={"ID":"51dc9706-6870-41fb-973b-e59108559437","Type":"ContainerStarted","Data":"ae36786ea1ba136e5c096636d163a874e518c44baf13430d4d9b8eac4ca944ee"} Feb 25 07:11:41 crc kubenswrapper[4978]: I0225 07:11:41.543455 4978 generic.go:334] "Generic (PLEG): container finished" podID="51dc9706-6870-41fb-973b-e59108559437" containerID="9fae477b6572ba1d21033dc861962d4c682f77d7ffd70b7de1f3bc573df560a8" exitCode=0 Feb 25 07:11:41 crc kubenswrapper[4978]: I0225 07:11:41.543549 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb7nx" event={"ID":"51dc9706-6870-41fb-973b-e59108559437","Type":"ContainerDied","Data":"9fae477b6572ba1d21033dc861962d4c682f77d7ffd70b7de1f3bc573df560a8"} Feb 25 07:11:42 crc kubenswrapper[4978]: I0225 07:11:42.557523 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb7nx" event={"ID":"51dc9706-6870-41fb-973b-e59108559437","Type":"ContainerStarted","Data":"443289f7d81ce2303786095bbcd980c547065a02cf37cfafcf7fc57dabb2fc54"} Feb 25 07:11:43 crc kubenswrapper[4978]: I0225 07:11:43.570558 4978 generic.go:334] "Generic (PLEG): container finished" podID="51dc9706-6870-41fb-973b-e59108559437" containerID="443289f7d81ce2303786095bbcd980c547065a02cf37cfafcf7fc57dabb2fc54" exitCode=0 Feb 25 07:11:43 crc kubenswrapper[4978]: I0225 07:11:43.570631 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb7nx" event={"ID":"51dc9706-6870-41fb-973b-e59108559437","Type":"ContainerDied","Data":"443289f7d81ce2303786095bbcd980c547065a02cf37cfafcf7fc57dabb2fc54"} Feb 25 07:11:44 crc kubenswrapper[4978]: I0225 07:11:44.328423 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:11:44 crc kubenswrapper[4978]: E0225 07:11:44.329311 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:11:44 crc kubenswrapper[4978]: I0225 07:11:44.582259 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb7nx" event={"ID":"51dc9706-6870-41fb-973b-e59108559437","Type":"ContainerStarted","Data":"cdf774c0187d2f474f7178bb1e55466ca1ff546c3d6382e37a85ae569ebe6d98"} Feb 25 07:11:44 crc kubenswrapper[4978]: I0225 07:11:44.614281 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hb7nx" podStartSLOduration=3.211539189 podStartE2EDuration="5.614249996s" podCreationTimestamp="2026-02-25 07:11:39 +0000 UTC" firstStartedPulling="2026-02-25 07:11:41.546924252 +0000 UTC m=+1594.986180721" lastFinishedPulling="2026-02-25 07:11:43.949635039 +0000 UTC m=+1597.388891528" observedRunningTime="2026-02-25 07:11:44.609140318 +0000 UTC m=+1598.048396817" watchObservedRunningTime="2026-02-25 07:11:44.614249996 +0000 UTC m=+1598.053506505" Feb 25 07:11:49 crc kubenswrapper[4978]: I0225 07:11:49.972913 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:49 crc kubenswrapper[4978]: I0225 07:11:49.973448 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:50 crc kubenswrapper[4978]: I0225 07:11:50.070921 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:50 crc kubenswrapper[4978]: I0225 07:11:50.712475 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:50 crc kubenswrapper[4978]: I0225 07:11:50.786088 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hb7nx"] Feb 25 07:11:52 crc kubenswrapper[4978]: I0225 07:11:52.656752 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hb7nx" podUID="51dc9706-6870-41fb-973b-e59108559437" containerName="registry-server" containerID="cri-o://cdf774c0187d2f474f7178bb1e55466ca1ff546c3d6382e37a85ae569ebe6d98" gracePeriod=2 Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.674938 4978 generic.go:334] "Generic (PLEG): container finished" podID="51dc9706-6870-41fb-973b-e59108559437" containerID="cdf774c0187d2f474f7178bb1e55466ca1ff546c3d6382e37a85ae569ebe6d98" exitCode=0 Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.675333 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb7nx" event={"ID":"51dc9706-6870-41fb-973b-e59108559437","Type":"ContainerDied","Data":"cdf774c0187d2f474f7178bb1e55466ca1ff546c3d6382e37a85ae569ebe6d98"} Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.675412 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hb7nx" event={"ID":"51dc9706-6870-41fb-973b-e59108559437","Type":"ContainerDied","Data":"ae36786ea1ba136e5c096636d163a874e518c44baf13430d4d9b8eac4ca944ee"} Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.675441 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae36786ea1ba136e5c096636d163a874e518c44baf13430d4d9b8eac4ca944ee" Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.676063 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.725600 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-utilities\") pod \"51dc9706-6870-41fb-973b-e59108559437\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.725674 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wpkb\" (UniqueName: \"kubernetes.io/projected/51dc9706-6870-41fb-973b-e59108559437-kube-api-access-5wpkb\") pod \"51dc9706-6870-41fb-973b-e59108559437\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.725716 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-catalog-content\") pod \"51dc9706-6870-41fb-973b-e59108559437\" (UID: \"51dc9706-6870-41fb-973b-e59108559437\") " Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.726739 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-utilities" (OuterVolumeSpecName: "utilities") pod "51dc9706-6870-41fb-973b-e59108559437" (UID: "51dc9706-6870-41fb-973b-e59108559437"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.734914 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51dc9706-6870-41fb-973b-e59108559437-kube-api-access-5wpkb" (OuterVolumeSpecName: "kube-api-access-5wpkb") pod "51dc9706-6870-41fb-973b-e59108559437" (UID: "51dc9706-6870-41fb-973b-e59108559437"). InnerVolumeSpecName "kube-api-access-5wpkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.827792 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:53 crc kubenswrapper[4978]: I0225 07:11:53.827848 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wpkb\" (UniqueName: \"kubernetes.io/projected/51dc9706-6870-41fb-973b-e59108559437-kube-api-access-5wpkb\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:54 crc kubenswrapper[4978]: I0225 07:11:54.047189 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51dc9706-6870-41fb-973b-e59108559437" (UID: "51dc9706-6870-41fb-973b-e59108559437"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:11:54 crc kubenswrapper[4978]: I0225 07:11:54.133987 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51dc9706-6870-41fb-973b-e59108559437-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:11:54 crc kubenswrapper[4978]: I0225 07:11:54.692815 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hb7nx" Feb 25 07:11:54 crc kubenswrapper[4978]: I0225 07:11:54.751988 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hb7nx"] Feb 25 07:11:54 crc kubenswrapper[4978]: I0225 07:11:54.756592 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hb7nx"] Feb 25 07:11:55 crc kubenswrapper[4978]: I0225 07:11:55.343546 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51dc9706-6870-41fb-973b-e59108559437" path="/var/lib/kubelet/pods/51dc9706-6870-41fb-973b-e59108559437/volumes" Feb 25 07:11:59 crc kubenswrapper[4978]: I0225 07:11:59.328028 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:11:59 crc kubenswrapper[4978]: E0225 07:11:59.328790 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.168739 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533392-5gp6f"] Feb 25 07:12:00 crc kubenswrapper[4978]: E0225 07:12:00.169617 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51dc9706-6870-41fb-973b-e59108559437" containerName="extract-utilities" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.169647 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="51dc9706-6870-41fb-973b-e59108559437" containerName="extract-utilities" Feb 25 07:12:00 crc kubenswrapper[4978]: E0225 07:12:00.169665 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51dc9706-6870-41fb-973b-e59108559437" containerName="registry-server" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.169678 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="51dc9706-6870-41fb-973b-e59108559437" containerName="registry-server" Feb 25 07:12:00 crc kubenswrapper[4978]: E0225 07:12:00.169698 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51dc9706-6870-41fb-973b-e59108559437" containerName="extract-content" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.169713 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="51dc9706-6870-41fb-973b-e59108559437" containerName="extract-content" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.170005 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="51dc9706-6870-41fb-973b-e59108559437" containerName="registry-server" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.170777 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533392-5gp6f" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.175122 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.176442 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.176506 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.182067 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533392-5gp6f"] Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.250010 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58jsg\" (UniqueName: \"kubernetes.io/projected/0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f-kube-api-access-58jsg\") pod \"auto-csr-approver-29533392-5gp6f\" (UID: \"0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f\") " pod="openshift-infra/auto-csr-approver-29533392-5gp6f" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.351313 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58jsg\" (UniqueName: \"kubernetes.io/projected/0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f-kube-api-access-58jsg\") pod \"auto-csr-approver-29533392-5gp6f\" (UID: \"0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f\") " pod="openshift-infra/auto-csr-approver-29533392-5gp6f" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.385806 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58jsg\" (UniqueName: \"kubernetes.io/projected/0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f-kube-api-access-58jsg\") pod \"auto-csr-approver-29533392-5gp6f\" (UID: \"0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f\") " pod="openshift-infra/auto-csr-approver-29533392-5gp6f" Feb 25 07:12:00 crc kubenswrapper[4978]: I0225 07:12:00.500754 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533392-5gp6f" Feb 25 07:12:01 crc kubenswrapper[4978]: I0225 07:12:01.079731 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533392-5gp6f"] Feb 25 07:12:01 crc kubenswrapper[4978]: W0225 07:12:01.083177 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0de7f76a_e2aa_4040_ae0b_5f65c8e6d40f.slice/crio-37eab5fc027645ae49839b55e778c2c51ed390447de3370310386824ac620424 WatchSource:0}: Error finding container 37eab5fc027645ae49839b55e778c2c51ed390447de3370310386824ac620424: Status 404 returned error can't find the container with id 37eab5fc027645ae49839b55e778c2c51ed390447de3370310386824ac620424 Feb 25 07:12:01 crc kubenswrapper[4978]: I0225 07:12:01.761986 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533392-5gp6f" event={"ID":"0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f","Type":"ContainerStarted","Data":"37eab5fc027645ae49839b55e778c2c51ed390447de3370310386824ac620424"} Feb 25 07:12:02 crc kubenswrapper[4978]: I0225 07:12:02.775027 4978 generic.go:334] "Generic (PLEG): container finished" podID="0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f" containerID="e112f4f95858aff97866911579a87e1a58622983526305400a4430c3e152d8ff" exitCode=0 Feb 25 07:12:02 crc kubenswrapper[4978]: I0225 07:12:02.775145 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533392-5gp6f" event={"ID":"0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f","Type":"ContainerDied","Data":"e112f4f95858aff97866911579a87e1a58622983526305400a4430c3e152d8ff"} Feb 25 07:12:04 crc kubenswrapper[4978]: I0225 07:12:04.159784 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533392-5gp6f" Feb 25 07:12:04 crc kubenswrapper[4978]: I0225 07:12:04.329833 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58jsg\" (UniqueName: \"kubernetes.io/projected/0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f-kube-api-access-58jsg\") pod \"0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f\" (UID: \"0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f\") " Feb 25 07:12:04 crc kubenswrapper[4978]: I0225 07:12:04.337610 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f-kube-api-access-58jsg" (OuterVolumeSpecName: "kube-api-access-58jsg") pod "0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f" (UID: "0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f"). InnerVolumeSpecName "kube-api-access-58jsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:12:04 crc kubenswrapper[4978]: I0225 07:12:04.432658 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58jsg\" (UniqueName: \"kubernetes.io/projected/0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f-kube-api-access-58jsg\") on node \"crc\" DevicePath \"\"" Feb 25 07:12:04 crc kubenswrapper[4978]: I0225 07:12:04.800458 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533392-5gp6f" event={"ID":"0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f","Type":"ContainerDied","Data":"37eab5fc027645ae49839b55e778c2c51ed390447de3370310386824ac620424"} Feb 25 07:12:04 crc kubenswrapper[4978]: I0225 07:12:04.800519 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37eab5fc027645ae49839b55e778c2c51ed390447de3370310386824ac620424" Feb 25 07:12:04 crc kubenswrapper[4978]: I0225 07:12:04.800554 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533392-5gp6f" Feb 25 07:12:05 crc kubenswrapper[4978]: I0225 07:12:05.255620 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533386-94rxl"] Feb 25 07:12:05 crc kubenswrapper[4978]: I0225 07:12:05.264080 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533386-94rxl"] Feb 25 07:12:05 crc kubenswrapper[4978]: I0225 07:12:05.339833 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fff273d8-333d-435b-86cc-613bb1c53b7c" path="/var/lib/kubelet/pods/fff273d8-333d-435b-86cc-613bb1c53b7c/volumes" Feb 25 07:12:10 crc kubenswrapper[4978]: I0225 07:12:10.327818 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:12:10 crc kubenswrapper[4978]: E0225 07:12:10.328918 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:12:18 crc kubenswrapper[4978]: I0225 07:12:18.693907 4978 scope.go:117] "RemoveContainer" containerID="5b819e72e4b09f85c33501291e203299fcd870de1d74deb4a6f337a81b3e6c57" Feb 25 07:12:18 crc kubenswrapper[4978]: I0225 07:12:18.717593 4978 scope.go:117] "RemoveContainer" containerID="38fe35a66c4eb783bff693706b13667ecdb4496a617dc255fc5aac1f2f5a3bc3" Feb 25 07:12:18 crc kubenswrapper[4978]: I0225 07:12:18.752730 4978 scope.go:117] "RemoveContainer" containerID="9569ef92f4606b61b4ee144b77430b5109d298241ce9e4d4f08627b799dfe74a" Feb 25 07:12:18 crc kubenswrapper[4978]: I0225 07:12:18.781476 4978 scope.go:117] "RemoveContainer" containerID="07a057de05f5bc6d8d6e12685b1aaa0d0969c94b182f9211e4ace6841f755b5b" Feb 25 07:12:18 crc kubenswrapper[4978]: I0225 07:12:18.813100 4978 scope.go:117] "RemoveContainer" containerID="22db8e67ef3c374833a5a7d3778dc19f3cb69c3620c3009d7badbe8c31ef7006" Feb 25 07:12:18 crc kubenswrapper[4978]: I0225 07:12:18.843043 4978 scope.go:117] "RemoveContainer" containerID="62b83f2cbf47e13ff7892d9042365b10f33ef61ca05bc25efd54f299ca6342e7" Feb 25 07:12:18 crc kubenswrapper[4978]: I0225 07:12:18.871201 4978 scope.go:117] "RemoveContainer" containerID="ea0f40bdc05d6dee36735b9c80c86d91e506b9d4f91466ebf56de31a28074aa5" Feb 25 07:12:18 crc kubenswrapper[4978]: I0225 07:12:18.919076 4978 scope.go:117] "RemoveContainer" containerID="a5ccdb669ac03bfcffb881734e74afcc02b508fb8978b1671db8c00879702e58" Feb 25 07:12:18 crc kubenswrapper[4978]: I0225 07:12:18.954503 4978 scope.go:117] "RemoveContainer" containerID="8989205d360d992ab99ca3347077abbe29a1d3dcc44b9b0e61af2f88c88568b8" Feb 25 07:12:25 crc kubenswrapper[4978]: I0225 07:12:25.328365 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:12:25 crc kubenswrapper[4978]: E0225 07:12:25.329616 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:12:38 crc kubenswrapper[4978]: I0225 07:12:38.328950 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:12:38 crc kubenswrapper[4978]: E0225 07:12:38.329679 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:12:52 crc kubenswrapper[4978]: I0225 07:12:52.328620 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:12:52 crc kubenswrapper[4978]: E0225 07:12:52.330553 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:13:05 crc kubenswrapper[4978]: I0225 07:13:05.327253 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:13:05 crc kubenswrapper[4978]: E0225 07:13:05.328244 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:13:18 crc kubenswrapper[4978]: I0225 07:13:18.327836 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:13:18 crc kubenswrapper[4978]: E0225 07:13:18.328851 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.093080 4978 scope.go:117] "RemoveContainer" containerID="cd94dd53f9e56529db66e99175ac2ac797be2dd86cf8eb9e8918194438ca01bf" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.158339 4978 scope.go:117] "RemoveContainer" containerID="51facbbef619d03278afe02c3c8fbe41a2c4e98c1f3161602d2cb500952a63a4" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.185254 4978 scope.go:117] "RemoveContainer" containerID="ea46e9a5ff7500141828c5e39da95b38f857b39bf42022aa0050cb398f5fd324" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.226736 4978 scope.go:117] "RemoveContainer" containerID="5b1a33fcd51e2cec3e100c3d54c2ea14e7ea10f1fc1625a323da5788dd492729" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.253805 4978 scope.go:117] "RemoveContainer" containerID="392249f3ec789cc057ca60e633474a4a6604c180310a30bdc6ddcca23a6427f4" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.278128 4978 scope.go:117] "RemoveContainer" containerID="5520aff92f97a896a5a806d71f4fb6585ad2a8f952fe7c239751ef1d30bae74e" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.304047 4978 scope.go:117] "RemoveContainer" containerID="1d9509bb185bfd8f66b16ea5e86fc0713bd07983268184b5f3f6005b2e7ab1b8" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.329814 4978 scope.go:117] "RemoveContainer" containerID="6341648a449ad9f4c7c225487199328ca24ce228204b741a39e695b53437b28d" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.399794 4978 scope.go:117] "RemoveContainer" containerID="6d5a6575f5371601ecab8ff2f8053d7907cb7b78d92f5ec99f37ea255f9dd065" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.445932 4978 scope.go:117] "RemoveContainer" containerID="2b9d4c187e335c3723ec5d19e25fdb6cedda0a9852c880709cea81dfaf1e5c3b" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.468077 4978 scope.go:117] "RemoveContainer" containerID="caed680d248b73867bcd7477d8eb3c3cdbb4b1d1ceff2297baed0b3c1eed6964" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.510042 4978 scope.go:117] "RemoveContainer" containerID="15a2b4fd13fadf8eed227080adbec60612493687b1b860efa155a9d0a8f50be7" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.551643 4978 scope.go:117] "RemoveContainer" containerID="8032219bff21948de2c2feac0cab9661d2687ea4f33eb4dc5b640069888bce22" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.597936 4978 scope.go:117] "RemoveContainer" containerID="3c6dcd5048df9f2163eaef44d3728e6bf73e055ec2ab84cad23993f70dbd7b2e" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.630480 4978 scope.go:117] "RemoveContainer" containerID="8e2f22cb510c2be5bede5f9167ccd0a70c7f5db1dec604bd4fea4484d24efb52" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.650233 4978 scope.go:117] "RemoveContainer" containerID="3027c6d98887a040746bbb52842379fe73693d836e727a5d3de95c63fc2b5e30" Feb 25 07:13:19 crc kubenswrapper[4978]: I0225 07:13:19.665325 4978 scope.go:117] "RemoveContainer" containerID="6921f9a5eef1ac8d074e2d1d00534d70d54e38bc5537860be66cb698876cf0e1" Feb 25 07:13:29 crc kubenswrapper[4978]: I0225 07:13:29.328206 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:13:29 crc kubenswrapper[4978]: E0225 07:13:29.328913 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:13:44 crc kubenswrapper[4978]: I0225 07:13:44.328325 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:13:44 crc kubenswrapper[4978]: E0225 07:13:44.329310 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:13:59 crc kubenswrapper[4978]: I0225 07:13:59.328710 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:13:59 crc kubenswrapper[4978]: E0225 07:13:59.330158 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.170844 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533394-9dtlp"] Feb 25 07:14:00 crc kubenswrapper[4978]: E0225 07:14:00.171313 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f" containerName="oc" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.171343 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f" containerName="oc" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.171643 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f" containerName="oc" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.175841 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533394-9dtlp" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.181601 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.183548 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.183554 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.195615 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533394-9dtlp"] Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.343023 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjtll\" (UniqueName: \"kubernetes.io/projected/96ffa0c4-8873-43fd-948c-9476708ae2d7-kube-api-access-vjtll\") pod \"auto-csr-approver-29533394-9dtlp\" (UID: \"96ffa0c4-8873-43fd-948c-9476708ae2d7\") " pod="openshift-infra/auto-csr-approver-29533394-9dtlp" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.444822 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjtll\" (UniqueName: \"kubernetes.io/projected/96ffa0c4-8873-43fd-948c-9476708ae2d7-kube-api-access-vjtll\") pod \"auto-csr-approver-29533394-9dtlp\" (UID: \"96ffa0c4-8873-43fd-948c-9476708ae2d7\") " pod="openshift-infra/auto-csr-approver-29533394-9dtlp" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.477923 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjtll\" (UniqueName: \"kubernetes.io/projected/96ffa0c4-8873-43fd-948c-9476708ae2d7-kube-api-access-vjtll\") pod \"auto-csr-approver-29533394-9dtlp\" (UID: \"96ffa0c4-8873-43fd-948c-9476708ae2d7\") " pod="openshift-infra/auto-csr-approver-29533394-9dtlp" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.507570 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533394-9dtlp" Feb 25 07:14:00 crc kubenswrapper[4978]: I0225 07:14:00.798060 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533394-9dtlp"] Feb 25 07:14:01 crc kubenswrapper[4978]: I0225 07:14:01.067344 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533394-9dtlp" event={"ID":"96ffa0c4-8873-43fd-948c-9476708ae2d7","Type":"ContainerStarted","Data":"b783db95f42a3d615aab5728cfacd4c9e3fc30d0c263987f104d916197d188a8"} Feb 25 07:14:03 crc kubenswrapper[4978]: I0225 07:14:03.086439 4978 generic.go:334] "Generic (PLEG): container finished" podID="96ffa0c4-8873-43fd-948c-9476708ae2d7" containerID="1b582cd9e957554846751ed6120adf2b47e2b8bc9eec07aa339deb6a1249ae8a" exitCode=0 Feb 25 07:14:03 crc kubenswrapper[4978]: I0225 07:14:03.086520 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533394-9dtlp" event={"ID":"96ffa0c4-8873-43fd-948c-9476708ae2d7","Type":"ContainerDied","Data":"1b582cd9e957554846751ed6120adf2b47e2b8bc9eec07aa339deb6a1249ae8a"} Feb 25 07:14:04 crc kubenswrapper[4978]: I0225 07:14:04.416639 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533394-9dtlp" Feb 25 07:14:04 crc kubenswrapper[4978]: I0225 07:14:04.519753 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjtll\" (UniqueName: \"kubernetes.io/projected/96ffa0c4-8873-43fd-948c-9476708ae2d7-kube-api-access-vjtll\") pod \"96ffa0c4-8873-43fd-948c-9476708ae2d7\" (UID: \"96ffa0c4-8873-43fd-948c-9476708ae2d7\") " Feb 25 07:14:04 crc kubenswrapper[4978]: I0225 07:14:04.526361 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96ffa0c4-8873-43fd-948c-9476708ae2d7-kube-api-access-vjtll" (OuterVolumeSpecName: "kube-api-access-vjtll") pod "96ffa0c4-8873-43fd-948c-9476708ae2d7" (UID: "96ffa0c4-8873-43fd-948c-9476708ae2d7"). InnerVolumeSpecName "kube-api-access-vjtll". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:14:04 crc kubenswrapper[4978]: I0225 07:14:04.621932 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjtll\" (UniqueName: \"kubernetes.io/projected/96ffa0c4-8873-43fd-948c-9476708ae2d7-kube-api-access-vjtll\") on node \"crc\" DevicePath \"\"" Feb 25 07:14:05 crc kubenswrapper[4978]: I0225 07:14:05.107834 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533394-9dtlp" event={"ID":"96ffa0c4-8873-43fd-948c-9476708ae2d7","Type":"ContainerDied","Data":"b783db95f42a3d615aab5728cfacd4c9e3fc30d0c263987f104d916197d188a8"} Feb 25 07:14:05 crc kubenswrapper[4978]: I0225 07:14:05.107895 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b783db95f42a3d615aab5728cfacd4c9e3fc30d0c263987f104d916197d188a8" Feb 25 07:14:05 crc kubenswrapper[4978]: I0225 07:14:05.107954 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533394-9dtlp" Feb 25 07:14:05 crc kubenswrapper[4978]: I0225 07:14:05.515801 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533388-rqj86"] Feb 25 07:14:05 crc kubenswrapper[4978]: I0225 07:14:05.526197 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533388-rqj86"] Feb 25 07:14:07 crc kubenswrapper[4978]: I0225 07:14:07.344995 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a77f5934-71ee-4680-979b-94d1e4b9feef" path="/var/lib/kubelet/pods/a77f5934-71ee-4680-979b-94d1e4b9feef/volumes" Feb 25 07:14:11 crc kubenswrapper[4978]: I0225 07:14:11.327237 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:14:11 crc kubenswrapper[4978]: E0225 07:14:11.328038 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:14:19 crc kubenswrapper[4978]: I0225 07:14:19.970215 4978 scope.go:117] "RemoveContainer" containerID="e0624efb2273218ddfa7b0599230cfa818769d0d5c002395bf7506f0625b968f" Feb 25 07:14:20 crc kubenswrapper[4978]: I0225 07:14:20.001853 4978 scope.go:117] "RemoveContainer" containerID="fdd96869b4fb12af351379969d60b7fe078542fa9ee866fff1cb955e15ffee92" Feb 25 07:14:20 crc kubenswrapper[4978]: I0225 07:14:20.058503 4978 scope.go:117] "RemoveContainer" containerID="6f42f16d8d601f4d2968ebd172a034ee1578a4272e4611cdbe696202023e8aad" Feb 25 07:14:20 crc kubenswrapper[4978]: I0225 07:14:20.084350 4978 scope.go:117] "RemoveContainer" containerID="d09e81920a08e75b18da1c30ea69b3f15f59cf1ccbbf32b83855123548ed4933" Feb 25 07:14:20 crc kubenswrapper[4978]: I0225 07:14:20.116262 4978 scope.go:117] "RemoveContainer" containerID="d795d32225060d81627ee5c37db3d8d00175f8bc43ea39f4af9cbf6719fd2309" Feb 25 07:14:20 crc kubenswrapper[4978]: I0225 07:14:20.151866 4978 scope.go:117] "RemoveContainer" containerID="883a527a9d84208558ca0007c0a8c77189f06bf0927bb703145ef20bcf5a9c9b" Feb 25 07:14:20 crc kubenswrapper[4978]: I0225 07:14:20.199542 4978 scope.go:117] "RemoveContainer" containerID="61c18e6841b098b42dd54f54e6c6ae0ce96ad977d7e9e58e7eafa71aa797add1" Feb 25 07:14:20 crc kubenswrapper[4978]: I0225 07:14:20.216201 4978 scope.go:117] "RemoveContainer" containerID="09f6efb9225a38d05749f4a3b1a4a38d04bdd5f6081c1c1cc01280de268c9d65" Feb 25 07:14:20 crc kubenswrapper[4978]: I0225 07:14:20.235878 4978 scope.go:117] "RemoveContainer" containerID="04bf0ba3b369628fea98361e5da48ba3bc1bc5dde35d641c7aaf36fd44c0ff1c" Feb 25 07:14:24 crc kubenswrapper[4978]: I0225 07:14:24.328059 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:14:24 crc kubenswrapper[4978]: E0225 07:14:24.329018 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:14:36 crc kubenswrapper[4978]: I0225 07:14:36.327860 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:14:36 crc kubenswrapper[4978]: E0225 07:14:36.328742 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:14:49 crc kubenswrapper[4978]: I0225 07:14:49.328468 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:14:49 crc kubenswrapper[4978]: E0225 07:14:49.329308 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.156904 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6"] Feb 25 07:15:00 crc kubenswrapper[4978]: E0225 07:15:00.158123 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96ffa0c4-8873-43fd-948c-9476708ae2d7" containerName="oc" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.158143 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="96ffa0c4-8873-43fd-948c-9476708ae2d7" containerName="oc" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.158666 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="96ffa0c4-8873-43fd-948c-9476708ae2d7" containerName="oc" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.159396 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.164255 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.164293 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.171809 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6"] Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.237294 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fada399d-f50f-4e12-bceb-dc03a02b25c9-secret-volume\") pod \"collect-profiles-29533395-qmcs6\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.237395 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fada399d-f50f-4e12-bceb-dc03a02b25c9-config-volume\") pod \"collect-profiles-29533395-qmcs6\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.237534 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqxcz\" (UniqueName: \"kubernetes.io/projected/fada399d-f50f-4e12-bceb-dc03a02b25c9-kube-api-access-lqxcz\") pod \"collect-profiles-29533395-qmcs6\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.340447 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqxcz\" (UniqueName: \"kubernetes.io/projected/fada399d-f50f-4e12-bceb-dc03a02b25c9-kube-api-access-lqxcz\") pod \"collect-profiles-29533395-qmcs6\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.340593 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fada399d-f50f-4e12-bceb-dc03a02b25c9-secret-volume\") pod \"collect-profiles-29533395-qmcs6\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.340661 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fada399d-f50f-4e12-bceb-dc03a02b25c9-config-volume\") pod \"collect-profiles-29533395-qmcs6\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.342496 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fada399d-f50f-4e12-bceb-dc03a02b25c9-config-volume\") pod \"collect-profiles-29533395-qmcs6\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.352631 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fada399d-f50f-4e12-bceb-dc03a02b25c9-secret-volume\") pod \"collect-profiles-29533395-qmcs6\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.374337 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqxcz\" (UniqueName: \"kubernetes.io/projected/fada399d-f50f-4e12-bceb-dc03a02b25c9-kube-api-access-lqxcz\") pod \"collect-profiles-29533395-qmcs6\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.483676 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:00 crc kubenswrapper[4978]: I0225 07:15:00.745234 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6"] Feb 25 07:15:01 crc kubenswrapper[4978]: I0225 07:15:01.689016 4978 generic.go:334] "Generic (PLEG): container finished" podID="fada399d-f50f-4e12-bceb-dc03a02b25c9" containerID="a7950e92dd757d74be2aa8134bb0bfe21452f72b9cf5dfadecd2140af87a1cf7" exitCode=0 Feb 25 07:15:01 crc kubenswrapper[4978]: I0225 07:15:01.689109 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" event={"ID":"fada399d-f50f-4e12-bceb-dc03a02b25c9","Type":"ContainerDied","Data":"a7950e92dd757d74be2aa8134bb0bfe21452f72b9cf5dfadecd2140af87a1cf7"} Feb 25 07:15:01 crc kubenswrapper[4978]: I0225 07:15:01.689345 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" event={"ID":"fada399d-f50f-4e12-bceb-dc03a02b25c9","Type":"ContainerStarted","Data":"f40073892673e51a01215ff7fc1821be2b1f389fcdc5a4af3bea1084aa204f61"} Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.082398 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.284126 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fada399d-f50f-4e12-bceb-dc03a02b25c9-config-volume\") pod \"fada399d-f50f-4e12-bceb-dc03a02b25c9\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.284314 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqxcz\" (UniqueName: \"kubernetes.io/projected/fada399d-f50f-4e12-bceb-dc03a02b25c9-kube-api-access-lqxcz\") pod \"fada399d-f50f-4e12-bceb-dc03a02b25c9\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.284353 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fada399d-f50f-4e12-bceb-dc03a02b25c9-secret-volume\") pod \"fada399d-f50f-4e12-bceb-dc03a02b25c9\" (UID: \"fada399d-f50f-4e12-bceb-dc03a02b25c9\") " Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.285000 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fada399d-f50f-4e12-bceb-dc03a02b25c9-config-volume" (OuterVolumeSpecName: "config-volume") pod "fada399d-f50f-4e12-bceb-dc03a02b25c9" (UID: "fada399d-f50f-4e12-bceb-dc03a02b25c9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.292508 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fada399d-f50f-4e12-bceb-dc03a02b25c9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fada399d-f50f-4e12-bceb-dc03a02b25c9" (UID: "fada399d-f50f-4e12-bceb-dc03a02b25c9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.292542 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fada399d-f50f-4e12-bceb-dc03a02b25c9-kube-api-access-lqxcz" (OuterVolumeSpecName: "kube-api-access-lqxcz") pod "fada399d-f50f-4e12-bceb-dc03a02b25c9" (UID: "fada399d-f50f-4e12-bceb-dc03a02b25c9"). InnerVolumeSpecName "kube-api-access-lqxcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.327362 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:15:03 crc kubenswrapper[4978]: E0225 07:15:03.327954 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.386991 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqxcz\" (UniqueName: \"kubernetes.io/projected/fada399d-f50f-4e12-bceb-dc03a02b25c9-kube-api-access-lqxcz\") on node \"crc\" DevicePath \"\"" Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.387028 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fada399d-f50f-4e12-bceb-dc03a02b25c9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.387041 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fada399d-f50f-4e12-bceb-dc03a02b25c9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.710256 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" event={"ID":"fada399d-f50f-4e12-bceb-dc03a02b25c9","Type":"ContainerDied","Data":"f40073892673e51a01215ff7fc1821be2b1f389fcdc5a4af3bea1084aa204f61"} Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.710762 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f40073892673e51a01215ff7fc1821be2b1f389fcdc5a4af3bea1084aa204f61" Feb 25 07:15:03 crc kubenswrapper[4978]: I0225 07:15:03.710354 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6" Feb 25 07:15:13 crc kubenswrapper[4978]: I0225 07:15:13.546554 4978 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-9d8lt container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 07:15:13 crc kubenswrapper[4978]: I0225 07:15:13.547192 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" podUID="d4b61b96-afd0-47a6-be02-ead379d948b2" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 07:15:13 crc kubenswrapper[4978]: I0225 07:15:13.556837 4978 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-9d8lt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 25 07:15:13 crc kubenswrapper[4978]: I0225 07:15:13.556932 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-9d8lt" podUID="d4b61b96-afd0-47a6-be02-ead379d948b2" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 07:15:14 crc kubenswrapper[4978]: I0225 07:15:14.328189 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:15:14 crc kubenswrapper[4978]: E0225 07:15:14.328710 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:15:20 crc kubenswrapper[4978]: I0225 07:15:20.439362 4978 scope.go:117] "RemoveContainer" containerID="e2d52313a54ebf436eac7928a9eab7d77dd7826b23c86f93edd92380125e47dd" Feb 25 07:15:20 crc kubenswrapper[4978]: I0225 07:15:20.475601 4978 scope.go:117] "RemoveContainer" containerID="348413bef14e6c205395d2c485c730cf829af44894b23680e9874614a44661ea" Feb 25 07:15:20 crc kubenswrapper[4978]: I0225 07:15:20.549110 4978 scope.go:117] "RemoveContainer" containerID="e227b0a33aecace45163696f40de57e5900930437a82fb19dd2b00654d1ceff6" Feb 25 07:15:20 crc kubenswrapper[4978]: I0225 07:15:20.573665 4978 scope.go:117] "RemoveContainer" containerID="d6443061bb18b3e732682398607026b9a8bfff4e6eda5690b8cd984c7e598a37" Feb 25 07:15:20 crc kubenswrapper[4978]: I0225 07:15:20.625057 4978 scope.go:117] "RemoveContainer" containerID="e36db710c1ab7f0d8cb9355b7504726128ca6de40de4ab2bc21a32b4e767925d" Feb 25 07:15:28 crc kubenswrapper[4978]: I0225 07:15:28.328220 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:15:28 crc kubenswrapper[4978]: E0225 07:15:28.329224 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:15:39 crc kubenswrapper[4978]: I0225 07:15:39.327871 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:15:39 crc kubenswrapper[4978]: E0225 07:15:39.328468 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:15:53 crc kubenswrapper[4978]: I0225 07:15:53.327611 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:15:53 crc kubenswrapper[4978]: E0225 07:15:53.330816 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.166986 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533396-bcqvb"] Feb 25 07:16:00 crc kubenswrapper[4978]: E0225 07:16:00.169321 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fada399d-f50f-4e12-bceb-dc03a02b25c9" containerName="collect-profiles" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.169352 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="fada399d-f50f-4e12-bceb-dc03a02b25c9" containerName="collect-profiles" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.169689 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="fada399d-f50f-4e12-bceb-dc03a02b25c9" containerName="collect-profiles" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.170435 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533396-bcqvb" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.173784 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.173833 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.174069 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.180958 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533396-bcqvb"] Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.345264 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgtsp\" (UniqueName: \"kubernetes.io/projected/7c1916c8-8177-427f-86fb-2e70a3217f3b-kube-api-access-qgtsp\") pod \"auto-csr-approver-29533396-bcqvb\" (UID: \"7c1916c8-8177-427f-86fb-2e70a3217f3b\") " pod="openshift-infra/auto-csr-approver-29533396-bcqvb" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.447703 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgtsp\" (UniqueName: \"kubernetes.io/projected/7c1916c8-8177-427f-86fb-2e70a3217f3b-kube-api-access-qgtsp\") pod \"auto-csr-approver-29533396-bcqvb\" (UID: \"7c1916c8-8177-427f-86fb-2e70a3217f3b\") " pod="openshift-infra/auto-csr-approver-29533396-bcqvb" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.483482 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgtsp\" (UniqueName: \"kubernetes.io/projected/7c1916c8-8177-427f-86fb-2e70a3217f3b-kube-api-access-qgtsp\") pod \"auto-csr-approver-29533396-bcqvb\" (UID: \"7c1916c8-8177-427f-86fb-2e70a3217f3b\") " pod="openshift-infra/auto-csr-approver-29533396-bcqvb" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.502439 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533396-bcqvb" Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.795996 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533396-bcqvb"] Feb 25 07:16:00 crc kubenswrapper[4978]: I0225 07:16:00.805636 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 07:16:01 crc kubenswrapper[4978]: I0225 07:16:01.257092 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533396-bcqvb" event={"ID":"7c1916c8-8177-427f-86fb-2e70a3217f3b","Type":"ContainerStarted","Data":"8cd9358c6e793aeec4e7f62e6eb7622d51fd2ffd0be81ed7b9fcf0283e6af292"} Feb 25 07:16:02 crc kubenswrapper[4978]: I0225 07:16:02.267203 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533396-bcqvb" event={"ID":"7c1916c8-8177-427f-86fb-2e70a3217f3b","Type":"ContainerStarted","Data":"e8df7b3ef6c0b2c8d219d3cb017485fe2e60d345d87e97a37631514fe68a6007"} Feb 25 07:16:02 crc kubenswrapper[4978]: I0225 07:16:02.286107 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533396-bcqvb" podStartSLOduration=1.27998152 podStartE2EDuration="2.286086882s" podCreationTimestamp="2026-02-25 07:16:00 +0000 UTC" firstStartedPulling="2026-02-25 07:16:00.805249591 +0000 UTC m=+1854.244506060" lastFinishedPulling="2026-02-25 07:16:01.811354923 +0000 UTC m=+1855.250611422" observedRunningTime="2026-02-25 07:16:02.278694871 +0000 UTC m=+1855.717951360" watchObservedRunningTime="2026-02-25 07:16:02.286086882 +0000 UTC m=+1855.725343351" Feb 25 07:16:03 crc kubenswrapper[4978]: I0225 07:16:03.277522 4978 generic.go:334] "Generic (PLEG): container finished" podID="7c1916c8-8177-427f-86fb-2e70a3217f3b" containerID="e8df7b3ef6c0b2c8d219d3cb017485fe2e60d345d87e97a37631514fe68a6007" exitCode=0 Feb 25 07:16:03 crc kubenswrapper[4978]: I0225 07:16:03.277630 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533396-bcqvb" event={"ID":"7c1916c8-8177-427f-86fb-2e70a3217f3b","Type":"ContainerDied","Data":"e8df7b3ef6c0b2c8d219d3cb017485fe2e60d345d87e97a37631514fe68a6007"} Feb 25 07:16:04 crc kubenswrapper[4978]: I0225 07:16:04.678609 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533396-bcqvb" Feb 25 07:16:04 crc kubenswrapper[4978]: I0225 07:16:04.842123 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgtsp\" (UniqueName: \"kubernetes.io/projected/7c1916c8-8177-427f-86fb-2e70a3217f3b-kube-api-access-qgtsp\") pod \"7c1916c8-8177-427f-86fb-2e70a3217f3b\" (UID: \"7c1916c8-8177-427f-86fb-2e70a3217f3b\") " Feb 25 07:16:04 crc kubenswrapper[4978]: I0225 07:16:04.850792 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c1916c8-8177-427f-86fb-2e70a3217f3b-kube-api-access-qgtsp" (OuterVolumeSpecName: "kube-api-access-qgtsp") pod "7c1916c8-8177-427f-86fb-2e70a3217f3b" (UID: "7c1916c8-8177-427f-86fb-2e70a3217f3b"). InnerVolumeSpecName "kube-api-access-qgtsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:16:04 crc kubenswrapper[4978]: I0225 07:16:04.945603 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgtsp\" (UniqueName: \"kubernetes.io/projected/7c1916c8-8177-427f-86fb-2e70a3217f3b-kube-api-access-qgtsp\") on node \"crc\" DevicePath \"\"" Feb 25 07:16:05 crc kubenswrapper[4978]: I0225 07:16:05.302450 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533396-bcqvb" event={"ID":"7c1916c8-8177-427f-86fb-2e70a3217f3b","Type":"ContainerDied","Data":"8cd9358c6e793aeec4e7f62e6eb7622d51fd2ffd0be81ed7b9fcf0283e6af292"} Feb 25 07:16:05 crc kubenswrapper[4978]: I0225 07:16:05.302510 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cd9358c6e793aeec4e7f62e6eb7622d51fd2ffd0be81ed7b9fcf0283e6af292" Feb 25 07:16:05 crc kubenswrapper[4978]: I0225 07:16:05.302614 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533396-bcqvb" Feb 25 07:16:05 crc kubenswrapper[4978]: I0225 07:16:05.327761 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:16:05 crc kubenswrapper[4978]: E0225 07:16:05.328338 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:16:05 crc kubenswrapper[4978]: I0225 07:16:05.363622 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533390-2vswx"] Feb 25 07:16:05 crc kubenswrapper[4978]: I0225 07:16:05.373613 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533390-2vswx"] Feb 25 07:16:07 crc kubenswrapper[4978]: I0225 07:16:07.350167 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c029995-b152-4e0a-84f3-0e9c617ee812" path="/var/lib/kubelet/pods/0c029995-b152-4e0a-84f3-0e9c617ee812/volumes" Feb 25 07:16:16 crc kubenswrapper[4978]: I0225 07:16:16.327346 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:16:16 crc kubenswrapper[4978]: E0225 07:16:16.327946 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:16:20 crc kubenswrapper[4978]: I0225 07:16:20.784175 4978 scope.go:117] "RemoveContainer" containerID="c61141b61f21cb127aa4ec4c820436879baedf614cff727d7cb229d414bd57ad" Feb 25 07:16:20 crc kubenswrapper[4978]: I0225 07:16:20.869014 4978 scope.go:117] "RemoveContainer" containerID="09b0228a0ba6058441efd4b4d0555b7a4cdbaf500b54897828dd5c3c56a204ef" Feb 25 07:16:29 crc kubenswrapper[4978]: I0225 07:16:29.328007 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:16:29 crc kubenswrapper[4978]: I0225 07:16:29.533488 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"4d066de9c5cf5b617ddc67630504439caf4b63e08b3c93deb65061b29eb89656"} Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.169118 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533398-85vk8"] Feb 25 07:18:00 crc kubenswrapper[4978]: E0225 07:18:00.171890 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c1916c8-8177-427f-86fb-2e70a3217f3b" containerName="oc" Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.172278 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c1916c8-8177-427f-86fb-2e70a3217f3b" containerName="oc" Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.172884 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c1916c8-8177-427f-86fb-2e70a3217f3b" containerName="oc" Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.174120 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533398-85vk8" Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.176535 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.176641 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.177192 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.183284 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533398-85vk8"] Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.313998 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8jsv\" (UniqueName: \"kubernetes.io/projected/646a0605-333d-4033-ade8-5309b201ecf3-kube-api-access-b8jsv\") pod \"auto-csr-approver-29533398-85vk8\" (UID: \"646a0605-333d-4033-ade8-5309b201ecf3\") " pod="openshift-infra/auto-csr-approver-29533398-85vk8" Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.415328 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8jsv\" (UniqueName: \"kubernetes.io/projected/646a0605-333d-4033-ade8-5309b201ecf3-kube-api-access-b8jsv\") pod \"auto-csr-approver-29533398-85vk8\" (UID: \"646a0605-333d-4033-ade8-5309b201ecf3\") " pod="openshift-infra/auto-csr-approver-29533398-85vk8" Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.442528 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8jsv\" (UniqueName: \"kubernetes.io/projected/646a0605-333d-4033-ade8-5309b201ecf3-kube-api-access-b8jsv\") pod \"auto-csr-approver-29533398-85vk8\" (UID: \"646a0605-333d-4033-ade8-5309b201ecf3\") " pod="openshift-infra/auto-csr-approver-29533398-85vk8" Feb 25 07:18:00 crc kubenswrapper[4978]: I0225 07:18:00.540088 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533398-85vk8" Feb 25 07:18:01 crc kubenswrapper[4978]: I0225 07:18:01.075604 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533398-85vk8"] Feb 25 07:18:01 crc kubenswrapper[4978]: I0225 07:18:01.385990 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533398-85vk8" event={"ID":"646a0605-333d-4033-ade8-5309b201ecf3","Type":"ContainerStarted","Data":"c86b60585cf0e59b7326fa5ea93d0e6cb862a91527a21d1664ffdee8998b6ee0"} Feb 25 07:18:03 crc kubenswrapper[4978]: I0225 07:18:03.404865 4978 generic.go:334] "Generic (PLEG): container finished" podID="646a0605-333d-4033-ade8-5309b201ecf3" containerID="339fa3d72c25703e1c07c7863cccf0f22bc75972f7ad4dff7c9f150d28757970" exitCode=0 Feb 25 07:18:03 crc kubenswrapper[4978]: I0225 07:18:03.404951 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533398-85vk8" event={"ID":"646a0605-333d-4033-ade8-5309b201ecf3","Type":"ContainerDied","Data":"339fa3d72c25703e1c07c7863cccf0f22bc75972f7ad4dff7c9f150d28757970"} Feb 25 07:18:04 crc kubenswrapper[4978]: I0225 07:18:04.803830 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533398-85vk8" Feb 25 07:18:04 crc kubenswrapper[4978]: I0225 07:18:04.916180 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8jsv\" (UniqueName: \"kubernetes.io/projected/646a0605-333d-4033-ade8-5309b201ecf3-kube-api-access-b8jsv\") pod \"646a0605-333d-4033-ade8-5309b201ecf3\" (UID: \"646a0605-333d-4033-ade8-5309b201ecf3\") " Feb 25 07:18:04 crc kubenswrapper[4978]: I0225 07:18:04.924067 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/646a0605-333d-4033-ade8-5309b201ecf3-kube-api-access-b8jsv" (OuterVolumeSpecName: "kube-api-access-b8jsv") pod "646a0605-333d-4033-ade8-5309b201ecf3" (UID: "646a0605-333d-4033-ade8-5309b201ecf3"). InnerVolumeSpecName "kube-api-access-b8jsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:18:05 crc kubenswrapper[4978]: I0225 07:18:05.018980 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8jsv\" (UniqueName: \"kubernetes.io/projected/646a0605-333d-4033-ade8-5309b201ecf3-kube-api-access-b8jsv\") on node \"crc\" DevicePath \"\"" Feb 25 07:18:05 crc kubenswrapper[4978]: I0225 07:18:05.421582 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533398-85vk8" event={"ID":"646a0605-333d-4033-ade8-5309b201ecf3","Type":"ContainerDied","Data":"c86b60585cf0e59b7326fa5ea93d0e6cb862a91527a21d1664ffdee8998b6ee0"} Feb 25 07:18:05 crc kubenswrapper[4978]: I0225 07:18:05.421864 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c86b60585cf0e59b7326fa5ea93d0e6cb862a91527a21d1664ffdee8998b6ee0" Feb 25 07:18:05 crc kubenswrapper[4978]: I0225 07:18:05.421809 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533398-85vk8" Feb 25 07:18:05 crc kubenswrapper[4978]: I0225 07:18:05.891248 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533392-5gp6f"] Feb 25 07:18:05 crc kubenswrapper[4978]: I0225 07:18:05.900107 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533392-5gp6f"] Feb 25 07:18:07 crc kubenswrapper[4978]: I0225 07:18:07.346726 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f" path="/var/lib/kubelet/pods/0de7f76a-e2aa-4040-ae0b-5f65c8e6d40f/volumes" Feb 25 07:18:20 crc kubenswrapper[4978]: I0225 07:18:20.967563 4978 scope.go:117] "RemoveContainer" containerID="443289f7d81ce2303786095bbcd980c547065a02cf37cfafcf7fc57dabb2fc54" Feb 25 07:18:21 crc kubenswrapper[4978]: I0225 07:18:21.003960 4978 scope.go:117] "RemoveContainer" containerID="9fae477b6572ba1d21033dc861962d4c682f77d7ffd70b7de1f3bc573df560a8" Feb 25 07:18:21 crc kubenswrapper[4978]: I0225 07:18:21.040715 4978 scope.go:117] "RemoveContainer" containerID="e112f4f95858aff97866911579a87e1a58622983526305400a4430c3e152d8ff" Feb 25 07:18:21 crc kubenswrapper[4978]: I0225 07:18:21.091799 4978 scope.go:117] "RemoveContainer" containerID="cdf774c0187d2f474f7178bb1e55466ca1ff546c3d6382e37a85ae569ebe6d98" Feb 25 07:18:46 crc kubenswrapper[4978]: I0225 07:18:46.540903 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:18:46 crc kubenswrapper[4978]: I0225 07:18:46.541531 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:19:16 crc kubenswrapper[4978]: I0225 07:19:16.540896 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:19:16 crc kubenswrapper[4978]: I0225 07:19:16.541648 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:19:46 crc kubenswrapper[4978]: I0225 07:19:46.540965 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:19:46 crc kubenswrapper[4978]: I0225 07:19:46.541654 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:19:46 crc kubenswrapper[4978]: I0225 07:19:46.541722 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:19:46 crc kubenswrapper[4978]: I0225 07:19:46.542905 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4d066de9c5cf5b617ddc67630504439caf4b63e08b3c93deb65061b29eb89656"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:19:46 crc kubenswrapper[4978]: I0225 07:19:46.543002 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://4d066de9c5cf5b617ddc67630504439caf4b63e08b3c93deb65061b29eb89656" gracePeriod=600 Feb 25 07:19:47 crc kubenswrapper[4978]: I0225 07:19:47.396180 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="4d066de9c5cf5b617ddc67630504439caf4b63e08b3c93deb65061b29eb89656" exitCode=0 Feb 25 07:19:47 crc kubenswrapper[4978]: I0225 07:19:47.396309 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"4d066de9c5cf5b617ddc67630504439caf4b63e08b3c93deb65061b29eb89656"} Feb 25 07:19:47 crc kubenswrapper[4978]: I0225 07:19:47.397127 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c"} Feb 25 07:19:47 crc kubenswrapper[4978]: I0225 07:19:47.397170 4978 scope.go:117] "RemoveContainer" containerID="158a920977592f371e64c7e268d06c7ab87cc2355a70ff37783e063ed48674e5" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.166577 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533400-qgkvz"] Feb 25 07:20:00 crc kubenswrapper[4978]: E0225 07:20:00.169630 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="646a0605-333d-4033-ade8-5309b201ecf3" containerName="oc" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.169806 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="646a0605-333d-4033-ade8-5309b201ecf3" containerName="oc" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.170146 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="646a0605-333d-4033-ade8-5309b201ecf3" containerName="oc" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.171001 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533400-qgkvz" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.173948 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.179023 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533400-qgkvz"] Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.181797 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.183047 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.257004 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lthnq\" (UniqueName: \"kubernetes.io/projected/dca72eca-d532-4e96-a7c1-ef885235eab2-kube-api-access-lthnq\") pod \"auto-csr-approver-29533400-qgkvz\" (UID: \"dca72eca-d532-4e96-a7c1-ef885235eab2\") " pod="openshift-infra/auto-csr-approver-29533400-qgkvz" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.358462 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lthnq\" (UniqueName: \"kubernetes.io/projected/dca72eca-d532-4e96-a7c1-ef885235eab2-kube-api-access-lthnq\") pod \"auto-csr-approver-29533400-qgkvz\" (UID: \"dca72eca-d532-4e96-a7c1-ef885235eab2\") " pod="openshift-infra/auto-csr-approver-29533400-qgkvz" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.399631 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lthnq\" (UniqueName: \"kubernetes.io/projected/dca72eca-d532-4e96-a7c1-ef885235eab2-kube-api-access-lthnq\") pod \"auto-csr-approver-29533400-qgkvz\" (UID: \"dca72eca-d532-4e96-a7c1-ef885235eab2\") " pod="openshift-infra/auto-csr-approver-29533400-qgkvz" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.510105 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533400-qgkvz" Feb 25 07:20:00 crc kubenswrapper[4978]: I0225 07:20:00.810488 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533400-qgkvz"] Feb 25 07:20:01 crc kubenswrapper[4978]: I0225 07:20:01.573608 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533400-qgkvz" event={"ID":"dca72eca-d532-4e96-a7c1-ef885235eab2","Type":"ContainerStarted","Data":"8cd59f6b6aa006b3542a0e288b62044126ab1f90c3516e7d4741934e9bc78dde"} Feb 25 07:20:02 crc kubenswrapper[4978]: I0225 07:20:02.584363 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533400-qgkvz" event={"ID":"dca72eca-d532-4e96-a7c1-ef885235eab2","Type":"ContainerStarted","Data":"dc5c78c9f8b9657b1d4706fbe0a6320719bf4b7fc37bfd9715012051e712e823"} Feb 25 07:20:02 crc kubenswrapper[4978]: I0225 07:20:02.608725 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533400-qgkvz" podStartSLOduration=1.493996558 podStartE2EDuration="2.60867618s" podCreationTimestamp="2026-02-25 07:20:00 +0000 UTC" firstStartedPulling="2026-02-25 07:20:00.822519021 +0000 UTC m=+2094.261775520" lastFinishedPulling="2026-02-25 07:20:01.937198653 +0000 UTC m=+2095.376455142" observedRunningTime="2026-02-25 07:20:02.604734899 +0000 UTC m=+2096.043991368" watchObservedRunningTime="2026-02-25 07:20:02.60867618 +0000 UTC m=+2096.047932649" Feb 25 07:20:03 crc kubenswrapper[4978]: I0225 07:20:03.597638 4978 generic.go:334] "Generic (PLEG): container finished" podID="dca72eca-d532-4e96-a7c1-ef885235eab2" containerID="dc5c78c9f8b9657b1d4706fbe0a6320719bf4b7fc37bfd9715012051e712e823" exitCode=0 Feb 25 07:20:03 crc kubenswrapper[4978]: I0225 07:20:03.597719 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533400-qgkvz" event={"ID":"dca72eca-d532-4e96-a7c1-ef885235eab2","Type":"ContainerDied","Data":"dc5c78c9f8b9657b1d4706fbe0a6320719bf4b7fc37bfd9715012051e712e823"} Feb 25 07:20:04 crc kubenswrapper[4978]: I0225 07:20:04.990092 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533400-qgkvz" Feb 25 07:20:05 crc kubenswrapper[4978]: I0225 07:20:05.170571 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lthnq\" (UniqueName: \"kubernetes.io/projected/dca72eca-d532-4e96-a7c1-ef885235eab2-kube-api-access-lthnq\") pod \"dca72eca-d532-4e96-a7c1-ef885235eab2\" (UID: \"dca72eca-d532-4e96-a7c1-ef885235eab2\") " Feb 25 07:20:05 crc kubenswrapper[4978]: I0225 07:20:05.179929 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dca72eca-d532-4e96-a7c1-ef885235eab2-kube-api-access-lthnq" (OuterVolumeSpecName: "kube-api-access-lthnq") pod "dca72eca-d532-4e96-a7c1-ef885235eab2" (UID: "dca72eca-d532-4e96-a7c1-ef885235eab2"). InnerVolumeSpecName "kube-api-access-lthnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:20:05 crc kubenswrapper[4978]: I0225 07:20:05.272830 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lthnq\" (UniqueName: \"kubernetes.io/projected/dca72eca-d532-4e96-a7c1-ef885235eab2-kube-api-access-lthnq\") on node \"crc\" DevicePath \"\"" Feb 25 07:20:05 crc kubenswrapper[4978]: I0225 07:20:05.619876 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533400-qgkvz" event={"ID":"dca72eca-d532-4e96-a7c1-ef885235eab2","Type":"ContainerDied","Data":"8cd59f6b6aa006b3542a0e288b62044126ab1f90c3516e7d4741934e9bc78dde"} Feb 25 07:20:05 crc kubenswrapper[4978]: I0225 07:20:05.620224 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cd59f6b6aa006b3542a0e288b62044126ab1f90c3516e7d4741934e9bc78dde" Feb 25 07:20:05 crc kubenswrapper[4978]: I0225 07:20:05.620006 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533400-qgkvz" Feb 25 07:20:05 crc kubenswrapper[4978]: I0225 07:20:05.695852 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533394-9dtlp"] Feb 25 07:20:05 crc kubenswrapper[4978]: I0225 07:20:05.704944 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533394-9dtlp"] Feb 25 07:20:07 crc kubenswrapper[4978]: I0225 07:20:07.348617 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96ffa0c4-8873-43fd-948c-9476708ae2d7" path="/var/lib/kubelet/pods/96ffa0c4-8873-43fd-948c-9476708ae2d7/volumes" Feb 25 07:20:21 crc kubenswrapper[4978]: I0225 07:20:21.189827 4978 scope.go:117] "RemoveContainer" containerID="1b582cd9e957554846751ed6120adf2b47e2b8bc9eec07aa339deb6a1249ae8a" Feb 25 07:20:30 crc kubenswrapper[4978]: I0225 07:20:30.983835 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nwwr9"] Feb 25 07:20:30 crc kubenswrapper[4978]: E0225 07:20:30.985039 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dca72eca-d532-4e96-a7c1-ef885235eab2" containerName="oc" Feb 25 07:20:30 crc kubenswrapper[4978]: I0225 07:20:30.985061 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="dca72eca-d532-4e96-a7c1-ef885235eab2" containerName="oc" Feb 25 07:20:30 crc kubenswrapper[4978]: I0225 07:20:30.985325 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="dca72eca-d532-4e96-a7c1-ef885235eab2" containerName="oc" Feb 25 07:20:30 crc kubenswrapper[4978]: I0225 07:20:30.986904 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.004037 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nwwr9"] Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.023296 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-utilities\") pod \"certified-operators-nwwr9\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.023341 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-catalog-content\") pod \"certified-operators-nwwr9\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.023399 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgfc2\" (UniqueName: \"kubernetes.io/projected/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-kube-api-access-qgfc2\") pod \"certified-operators-nwwr9\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.125320 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-utilities\") pod \"certified-operators-nwwr9\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.125878 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-catalog-content\") pod \"certified-operators-nwwr9\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.125886 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-utilities\") pod \"certified-operators-nwwr9\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.125986 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgfc2\" (UniqueName: \"kubernetes.io/projected/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-kube-api-access-qgfc2\") pod \"certified-operators-nwwr9\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.126137 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-catalog-content\") pod \"certified-operators-nwwr9\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.167121 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgfc2\" (UniqueName: \"kubernetes.io/projected/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-kube-api-access-qgfc2\") pod \"certified-operators-nwwr9\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.323732 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.816852 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nwwr9"] Feb 25 07:20:31 crc kubenswrapper[4978]: I0225 07:20:31.883034 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwwr9" event={"ID":"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff","Type":"ContainerStarted","Data":"9bac56112b320436e3739e7beca651fb385cc584ff177b7676c48f54e9b3bfca"} Feb 25 07:20:32 crc kubenswrapper[4978]: I0225 07:20:32.896899 4978 generic.go:334] "Generic (PLEG): container finished" podID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerID="831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3" exitCode=0 Feb 25 07:20:32 crc kubenswrapper[4978]: I0225 07:20:32.896968 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwwr9" event={"ID":"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff","Type":"ContainerDied","Data":"831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3"} Feb 25 07:20:33 crc kubenswrapper[4978]: I0225 07:20:33.908221 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwwr9" event={"ID":"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff","Type":"ContainerStarted","Data":"bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112"} Feb 25 07:20:34 crc kubenswrapper[4978]: I0225 07:20:34.920635 4978 generic.go:334] "Generic (PLEG): container finished" podID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerID="bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112" exitCode=0 Feb 25 07:20:34 crc kubenswrapper[4978]: I0225 07:20:34.920753 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwwr9" event={"ID":"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff","Type":"ContainerDied","Data":"bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112"} Feb 25 07:20:35 crc kubenswrapper[4978]: I0225 07:20:35.930850 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwwr9" event={"ID":"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff","Type":"ContainerStarted","Data":"f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7"} Feb 25 07:20:35 crc kubenswrapper[4978]: I0225 07:20:35.951041 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nwwr9" podStartSLOduration=3.542465371 podStartE2EDuration="5.951027326s" podCreationTimestamp="2026-02-25 07:20:30 +0000 UTC" firstStartedPulling="2026-02-25 07:20:32.908364896 +0000 UTC m=+2126.347654356" lastFinishedPulling="2026-02-25 07:20:35.316959812 +0000 UTC m=+2128.756216311" observedRunningTime="2026-02-25 07:20:35.945531587 +0000 UTC m=+2129.384788146" watchObservedRunningTime="2026-02-25 07:20:35.951027326 +0000 UTC m=+2129.390283785" Feb 25 07:20:41 crc kubenswrapper[4978]: I0225 07:20:41.324645 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:41 crc kubenswrapper[4978]: I0225 07:20:41.325363 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:41 crc kubenswrapper[4978]: I0225 07:20:41.393317 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:42 crc kubenswrapper[4978]: I0225 07:20:42.047413 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:42 crc kubenswrapper[4978]: I0225 07:20:42.155223 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nwwr9"] Feb 25 07:20:44 crc kubenswrapper[4978]: I0225 07:20:44.005775 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nwwr9" podUID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerName="registry-server" containerID="cri-o://f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7" gracePeriod=2 Feb 25 07:20:44 crc kubenswrapper[4978]: I0225 07:20:44.510950 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:44 crc kubenswrapper[4978]: I0225 07:20:44.641764 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgfc2\" (UniqueName: \"kubernetes.io/projected/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-kube-api-access-qgfc2\") pod \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " Feb 25 07:20:44 crc kubenswrapper[4978]: I0225 07:20:44.641957 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-catalog-content\") pod \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " Feb 25 07:20:44 crc kubenswrapper[4978]: I0225 07:20:44.642014 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-utilities\") pod \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\" (UID: \"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff\") " Feb 25 07:20:44 crc kubenswrapper[4978]: I0225 07:20:44.643137 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-utilities" (OuterVolumeSpecName: "utilities") pod "b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" (UID: "b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:20:44 crc kubenswrapper[4978]: I0225 07:20:44.651733 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-kube-api-access-qgfc2" (OuterVolumeSpecName: "kube-api-access-qgfc2") pod "b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" (UID: "b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff"). InnerVolumeSpecName "kube-api-access-qgfc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:20:44 crc kubenswrapper[4978]: I0225 07:20:44.743457 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:20:44 crc kubenswrapper[4978]: I0225 07:20:44.743514 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgfc2\" (UniqueName: \"kubernetes.io/projected/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-kube-api-access-qgfc2\") on node \"crc\" DevicePath \"\"" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.027589 4978 generic.go:334] "Generic (PLEG): container finished" podID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerID="f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7" exitCode=0 Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.027659 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwwr9" event={"ID":"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff","Type":"ContainerDied","Data":"f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7"} Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.027702 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nwwr9" event={"ID":"b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff","Type":"ContainerDied","Data":"9bac56112b320436e3739e7beca651fb385cc584ff177b7676c48f54e9b3bfca"} Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.027809 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nwwr9" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.027809 4978 scope.go:117] "RemoveContainer" containerID="f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.062246 4978 scope.go:117] "RemoveContainer" containerID="bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.097756 4978 scope.go:117] "RemoveContainer" containerID="831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.139341 4978 scope.go:117] "RemoveContainer" containerID="f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7" Feb 25 07:20:45 crc kubenswrapper[4978]: E0225 07:20:45.140152 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7\": container with ID starting with f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7 not found: ID does not exist" containerID="f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.140221 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7"} err="failed to get container status \"f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7\": rpc error: code = NotFound desc = could not find container \"f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7\": container with ID starting with f4e4d8982edd7f2184f0e07a334adeabba68c64f76eaf9cd868ec79cc91e8fb7 not found: ID does not exist" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.140261 4978 scope.go:117] "RemoveContainer" containerID="bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112" Feb 25 07:20:45 crc kubenswrapper[4978]: E0225 07:20:45.142967 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112\": container with ID starting with bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112 not found: ID does not exist" containerID="bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.143053 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112"} err="failed to get container status \"bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112\": rpc error: code = NotFound desc = could not find container \"bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112\": container with ID starting with bf6811c055327aff592150a56c7f681caf65dd17a5a87c0ce3e28e3448aa5112 not found: ID does not exist" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.143098 4978 scope.go:117] "RemoveContainer" containerID="831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3" Feb 25 07:20:45 crc kubenswrapper[4978]: E0225 07:20:45.143702 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3\": container with ID starting with 831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3 not found: ID does not exist" containerID="831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.143750 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3"} err="failed to get container status \"831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3\": rpc error: code = NotFound desc = could not find container \"831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3\": container with ID starting with 831539e314ae69cb13c5c4732dcb059bb42326dd7a08950848bc027912f23ab3 not found: ID does not exist" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.156767 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" (UID: "b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.252076 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.371218 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nwwr9"] Feb 25 07:20:45 crc kubenswrapper[4978]: I0225 07:20:45.377716 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nwwr9"] Feb 25 07:20:47 crc kubenswrapper[4978]: I0225 07:20:47.343553 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" path="/var/lib/kubelet/pods/b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff/volumes" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.236760 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7src5"] Feb 25 07:21:04 crc kubenswrapper[4978]: E0225 07:21:04.237633 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerName="extract-content" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.237649 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerName="extract-content" Feb 25 07:21:04 crc kubenswrapper[4978]: E0225 07:21:04.237682 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerName="extract-utilities" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.237691 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerName="extract-utilities" Feb 25 07:21:04 crc kubenswrapper[4978]: E0225 07:21:04.237701 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerName="registry-server" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.237710 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerName="registry-server" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.237867 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ccfa7c-0e07-46a1-94bc-5d0f8401cbff" containerName="registry-server" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.239131 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.256990 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7src5"] Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.398442 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gcsh\" (UniqueName: \"kubernetes.io/projected/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-kube-api-access-8gcsh\") pod \"redhat-operators-7src5\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.398781 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-catalog-content\") pod \"redhat-operators-7src5\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.398921 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-utilities\") pod \"redhat-operators-7src5\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.500351 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-catalog-content\") pod \"redhat-operators-7src5\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.500429 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-utilities\") pod \"redhat-operators-7src5\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.500493 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gcsh\" (UniqueName: \"kubernetes.io/projected/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-kube-api-access-8gcsh\") pod \"redhat-operators-7src5\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.501274 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-catalog-content\") pod \"redhat-operators-7src5\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.501606 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-utilities\") pod \"redhat-operators-7src5\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.519507 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gcsh\" (UniqueName: \"kubernetes.io/projected/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-kube-api-access-8gcsh\") pod \"redhat-operators-7src5\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:04 crc kubenswrapper[4978]: I0225 07:21:04.589777 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:05 crc kubenswrapper[4978]: I0225 07:21:05.102208 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7src5"] Feb 25 07:21:05 crc kubenswrapper[4978]: I0225 07:21:05.260636 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7src5" event={"ID":"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d","Type":"ContainerStarted","Data":"81eae5455acaa7919f98bb852d9483083c6354321033ac65cd6b5ae644ec5472"} Feb 25 07:21:06 crc kubenswrapper[4978]: I0225 07:21:06.283019 4978 generic.go:334] "Generic (PLEG): container finished" podID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerID="3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431" exitCode=0 Feb 25 07:21:06 crc kubenswrapper[4978]: I0225 07:21:06.283154 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7src5" event={"ID":"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d","Type":"ContainerDied","Data":"3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431"} Feb 25 07:21:06 crc kubenswrapper[4978]: I0225 07:21:06.286740 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 07:21:07 crc kubenswrapper[4978]: I0225 07:21:07.294980 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7src5" event={"ID":"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d","Type":"ContainerStarted","Data":"b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc"} Feb 25 07:21:08 crc kubenswrapper[4978]: I0225 07:21:08.307426 4978 generic.go:334] "Generic (PLEG): container finished" podID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerID="b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc" exitCode=0 Feb 25 07:21:08 crc kubenswrapper[4978]: I0225 07:21:08.307520 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7src5" event={"ID":"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d","Type":"ContainerDied","Data":"b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc"} Feb 25 07:21:09 crc kubenswrapper[4978]: I0225 07:21:09.316620 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7src5" event={"ID":"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d","Type":"ContainerStarted","Data":"1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e"} Feb 25 07:21:09 crc kubenswrapper[4978]: I0225 07:21:09.337998 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7src5" podStartSLOduration=2.901036828 podStartE2EDuration="5.337980036s" podCreationTimestamp="2026-02-25 07:21:04 +0000 UTC" firstStartedPulling="2026-02-25 07:21:06.286299498 +0000 UTC m=+2159.725555997" lastFinishedPulling="2026-02-25 07:21:08.723242746 +0000 UTC m=+2162.162499205" observedRunningTime="2026-02-25 07:21:09.337449189 +0000 UTC m=+2162.776705678" watchObservedRunningTime="2026-02-25 07:21:09.337980036 +0000 UTC m=+2162.777236495" Feb 25 07:21:14 crc kubenswrapper[4978]: I0225 07:21:14.590726 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:14 crc kubenswrapper[4978]: I0225 07:21:14.591360 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:15 crc kubenswrapper[4978]: I0225 07:21:15.669954 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7src5" podUID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerName="registry-server" probeResult="failure" output=< Feb 25 07:21:15 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 07:21:15 crc kubenswrapper[4978]: > Feb 25 07:21:24 crc kubenswrapper[4978]: I0225 07:21:24.650442 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:24 crc kubenswrapper[4978]: I0225 07:21:24.711043 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:24 crc kubenswrapper[4978]: I0225 07:21:24.902729 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7src5"] Feb 25 07:21:26 crc kubenswrapper[4978]: I0225 07:21:26.466555 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7src5" podUID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerName="registry-server" containerID="cri-o://1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e" gracePeriod=2 Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.009888 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.173046 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-catalog-content\") pod \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.173144 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gcsh\" (UniqueName: \"kubernetes.io/projected/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-kube-api-access-8gcsh\") pod \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.173204 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-utilities\") pod \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\" (UID: \"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d\") " Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.175131 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-utilities" (OuterVolumeSpecName: "utilities") pod "aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" (UID: "aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.184696 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-kube-api-access-8gcsh" (OuterVolumeSpecName: "kube-api-access-8gcsh") pod "aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" (UID: "aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d"). InnerVolumeSpecName "kube-api-access-8gcsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.274910 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.274965 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gcsh\" (UniqueName: \"kubernetes.io/projected/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-kube-api-access-8gcsh\") on node \"crc\" DevicePath \"\"" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.350795 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" (UID: "aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.377034 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.480346 4978 generic.go:334] "Generic (PLEG): container finished" podID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerID="1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e" exitCode=0 Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.480460 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7src5" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.480496 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7src5" event={"ID":"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d","Type":"ContainerDied","Data":"1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e"} Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.480666 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7src5" event={"ID":"aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d","Type":"ContainerDied","Data":"81eae5455acaa7919f98bb852d9483083c6354321033ac65cd6b5ae644ec5472"} Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.480704 4978 scope.go:117] "RemoveContainer" containerID="1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.514879 4978 scope.go:117] "RemoveContainer" containerID="b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.534430 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7src5"] Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.544648 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7src5"] Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.561519 4978 scope.go:117] "RemoveContainer" containerID="3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.596123 4978 scope.go:117] "RemoveContainer" containerID="1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e" Feb 25 07:21:27 crc kubenswrapper[4978]: E0225 07:21:27.596808 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e\": container with ID starting with 1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e not found: ID does not exist" containerID="1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.596897 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e"} err="failed to get container status \"1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e\": rpc error: code = NotFound desc = could not find container \"1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e\": container with ID starting with 1e46334da09044c72de7ec8e93267662e58956d51b29f780338bceec9fe35e1e not found: ID does not exist" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.596972 4978 scope.go:117] "RemoveContainer" containerID="b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc" Feb 25 07:21:27 crc kubenswrapper[4978]: E0225 07:21:27.597575 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc\": container with ID starting with b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc not found: ID does not exist" containerID="b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.597623 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc"} err="failed to get container status \"b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc\": rpc error: code = NotFound desc = could not find container \"b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc\": container with ID starting with b7c030f9c8b3b0c1ad44a9f026dce1bf434617005b732bcfdd7754a3b6ec7efc not found: ID does not exist" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.597649 4978 scope.go:117] "RemoveContainer" containerID="3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431" Feb 25 07:21:27 crc kubenswrapper[4978]: E0225 07:21:27.598205 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431\": container with ID starting with 3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431 not found: ID does not exist" containerID="3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431" Feb 25 07:21:27 crc kubenswrapper[4978]: I0225 07:21:27.598257 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431"} err="failed to get container status \"3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431\": rpc error: code = NotFound desc = could not find container \"3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431\": container with ID starting with 3a879a777898b923fd9dd07a33b7baf8310191abd82952646d1b6dbd48af5431 not found: ID does not exist" Feb 25 07:21:29 crc kubenswrapper[4978]: I0225 07:21:29.343001 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" path="/var/lib/kubelet/pods/aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d/volumes" Feb 25 07:21:46 crc kubenswrapper[4978]: I0225 07:21:46.540478 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:21:46 crc kubenswrapper[4978]: I0225 07:21:46.541210 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.169153 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533402-vl77q"] Feb 25 07:22:00 crc kubenswrapper[4978]: E0225 07:22:00.170313 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerName="extract-utilities" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.170342 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerName="extract-utilities" Feb 25 07:22:00 crc kubenswrapper[4978]: E0225 07:22:00.170393 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerName="registry-server" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.170407 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerName="registry-server" Feb 25 07:22:00 crc kubenswrapper[4978]: E0225 07:22:00.170452 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerName="extract-content" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.170465 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerName="extract-content" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.170705 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaedc1b4-a6f2-4944-a70b-8c506ae2cf0d" containerName="registry-server" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.171612 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533402-vl77q" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.174702 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.174918 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.175006 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.179750 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533402-vl77q"] Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.335775 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hctp7\" (UniqueName: \"kubernetes.io/projected/e58c0d5d-8529-4d34-9621-9f367f9d6d8d-kube-api-access-hctp7\") pod \"auto-csr-approver-29533402-vl77q\" (UID: \"e58c0d5d-8529-4d34-9621-9f367f9d6d8d\") " pod="openshift-infra/auto-csr-approver-29533402-vl77q" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.437336 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hctp7\" (UniqueName: \"kubernetes.io/projected/e58c0d5d-8529-4d34-9621-9f367f9d6d8d-kube-api-access-hctp7\") pod \"auto-csr-approver-29533402-vl77q\" (UID: \"e58c0d5d-8529-4d34-9621-9f367f9d6d8d\") " pod="openshift-infra/auto-csr-approver-29533402-vl77q" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.464946 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hctp7\" (UniqueName: \"kubernetes.io/projected/e58c0d5d-8529-4d34-9621-9f367f9d6d8d-kube-api-access-hctp7\") pod \"auto-csr-approver-29533402-vl77q\" (UID: \"e58c0d5d-8529-4d34-9621-9f367f9d6d8d\") " pod="openshift-infra/auto-csr-approver-29533402-vl77q" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.515402 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533402-vl77q" Feb 25 07:22:00 crc kubenswrapper[4978]: I0225 07:22:00.995103 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533402-vl77q"] Feb 25 07:22:01 crc kubenswrapper[4978]: I0225 07:22:01.803130 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533402-vl77q" event={"ID":"e58c0d5d-8529-4d34-9621-9f367f9d6d8d","Type":"ContainerStarted","Data":"b8073df9c1c398383b4ce95fab5d78f7e313b6f423409ffb997762292bac30a1"} Feb 25 07:22:02 crc kubenswrapper[4978]: I0225 07:22:02.812426 4978 generic.go:334] "Generic (PLEG): container finished" podID="e58c0d5d-8529-4d34-9621-9f367f9d6d8d" containerID="f32df57c3ddee4cbc7724e109664feeeec9dd24e1a73261957545b9eefb460b0" exitCode=0 Feb 25 07:22:02 crc kubenswrapper[4978]: I0225 07:22:02.812534 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533402-vl77q" event={"ID":"e58c0d5d-8529-4d34-9621-9f367f9d6d8d","Type":"ContainerDied","Data":"f32df57c3ddee4cbc7724e109664feeeec9dd24e1a73261957545b9eefb460b0"} Feb 25 07:22:04 crc kubenswrapper[4978]: I0225 07:22:04.229038 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533402-vl77q" Feb 25 07:22:04 crc kubenswrapper[4978]: I0225 07:22:04.404687 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hctp7\" (UniqueName: \"kubernetes.io/projected/e58c0d5d-8529-4d34-9621-9f367f9d6d8d-kube-api-access-hctp7\") pod \"e58c0d5d-8529-4d34-9621-9f367f9d6d8d\" (UID: \"e58c0d5d-8529-4d34-9621-9f367f9d6d8d\") " Feb 25 07:22:04 crc kubenswrapper[4978]: I0225 07:22:04.413403 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e58c0d5d-8529-4d34-9621-9f367f9d6d8d-kube-api-access-hctp7" (OuterVolumeSpecName: "kube-api-access-hctp7") pod "e58c0d5d-8529-4d34-9621-9f367f9d6d8d" (UID: "e58c0d5d-8529-4d34-9621-9f367f9d6d8d"). InnerVolumeSpecName "kube-api-access-hctp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:22:04 crc kubenswrapper[4978]: I0225 07:22:04.507080 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hctp7\" (UniqueName: \"kubernetes.io/projected/e58c0d5d-8529-4d34-9621-9f367f9d6d8d-kube-api-access-hctp7\") on node \"crc\" DevicePath \"\"" Feb 25 07:22:04 crc kubenswrapper[4978]: I0225 07:22:04.833464 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533402-vl77q" event={"ID":"e58c0d5d-8529-4d34-9621-9f367f9d6d8d","Type":"ContainerDied","Data":"b8073df9c1c398383b4ce95fab5d78f7e313b6f423409ffb997762292bac30a1"} Feb 25 07:22:04 crc kubenswrapper[4978]: I0225 07:22:04.833522 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8073df9c1c398383b4ce95fab5d78f7e313b6f423409ffb997762292bac30a1" Feb 25 07:22:04 crc kubenswrapper[4978]: I0225 07:22:04.833560 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533402-vl77q" Feb 25 07:22:05 crc kubenswrapper[4978]: I0225 07:22:05.334048 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533396-bcqvb"] Feb 25 07:22:05 crc kubenswrapper[4978]: I0225 07:22:05.357012 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533396-bcqvb"] Feb 25 07:22:07 crc kubenswrapper[4978]: I0225 07:22:07.340588 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c1916c8-8177-427f-86fb-2e70a3217f3b" path="/var/lib/kubelet/pods/7c1916c8-8177-427f-86fb-2e70a3217f3b/volumes" Feb 25 07:22:16 crc kubenswrapper[4978]: I0225 07:22:16.541009 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:22:16 crc kubenswrapper[4978]: I0225 07:22:16.541816 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:22:21 crc kubenswrapper[4978]: I0225 07:22:21.431364 4978 scope.go:117] "RemoveContainer" containerID="e8df7b3ef6c0b2c8d219d3cb017485fe2e60d345d87e97a37631514fe68a6007" Feb 25 07:22:46 crc kubenswrapper[4978]: I0225 07:22:46.540539 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:22:46 crc kubenswrapper[4978]: I0225 07:22:46.541101 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:22:46 crc kubenswrapper[4978]: I0225 07:22:46.541157 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:22:46 crc kubenswrapper[4978]: I0225 07:22:46.541830 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:22:46 crc kubenswrapper[4978]: I0225 07:22:46.541911 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" gracePeriod=600 Feb 25 07:22:46 crc kubenswrapper[4978]: E0225 07:22:46.667962 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.021898 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gpwz7"] Feb 25 07:22:47 crc kubenswrapper[4978]: E0225 07:22:47.022346 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e58c0d5d-8529-4d34-9621-9f367f9d6d8d" containerName="oc" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.022377 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e58c0d5d-8529-4d34-9621-9f367f9d6d8d" containerName="oc" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.022630 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e58c0d5d-8529-4d34-9621-9f367f9d6d8d" containerName="oc" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.024526 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.038194 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gpwz7"] Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.166856 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-utilities\") pod \"community-operators-gpwz7\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.167339 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-catalog-content\") pod \"community-operators-gpwz7\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.167428 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnklx\" (UniqueName: \"kubernetes.io/projected/830fcbdd-7741-402e-9e6a-4d942969bd59-kube-api-access-xnklx\") pod \"community-operators-gpwz7\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.224810 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" exitCode=0 Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.224857 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c"} Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.224924 4978 scope.go:117] "RemoveContainer" containerID="4d066de9c5cf5b617ddc67630504439caf4b63e08b3c93deb65061b29eb89656" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.225745 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:22:47 crc kubenswrapper[4978]: E0225 07:22:47.226120 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.268923 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-catalog-content\") pod \"community-operators-gpwz7\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.269002 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnklx\" (UniqueName: \"kubernetes.io/projected/830fcbdd-7741-402e-9e6a-4d942969bd59-kube-api-access-xnklx\") pod \"community-operators-gpwz7\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.269606 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-catalog-content\") pod \"community-operators-gpwz7\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.269782 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-utilities\") pod \"community-operators-gpwz7\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.270277 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-utilities\") pod \"community-operators-gpwz7\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.297618 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnklx\" (UniqueName: \"kubernetes.io/projected/830fcbdd-7741-402e-9e6a-4d942969bd59-kube-api-access-xnklx\") pod \"community-operators-gpwz7\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.355145 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:47 crc kubenswrapper[4978]: I0225 07:22:47.789817 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gpwz7"] Feb 25 07:22:48 crc kubenswrapper[4978]: I0225 07:22:48.243128 4978 generic.go:334] "Generic (PLEG): container finished" podID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerID="970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf" exitCode=0 Feb 25 07:22:48 crc kubenswrapper[4978]: I0225 07:22:48.243373 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpwz7" event={"ID":"830fcbdd-7741-402e-9e6a-4d942969bd59","Type":"ContainerDied","Data":"970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf"} Feb 25 07:22:48 crc kubenswrapper[4978]: I0225 07:22:48.243461 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpwz7" event={"ID":"830fcbdd-7741-402e-9e6a-4d942969bd59","Type":"ContainerStarted","Data":"f7eb670a8866ae24f70ad481020b7b407c76a47380364815f1c15f2d412c1114"} Feb 25 07:22:50 crc kubenswrapper[4978]: I0225 07:22:50.272230 4978 generic.go:334] "Generic (PLEG): container finished" podID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerID="6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736" exitCode=0 Feb 25 07:22:50 crc kubenswrapper[4978]: I0225 07:22:50.272362 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpwz7" event={"ID":"830fcbdd-7741-402e-9e6a-4d942969bd59","Type":"ContainerDied","Data":"6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736"} Feb 25 07:22:51 crc kubenswrapper[4978]: I0225 07:22:51.286477 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpwz7" event={"ID":"830fcbdd-7741-402e-9e6a-4d942969bd59","Type":"ContainerStarted","Data":"1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a"} Feb 25 07:22:51 crc kubenswrapper[4978]: I0225 07:22:51.309625 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gpwz7" podStartSLOduration=2.871142657 podStartE2EDuration="5.309591959s" podCreationTimestamp="2026-02-25 07:22:46 +0000 UTC" firstStartedPulling="2026-02-25 07:22:48.247343019 +0000 UTC m=+2261.686599468" lastFinishedPulling="2026-02-25 07:22:50.685792271 +0000 UTC m=+2264.125048770" observedRunningTime="2026-02-25 07:22:51.308319732 +0000 UTC m=+2264.747576301" watchObservedRunningTime="2026-02-25 07:22:51.309591959 +0000 UTC m=+2264.748848478" Feb 25 07:22:57 crc kubenswrapper[4978]: I0225 07:22:57.355340 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:57 crc kubenswrapper[4978]: I0225 07:22:57.356476 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:57 crc kubenswrapper[4978]: I0225 07:22:57.437667 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:58 crc kubenswrapper[4978]: I0225 07:22:58.438626 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:22:58 crc kubenswrapper[4978]: I0225 07:22:58.513645 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gpwz7"] Feb 25 07:22:59 crc kubenswrapper[4978]: I0225 07:22:59.328934 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:22:59 crc kubenswrapper[4978]: E0225 07:22:59.329422 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.381546 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gpwz7" podUID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerName="registry-server" containerID="cri-o://1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a" gracePeriod=2 Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.844867 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.894182 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-catalog-content\") pod \"830fcbdd-7741-402e-9e6a-4d942969bd59\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.894302 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnklx\" (UniqueName: \"kubernetes.io/projected/830fcbdd-7741-402e-9e6a-4d942969bd59-kube-api-access-xnklx\") pod \"830fcbdd-7741-402e-9e6a-4d942969bd59\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.894348 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-utilities\") pod \"830fcbdd-7741-402e-9e6a-4d942969bd59\" (UID: \"830fcbdd-7741-402e-9e6a-4d942969bd59\") " Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.895390 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-utilities" (OuterVolumeSpecName: "utilities") pod "830fcbdd-7741-402e-9e6a-4d942969bd59" (UID: "830fcbdd-7741-402e-9e6a-4d942969bd59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.903079 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/830fcbdd-7741-402e-9e6a-4d942969bd59-kube-api-access-xnklx" (OuterVolumeSpecName: "kube-api-access-xnklx") pod "830fcbdd-7741-402e-9e6a-4d942969bd59" (UID: "830fcbdd-7741-402e-9e6a-4d942969bd59"). InnerVolumeSpecName "kube-api-access-xnklx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.950128 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "830fcbdd-7741-402e-9e6a-4d942969bd59" (UID: "830fcbdd-7741-402e-9e6a-4d942969bd59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.995418 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.995472 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnklx\" (UniqueName: \"kubernetes.io/projected/830fcbdd-7741-402e-9e6a-4d942969bd59-kube-api-access-xnklx\") on node \"crc\" DevicePath \"\"" Feb 25 07:23:00 crc kubenswrapper[4978]: I0225 07:23:00.995488 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/830fcbdd-7741-402e-9e6a-4d942969bd59-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.393234 4978 generic.go:334] "Generic (PLEG): container finished" podID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerID="1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a" exitCode=0 Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.393293 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpwz7" event={"ID":"830fcbdd-7741-402e-9e6a-4d942969bd59","Type":"ContainerDied","Data":"1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a"} Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.393360 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpwz7" Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.393665 4978 scope.go:117] "RemoveContainer" containerID="1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a" Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.393648 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpwz7" event={"ID":"830fcbdd-7741-402e-9e6a-4d942969bd59","Type":"ContainerDied","Data":"f7eb670a8866ae24f70ad481020b7b407c76a47380364815f1c15f2d412c1114"} Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.425226 4978 scope.go:117] "RemoveContainer" containerID="6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736" Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.434054 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gpwz7"] Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.449695 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gpwz7"] Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.465337 4978 scope.go:117] "RemoveContainer" containerID="970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf" Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.490264 4978 scope.go:117] "RemoveContainer" containerID="1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a" Feb 25 07:23:01 crc kubenswrapper[4978]: E0225 07:23:01.490967 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a\": container with ID starting with 1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a not found: ID does not exist" containerID="1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a" Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.491033 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a"} err="failed to get container status \"1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a\": rpc error: code = NotFound desc = could not find container \"1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a\": container with ID starting with 1093d0395a9eba9edac4ad561e6275a3e6c0ec53695d665ee878f1a509b7417a not found: ID does not exist" Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.491080 4978 scope.go:117] "RemoveContainer" containerID="6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736" Feb 25 07:23:01 crc kubenswrapper[4978]: E0225 07:23:01.492445 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736\": container with ID starting with 6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736 not found: ID does not exist" containerID="6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736" Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.492508 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736"} err="failed to get container status \"6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736\": rpc error: code = NotFound desc = could not find container \"6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736\": container with ID starting with 6acac77f960ae45ac3f0a7a3e931c67929737ba7ddf99dfae4308aee80030736 not found: ID does not exist" Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.492553 4978 scope.go:117] "RemoveContainer" containerID="970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf" Feb 25 07:23:01 crc kubenswrapper[4978]: E0225 07:23:01.492944 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf\": container with ID starting with 970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf not found: ID does not exist" containerID="970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf" Feb 25 07:23:01 crc kubenswrapper[4978]: I0225 07:23:01.493005 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf"} err="failed to get container status \"970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf\": rpc error: code = NotFound desc = could not find container \"970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf\": container with ID starting with 970a24c7e60b3c5352e22781fc26424d37249e5de6fa81abfaee2af3a2d8c8cf not found: ID does not exist" Feb 25 07:23:03 crc kubenswrapper[4978]: I0225 07:23:03.342823 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="830fcbdd-7741-402e-9e6a-4d942969bd59" path="/var/lib/kubelet/pods/830fcbdd-7741-402e-9e6a-4d942969bd59/volumes" Feb 25 07:23:13 crc kubenswrapper[4978]: I0225 07:23:13.328144 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:23:13 crc kubenswrapper[4978]: E0225 07:23:13.329518 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:23:27 crc kubenswrapper[4978]: I0225 07:23:27.336244 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:23:27 crc kubenswrapper[4978]: E0225 07:23:27.337456 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.581616 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ngmqc"] Feb 25 07:23:37 crc kubenswrapper[4978]: E0225 07:23:37.582645 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerName="extract-content" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.582679 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerName="extract-content" Feb 25 07:23:37 crc kubenswrapper[4978]: E0225 07:23:37.582704 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerName="registry-server" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.582720 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerName="registry-server" Feb 25 07:23:37 crc kubenswrapper[4978]: E0225 07:23:37.582760 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerName="extract-utilities" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.582779 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerName="extract-utilities" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.583144 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="830fcbdd-7741-402e-9e6a-4d942969bd59" containerName="registry-server" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.585449 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.608759 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngmqc"] Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.720271 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-catalog-content\") pod \"redhat-marketplace-ngmqc\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.720335 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-utilities\") pod \"redhat-marketplace-ngmqc\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.720392 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-kube-api-access-dc5bv\") pod \"redhat-marketplace-ngmqc\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.822194 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-kube-api-access-dc5bv\") pod \"redhat-marketplace-ngmqc\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.822365 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-catalog-content\") pod \"redhat-marketplace-ngmqc\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.822419 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-utilities\") pod \"redhat-marketplace-ngmqc\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.822907 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-catalog-content\") pod \"redhat-marketplace-ngmqc\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.822945 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-utilities\") pod \"redhat-marketplace-ngmqc\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.846499 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-kube-api-access-dc5bv\") pod \"redhat-marketplace-ngmqc\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:37 crc kubenswrapper[4978]: I0225 07:23:37.937699 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:38 crc kubenswrapper[4978]: I0225 07:23:38.411338 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngmqc"] Feb 25 07:23:38 crc kubenswrapper[4978]: I0225 07:23:38.745495 4978 generic.go:334] "Generic (PLEG): container finished" podID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerID="fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1" exitCode=0 Feb 25 07:23:38 crc kubenswrapper[4978]: I0225 07:23:38.745573 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngmqc" event={"ID":"c04edcd6-db37-4a6a-ab0e-acb0b4a97316","Type":"ContainerDied","Data":"fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1"} Feb 25 07:23:38 crc kubenswrapper[4978]: I0225 07:23:38.745803 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngmqc" event={"ID":"c04edcd6-db37-4a6a-ab0e-acb0b4a97316","Type":"ContainerStarted","Data":"d28a58722d9cbf635230f87c49d446af46738404d4a3f5ad9c46e811d4e6386f"} Feb 25 07:23:39 crc kubenswrapper[4978]: I0225 07:23:39.755147 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngmqc" event={"ID":"c04edcd6-db37-4a6a-ab0e-acb0b4a97316","Type":"ContainerStarted","Data":"a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d"} Feb 25 07:23:40 crc kubenswrapper[4978]: I0225 07:23:40.327638 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:23:40 crc kubenswrapper[4978]: E0225 07:23:40.328248 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:23:40 crc kubenswrapper[4978]: I0225 07:23:40.764712 4978 generic.go:334] "Generic (PLEG): container finished" podID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerID="a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d" exitCode=0 Feb 25 07:23:40 crc kubenswrapper[4978]: I0225 07:23:40.764761 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngmqc" event={"ID":"c04edcd6-db37-4a6a-ab0e-acb0b4a97316","Type":"ContainerDied","Data":"a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d"} Feb 25 07:23:41 crc kubenswrapper[4978]: I0225 07:23:41.777098 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngmqc" event={"ID":"c04edcd6-db37-4a6a-ab0e-acb0b4a97316","Type":"ContainerStarted","Data":"af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434"} Feb 25 07:23:41 crc kubenswrapper[4978]: I0225 07:23:41.807035 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ngmqc" podStartSLOduration=2.40986308 podStartE2EDuration="4.807013906s" podCreationTimestamp="2026-02-25 07:23:37 +0000 UTC" firstStartedPulling="2026-02-25 07:23:38.747427312 +0000 UTC m=+2312.186683811" lastFinishedPulling="2026-02-25 07:23:41.144578168 +0000 UTC m=+2314.583834637" observedRunningTime="2026-02-25 07:23:41.80368575 +0000 UTC m=+2315.242942229" watchObservedRunningTime="2026-02-25 07:23:41.807013906 +0000 UTC m=+2315.246270365" Feb 25 07:23:47 crc kubenswrapper[4978]: I0225 07:23:47.938439 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:47 crc kubenswrapper[4978]: I0225 07:23:47.939351 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:48 crc kubenswrapper[4978]: I0225 07:23:48.018588 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:48 crc kubenswrapper[4978]: I0225 07:23:48.879515 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:51 crc kubenswrapper[4978]: I0225 07:23:51.561875 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngmqc"] Feb 25 07:23:51 crc kubenswrapper[4978]: I0225 07:23:51.859759 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ngmqc" podUID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerName="registry-server" containerID="cri-o://af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434" gracePeriod=2 Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.327929 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.457511 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-kube-api-access-dc5bv\") pod \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.457578 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-catalog-content\") pod \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.457656 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-utilities\") pod \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\" (UID: \"c04edcd6-db37-4a6a-ab0e-acb0b4a97316\") " Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.459772 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-utilities" (OuterVolumeSpecName: "utilities") pod "c04edcd6-db37-4a6a-ab0e-acb0b4a97316" (UID: "c04edcd6-db37-4a6a-ab0e-acb0b4a97316"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.467554 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-kube-api-access-dc5bv" (OuterVolumeSpecName: "kube-api-access-dc5bv") pod "c04edcd6-db37-4a6a-ab0e-acb0b4a97316" (UID: "c04edcd6-db37-4a6a-ab0e-acb0b4a97316"). InnerVolumeSpecName "kube-api-access-dc5bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.480734 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c04edcd6-db37-4a6a-ab0e-acb0b4a97316" (UID: "c04edcd6-db37-4a6a-ab0e-acb0b4a97316"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.559150 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.559185 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-kube-api-access-dc5bv\") on node \"crc\" DevicePath \"\"" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.559196 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c04edcd6-db37-4a6a-ab0e-acb0b4a97316-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.871521 4978 generic.go:334] "Generic (PLEG): container finished" podID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerID="af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434" exitCode=0 Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.871578 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngmqc" event={"ID":"c04edcd6-db37-4a6a-ab0e-acb0b4a97316","Type":"ContainerDied","Data":"af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434"} Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.871683 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngmqc" event={"ID":"c04edcd6-db37-4a6a-ab0e-acb0b4a97316","Type":"ContainerDied","Data":"d28a58722d9cbf635230f87c49d446af46738404d4a3f5ad9c46e811d4e6386f"} Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.871716 4978 scope.go:117] "RemoveContainer" containerID="af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.871717 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngmqc" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.898271 4978 scope.go:117] "RemoveContainer" containerID="a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.924745 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngmqc"] Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.929819 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngmqc"] Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.944008 4978 scope.go:117] "RemoveContainer" containerID="fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.969682 4978 scope.go:117] "RemoveContainer" containerID="af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434" Feb 25 07:23:52 crc kubenswrapper[4978]: E0225 07:23:52.970190 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434\": container with ID starting with af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434 not found: ID does not exist" containerID="af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.970248 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434"} err="failed to get container status \"af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434\": rpc error: code = NotFound desc = could not find container \"af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434\": container with ID starting with af4dd5fd7385397eb07f43b012a46185dd409f8665b50bb08e4e8cd654f05434 not found: ID does not exist" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.970275 4978 scope.go:117] "RemoveContainer" containerID="a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d" Feb 25 07:23:52 crc kubenswrapper[4978]: E0225 07:23:52.970570 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d\": container with ID starting with a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d not found: ID does not exist" containerID="a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.970598 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d"} err="failed to get container status \"a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d\": rpc error: code = NotFound desc = could not find container \"a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d\": container with ID starting with a768d4cdff5f529f04a13ebc3e0edb58b9ca34b6a533dfb01d73eb1b3c06609d not found: ID does not exist" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.970645 4978 scope.go:117] "RemoveContainer" containerID="fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1" Feb 25 07:23:52 crc kubenswrapper[4978]: E0225 07:23:52.970949 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1\": container with ID starting with fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1 not found: ID does not exist" containerID="fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1" Feb 25 07:23:52 crc kubenswrapper[4978]: I0225 07:23:52.970984 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1"} err="failed to get container status \"fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1\": rpc error: code = NotFound desc = could not find container \"fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1\": container with ID starting with fbeb48d9384b110cc60344b5d3df2c9ad1c29bc99959c90145ea332426dac3d1 not found: ID does not exist" Feb 25 07:23:53 crc kubenswrapper[4978]: I0225 07:23:53.336194 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" path="/var/lib/kubelet/pods/c04edcd6-db37-4a6a-ab0e-acb0b4a97316/volumes" Feb 25 07:23:54 crc kubenswrapper[4978]: I0225 07:23:54.327909 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:23:54 crc kubenswrapper[4978]: E0225 07:23:54.328183 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.148694 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533404-p6ms8"] Feb 25 07:24:00 crc kubenswrapper[4978]: E0225 07:24:00.149514 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerName="extract-content" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.149529 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerName="extract-content" Feb 25 07:24:00 crc kubenswrapper[4978]: E0225 07:24:00.149544 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerName="extract-utilities" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.149550 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerName="extract-utilities" Feb 25 07:24:00 crc kubenswrapper[4978]: E0225 07:24:00.149572 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerName="registry-server" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.149578 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerName="registry-server" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.149731 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c04edcd6-db37-4a6a-ab0e-acb0b4a97316" containerName="registry-server" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.150307 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533404-p6ms8" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.154277 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.156055 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.156059 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.174883 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533404-p6ms8"] Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.280157 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jh5v\" (UniqueName: \"kubernetes.io/projected/e641956c-57e8-4eed-9cb0-7f86c375dba7-kube-api-access-7jh5v\") pod \"auto-csr-approver-29533404-p6ms8\" (UID: \"e641956c-57e8-4eed-9cb0-7f86c375dba7\") " pod="openshift-infra/auto-csr-approver-29533404-p6ms8" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.382472 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jh5v\" (UniqueName: \"kubernetes.io/projected/e641956c-57e8-4eed-9cb0-7f86c375dba7-kube-api-access-7jh5v\") pod \"auto-csr-approver-29533404-p6ms8\" (UID: \"e641956c-57e8-4eed-9cb0-7f86c375dba7\") " pod="openshift-infra/auto-csr-approver-29533404-p6ms8" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.414860 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jh5v\" (UniqueName: \"kubernetes.io/projected/e641956c-57e8-4eed-9cb0-7f86c375dba7-kube-api-access-7jh5v\") pod \"auto-csr-approver-29533404-p6ms8\" (UID: \"e641956c-57e8-4eed-9cb0-7f86c375dba7\") " pod="openshift-infra/auto-csr-approver-29533404-p6ms8" Feb 25 07:24:00 crc kubenswrapper[4978]: I0225 07:24:00.482076 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533404-p6ms8" Feb 25 07:24:01 crc kubenswrapper[4978]: I0225 07:24:01.006458 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533404-p6ms8"] Feb 25 07:24:01 crc kubenswrapper[4978]: I0225 07:24:01.987123 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533404-p6ms8" event={"ID":"e641956c-57e8-4eed-9cb0-7f86c375dba7","Type":"ContainerStarted","Data":"999d1663da5e05623d6021c1828206320856b9cf35a227dde89c962c7f32cd7b"} Feb 25 07:24:03 crc kubenswrapper[4978]: I0225 07:24:03.003624 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533404-p6ms8" event={"ID":"e641956c-57e8-4eed-9cb0-7f86c375dba7","Type":"ContainerStarted","Data":"d86c99f4652655abe18d9b1aff127786d70e31b7a01a59b241f2a0cb90875d68"} Feb 25 07:24:04 crc kubenswrapper[4978]: I0225 07:24:04.014801 4978 generic.go:334] "Generic (PLEG): container finished" podID="e641956c-57e8-4eed-9cb0-7f86c375dba7" containerID="d86c99f4652655abe18d9b1aff127786d70e31b7a01a59b241f2a0cb90875d68" exitCode=0 Feb 25 07:24:04 crc kubenswrapper[4978]: I0225 07:24:04.014872 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533404-p6ms8" event={"ID":"e641956c-57e8-4eed-9cb0-7f86c375dba7","Type":"ContainerDied","Data":"d86c99f4652655abe18d9b1aff127786d70e31b7a01a59b241f2a0cb90875d68"} Feb 25 07:24:04 crc kubenswrapper[4978]: I0225 07:24:04.364420 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533404-p6ms8" Feb 25 07:24:04 crc kubenswrapper[4978]: I0225 07:24:04.446620 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jh5v\" (UniqueName: \"kubernetes.io/projected/e641956c-57e8-4eed-9cb0-7f86c375dba7-kube-api-access-7jh5v\") pod \"e641956c-57e8-4eed-9cb0-7f86c375dba7\" (UID: \"e641956c-57e8-4eed-9cb0-7f86c375dba7\") " Feb 25 07:24:04 crc kubenswrapper[4978]: I0225 07:24:04.454126 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e641956c-57e8-4eed-9cb0-7f86c375dba7-kube-api-access-7jh5v" (OuterVolumeSpecName: "kube-api-access-7jh5v") pod "e641956c-57e8-4eed-9cb0-7f86c375dba7" (UID: "e641956c-57e8-4eed-9cb0-7f86c375dba7"). InnerVolumeSpecName "kube-api-access-7jh5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:24:04 crc kubenswrapper[4978]: I0225 07:24:04.548588 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jh5v\" (UniqueName: \"kubernetes.io/projected/e641956c-57e8-4eed-9cb0-7f86c375dba7-kube-api-access-7jh5v\") on node \"crc\" DevicePath \"\"" Feb 25 07:24:05 crc kubenswrapper[4978]: I0225 07:24:05.024069 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533404-p6ms8" event={"ID":"e641956c-57e8-4eed-9cb0-7f86c375dba7","Type":"ContainerDied","Data":"999d1663da5e05623d6021c1828206320856b9cf35a227dde89c962c7f32cd7b"} Feb 25 07:24:05 crc kubenswrapper[4978]: I0225 07:24:05.024111 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="999d1663da5e05623d6021c1828206320856b9cf35a227dde89c962c7f32cd7b" Feb 25 07:24:05 crc kubenswrapper[4978]: I0225 07:24:05.024864 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533404-p6ms8" Feb 25 07:24:05 crc kubenswrapper[4978]: I0225 07:24:05.446855 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533398-85vk8"] Feb 25 07:24:05 crc kubenswrapper[4978]: I0225 07:24:05.457957 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533398-85vk8"] Feb 25 07:24:07 crc kubenswrapper[4978]: I0225 07:24:07.338810 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="646a0605-333d-4033-ade8-5309b201ecf3" path="/var/lib/kubelet/pods/646a0605-333d-4033-ade8-5309b201ecf3/volumes" Feb 25 07:24:08 crc kubenswrapper[4978]: I0225 07:24:08.327469 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:24:08 crc kubenswrapper[4978]: E0225 07:24:08.327884 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:24:21 crc kubenswrapper[4978]: I0225 07:24:21.578642 4978 scope.go:117] "RemoveContainer" containerID="339fa3d72c25703e1c07c7863cccf0f22bc75972f7ad4dff7c9f150d28757970" Feb 25 07:24:23 crc kubenswrapper[4978]: I0225 07:24:23.329011 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:24:23 crc kubenswrapper[4978]: E0225 07:24:23.329812 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:24:38 crc kubenswrapper[4978]: I0225 07:24:38.327628 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:24:38 crc kubenswrapper[4978]: E0225 07:24:38.328463 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:24:52 crc kubenswrapper[4978]: I0225 07:24:52.328253 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:24:52 crc kubenswrapper[4978]: E0225 07:24:52.330064 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:25:05 crc kubenswrapper[4978]: I0225 07:25:05.327971 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:25:05 crc kubenswrapper[4978]: E0225 07:25:05.329282 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:25:19 crc kubenswrapper[4978]: I0225 07:25:19.329475 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:25:19 crc kubenswrapper[4978]: E0225 07:25:19.330582 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:25:30 crc kubenswrapper[4978]: I0225 07:25:30.364326 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:25:30 crc kubenswrapper[4978]: E0225 07:25:30.365510 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:25:43 crc kubenswrapper[4978]: I0225 07:25:43.329098 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:25:43 crc kubenswrapper[4978]: E0225 07:25:43.330355 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:25:54 crc kubenswrapper[4978]: I0225 07:25:54.327778 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:25:54 crc kubenswrapper[4978]: E0225 07:25:54.328940 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.158417 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533406-hpwgx"] Feb 25 07:26:00 crc kubenswrapper[4978]: E0225 07:26:00.159497 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e641956c-57e8-4eed-9cb0-7f86c375dba7" containerName="oc" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.159532 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e641956c-57e8-4eed-9cb0-7f86c375dba7" containerName="oc" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.159943 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e641956c-57e8-4eed-9cb0-7f86c375dba7" containerName="oc" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.161002 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533406-hpwgx" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.163209 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.164390 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.165027 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.173827 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533406-hpwgx"] Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.195066 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwrj6\" (UniqueName: \"kubernetes.io/projected/6b116345-90ab-4ff1-a394-c50059ec64f9-kube-api-access-kwrj6\") pod \"auto-csr-approver-29533406-hpwgx\" (UID: \"6b116345-90ab-4ff1-a394-c50059ec64f9\") " pod="openshift-infra/auto-csr-approver-29533406-hpwgx" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.296667 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwrj6\" (UniqueName: \"kubernetes.io/projected/6b116345-90ab-4ff1-a394-c50059ec64f9-kube-api-access-kwrj6\") pod \"auto-csr-approver-29533406-hpwgx\" (UID: \"6b116345-90ab-4ff1-a394-c50059ec64f9\") " pod="openshift-infra/auto-csr-approver-29533406-hpwgx" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.317795 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwrj6\" (UniqueName: \"kubernetes.io/projected/6b116345-90ab-4ff1-a394-c50059ec64f9-kube-api-access-kwrj6\") pod \"auto-csr-approver-29533406-hpwgx\" (UID: \"6b116345-90ab-4ff1-a394-c50059ec64f9\") " pod="openshift-infra/auto-csr-approver-29533406-hpwgx" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.484966 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533406-hpwgx" Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.933291 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533406-hpwgx"] Feb 25 07:26:00 crc kubenswrapper[4978]: I0225 07:26:00.998355 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533406-hpwgx" event={"ID":"6b116345-90ab-4ff1-a394-c50059ec64f9","Type":"ContainerStarted","Data":"64be25bedfa8f328b50cd6338675f0be6fb87e799853271ed9575ae9850135ac"} Feb 25 07:26:03 crc kubenswrapper[4978]: I0225 07:26:03.020998 4978 generic.go:334] "Generic (PLEG): container finished" podID="6b116345-90ab-4ff1-a394-c50059ec64f9" containerID="035bceb27599e2dfaef5d6a3b5b2a5bd925498e17fda1d594b7f6929967d93a5" exitCode=0 Feb 25 07:26:03 crc kubenswrapper[4978]: I0225 07:26:03.021090 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533406-hpwgx" event={"ID":"6b116345-90ab-4ff1-a394-c50059ec64f9","Type":"ContainerDied","Data":"035bceb27599e2dfaef5d6a3b5b2a5bd925498e17fda1d594b7f6929967d93a5"} Feb 25 07:26:04 crc kubenswrapper[4978]: I0225 07:26:04.395798 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533406-hpwgx" Feb 25 07:26:04 crc kubenswrapper[4978]: I0225 07:26:04.564655 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwrj6\" (UniqueName: \"kubernetes.io/projected/6b116345-90ab-4ff1-a394-c50059ec64f9-kube-api-access-kwrj6\") pod \"6b116345-90ab-4ff1-a394-c50059ec64f9\" (UID: \"6b116345-90ab-4ff1-a394-c50059ec64f9\") " Feb 25 07:26:04 crc kubenswrapper[4978]: I0225 07:26:04.574603 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b116345-90ab-4ff1-a394-c50059ec64f9-kube-api-access-kwrj6" (OuterVolumeSpecName: "kube-api-access-kwrj6") pod "6b116345-90ab-4ff1-a394-c50059ec64f9" (UID: "6b116345-90ab-4ff1-a394-c50059ec64f9"). InnerVolumeSpecName "kube-api-access-kwrj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:26:04 crc kubenswrapper[4978]: I0225 07:26:04.667370 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwrj6\" (UniqueName: \"kubernetes.io/projected/6b116345-90ab-4ff1-a394-c50059ec64f9-kube-api-access-kwrj6\") on node \"crc\" DevicePath \"\"" Feb 25 07:26:05 crc kubenswrapper[4978]: I0225 07:26:05.043714 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533406-hpwgx" event={"ID":"6b116345-90ab-4ff1-a394-c50059ec64f9","Type":"ContainerDied","Data":"64be25bedfa8f328b50cd6338675f0be6fb87e799853271ed9575ae9850135ac"} Feb 25 07:26:05 crc kubenswrapper[4978]: I0225 07:26:05.043769 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64be25bedfa8f328b50cd6338675f0be6fb87e799853271ed9575ae9850135ac" Feb 25 07:26:05 crc kubenswrapper[4978]: I0225 07:26:05.043801 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533406-hpwgx" Feb 25 07:26:05 crc kubenswrapper[4978]: I0225 07:26:05.493934 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533400-qgkvz"] Feb 25 07:26:05 crc kubenswrapper[4978]: I0225 07:26:05.516736 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533400-qgkvz"] Feb 25 07:26:07 crc kubenswrapper[4978]: I0225 07:26:07.341248 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dca72eca-d532-4e96-a7c1-ef885235eab2" path="/var/lib/kubelet/pods/dca72eca-d532-4e96-a7c1-ef885235eab2/volumes" Feb 25 07:26:08 crc kubenswrapper[4978]: I0225 07:26:08.329186 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:26:08 crc kubenswrapper[4978]: E0225 07:26:08.329837 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:26:21 crc kubenswrapper[4978]: I0225 07:26:21.715987 4978 scope.go:117] "RemoveContainer" containerID="dc5c78c9f8b9657b1d4706fbe0a6320719bf4b7fc37bfd9715012051e712e823" Feb 25 07:26:23 crc kubenswrapper[4978]: I0225 07:26:23.328263 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:26:23 crc kubenswrapper[4978]: E0225 07:26:23.329022 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:26:35 crc kubenswrapper[4978]: I0225 07:26:35.328297 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:26:35 crc kubenswrapper[4978]: E0225 07:26:35.329248 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:26:47 crc kubenswrapper[4978]: I0225 07:26:47.331476 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:26:47 crc kubenswrapper[4978]: E0225 07:26:47.332258 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:26:59 crc kubenswrapper[4978]: I0225 07:26:59.328432 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:26:59 crc kubenswrapper[4978]: E0225 07:26:59.329121 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:27:10 crc kubenswrapper[4978]: I0225 07:27:10.328613 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:27:10 crc kubenswrapper[4978]: E0225 07:27:10.329411 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:27:25 crc kubenswrapper[4978]: I0225 07:27:25.327967 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:27:25 crc kubenswrapper[4978]: E0225 07:27:25.328817 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:27:40 crc kubenswrapper[4978]: I0225 07:27:40.328476 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:27:40 crc kubenswrapper[4978]: E0225 07:27:40.331130 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:27:52 crc kubenswrapper[4978]: I0225 07:27:52.327893 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:27:53 crc kubenswrapper[4978]: I0225 07:27:53.262983 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"164f20c4dafdd6bdb0816971bff3c38ceeea3c62076bcdad2e26cfca1cfb00a2"} Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.160814 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533408-lr9bq"] Feb 25 07:28:00 crc kubenswrapper[4978]: E0225 07:28:00.161616 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b116345-90ab-4ff1-a394-c50059ec64f9" containerName="oc" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.161637 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b116345-90ab-4ff1-a394-c50059ec64f9" containerName="oc" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.161858 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b116345-90ab-4ff1-a394-c50059ec64f9" containerName="oc" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.162570 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533408-lr9bq" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.166315 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.171793 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533408-lr9bq"] Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.175136 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.175737 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.322870 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmx6x\" (UniqueName: \"kubernetes.io/projected/a072887d-45d9-429c-866e-71d1dd6d726c-kube-api-access-bmx6x\") pod \"auto-csr-approver-29533408-lr9bq\" (UID: \"a072887d-45d9-429c-866e-71d1dd6d726c\") " pod="openshift-infra/auto-csr-approver-29533408-lr9bq" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.424810 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmx6x\" (UniqueName: \"kubernetes.io/projected/a072887d-45d9-429c-866e-71d1dd6d726c-kube-api-access-bmx6x\") pod \"auto-csr-approver-29533408-lr9bq\" (UID: \"a072887d-45d9-429c-866e-71d1dd6d726c\") " pod="openshift-infra/auto-csr-approver-29533408-lr9bq" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.469742 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmx6x\" (UniqueName: \"kubernetes.io/projected/a072887d-45d9-429c-866e-71d1dd6d726c-kube-api-access-bmx6x\") pod \"auto-csr-approver-29533408-lr9bq\" (UID: \"a072887d-45d9-429c-866e-71d1dd6d726c\") " pod="openshift-infra/auto-csr-approver-29533408-lr9bq" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.502528 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533408-lr9bq" Feb 25 07:28:00 crc kubenswrapper[4978]: I0225 07:28:00.990143 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533408-lr9bq"] Feb 25 07:28:01 crc kubenswrapper[4978]: I0225 07:28:01.004830 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 07:28:01 crc kubenswrapper[4978]: I0225 07:28:01.343807 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533408-lr9bq" event={"ID":"a072887d-45d9-429c-866e-71d1dd6d726c","Type":"ContainerStarted","Data":"8536d8ea897df8ff8867cd4d5f9be436eb4a1d8641c9d5f4ef3d004eaf76c279"} Feb 25 07:28:02 crc kubenswrapper[4978]: I0225 07:28:02.356824 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533408-lr9bq" event={"ID":"a072887d-45d9-429c-866e-71d1dd6d726c","Type":"ContainerStarted","Data":"355baa64dfffcefe6b7012720bc7447b6d9bf2ce88e72b6486e64dd219961ea2"} Feb 25 07:28:02 crc kubenswrapper[4978]: I0225 07:28:02.384432 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533408-lr9bq" podStartSLOduration=1.492154326 podStartE2EDuration="2.384410932s" podCreationTimestamp="2026-02-25 07:28:00 +0000 UTC" firstStartedPulling="2026-02-25 07:28:01.004284395 +0000 UTC m=+2574.443540894" lastFinishedPulling="2026-02-25 07:28:01.896541021 +0000 UTC m=+2575.335797500" observedRunningTime="2026-02-25 07:28:02.381827212 +0000 UTC m=+2575.821083751" watchObservedRunningTime="2026-02-25 07:28:02.384410932 +0000 UTC m=+2575.823667401" Feb 25 07:28:03 crc kubenswrapper[4978]: I0225 07:28:03.369430 4978 generic.go:334] "Generic (PLEG): container finished" podID="a072887d-45d9-429c-866e-71d1dd6d726c" containerID="355baa64dfffcefe6b7012720bc7447b6d9bf2ce88e72b6486e64dd219961ea2" exitCode=0 Feb 25 07:28:03 crc kubenswrapper[4978]: I0225 07:28:03.369506 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533408-lr9bq" event={"ID":"a072887d-45d9-429c-866e-71d1dd6d726c","Type":"ContainerDied","Data":"355baa64dfffcefe6b7012720bc7447b6d9bf2ce88e72b6486e64dd219961ea2"} Feb 25 07:28:04 crc kubenswrapper[4978]: I0225 07:28:04.675287 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533408-lr9bq" Feb 25 07:28:04 crc kubenswrapper[4978]: I0225 07:28:04.796524 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmx6x\" (UniqueName: \"kubernetes.io/projected/a072887d-45d9-429c-866e-71d1dd6d726c-kube-api-access-bmx6x\") pod \"a072887d-45d9-429c-866e-71d1dd6d726c\" (UID: \"a072887d-45d9-429c-866e-71d1dd6d726c\") " Feb 25 07:28:04 crc kubenswrapper[4978]: I0225 07:28:04.803847 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a072887d-45d9-429c-866e-71d1dd6d726c-kube-api-access-bmx6x" (OuterVolumeSpecName: "kube-api-access-bmx6x") pod "a072887d-45d9-429c-866e-71d1dd6d726c" (UID: "a072887d-45d9-429c-866e-71d1dd6d726c"). InnerVolumeSpecName "kube-api-access-bmx6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:28:04 crc kubenswrapper[4978]: I0225 07:28:04.898749 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmx6x\" (UniqueName: \"kubernetes.io/projected/a072887d-45d9-429c-866e-71d1dd6d726c-kube-api-access-bmx6x\") on node \"crc\" DevicePath \"\"" Feb 25 07:28:05 crc kubenswrapper[4978]: I0225 07:28:05.394279 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533408-lr9bq" event={"ID":"a072887d-45d9-429c-866e-71d1dd6d726c","Type":"ContainerDied","Data":"8536d8ea897df8ff8867cd4d5f9be436eb4a1d8641c9d5f4ef3d004eaf76c279"} Feb 25 07:28:05 crc kubenswrapper[4978]: I0225 07:28:05.394336 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8536d8ea897df8ff8867cd4d5f9be436eb4a1d8641c9d5f4ef3d004eaf76c279" Feb 25 07:28:05 crc kubenswrapper[4978]: I0225 07:28:05.394425 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533408-lr9bq" Feb 25 07:28:05 crc kubenswrapper[4978]: I0225 07:28:05.761212 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533402-vl77q"] Feb 25 07:28:05 crc kubenswrapper[4978]: I0225 07:28:05.766718 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533402-vl77q"] Feb 25 07:28:07 crc kubenswrapper[4978]: I0225 07:28:07.344363 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e58c0d5d-8529-4d34-9621-9f367f9d6d8d" path="/var/lib/kubelet/pods/e58c0d5d-8529-4d34-9621-9f367f9d6d8d/volumes" Feb 25 07:28:21 crc kubenswrapper[4978]: I0225 07:28:21.840322 4978 scope.go:117] "RemoveContainer" containerID="f32df57c3ddee4cbc7724e109664feeeec9dd24e1a73261957545b9eefb460b0" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.175940 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533410-z6bsp"] Feb 25 07:30:00 crc kubenswrapper[4978]: E0225 07:30:00.177168 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a072887d-45d9-429c-866e-71d1dd6d726c" containerName="oc" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.177193 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a072887d-45d9-429c-866e-71d1dd6d726c" containerName="oc" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.177611 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a072887d-45d9-429c-866e-71d1dd6d726c" containerName="oc" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.178726 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533410-z6bsp" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.185443 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.185491 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.185622 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.189019 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx"] Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.190425 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.192227 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.192737 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.201858 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533410-z6bsp"] Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.209643 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx"] Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.318476 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxtnf\" (UniqueName: \"kubernetes.io/projected/28650a47-547e-4ce5-b222-7c8ca2b6ebc4-kube-api-access-qxtnf\") pod \"auto-csr-approver-29533410-z6bsp\" (UID: \"28650a47-547e-4ce5-b222-7c8ca2b6ebc4\") " pod="openshift-infra/auto-csr-approver-29533410-z6bsp" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.318547 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-722bc\" (UniqueName: \"kubernetes.io/projected/c3fbc3ff-fd50-4030-be63-750709b2f3fb-kube-api-access-722bc\") pod \"collect-profiles-29533410-rzrpx\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.318723 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c3fbc3ff-fd50-4030-be63-750709b2f3fb-secret-volume\") pod \"collect-profiles-29533410-rzrpx\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.318777 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c3fbc3ff-fd50-4030-be63-750709b2f3fb-config-volume\") pod \"collect-profiles-29533410-rzrpx\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.419634 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c3fbc3ff-fd50-4030-be63-750709b2f3fb-secret-volume\") pod \"collect-profiles-29533410-rzrpx\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.420241 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c3fbc3ff-fd50-4030-be63-750709b2f3fb-config-volume\") pod \"collect-profiles-29533410-rzrpx\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.420586 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxtnf\" (UniqueName: \"kubernetes.io/projected/28650a47-547e-4ce5-b222-7c8ca2b6ebc4-kube-api-access-qxtnf\") pod \"auto-csr-approver-29533410-z6bsp\" (UID: \"28650a47-547e-4ce5-b222-7c8ca2b6ebc4\") " pod="openshift-infra/auto-csr-approver-29533410-z6bsp" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.420787 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-722bc\" (UniqueName: \"kubernetes.io/projected/c3fbc3ff-fd50-4030-be63-750709b2f3fb-kube-api-access-722bc\") pod \"collect-profiles-29533410-rzrpx\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.423066 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c3fbc3ff-fd50-4030-be63-750709b2f3fb-config-volume\") pod \"collect-profiles-29533410-rzrpx\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.440006 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c3fbc3ff-fd50-4030-be63-750709b2f3fb-secret-volume\") pod \"collect-profiles-29533410-rzrpx\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.445331 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxtnf\" (UniqueName: \"kubernetes.io/projected/28650a47-547e-4ce5-b222-7c8ca2b6ebc4-kube-api-access-qxtnf\") pod \"auto-csr-approver-29533410-z6bsp\" (UID: \"28650a47-547e-4ce5-b222-7c8ca2b6ebc4\") " pod="openshift-infra/auto-csr-approver-29533410-z6bsp" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.450924 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-722bc\" (UniqueName: \"kubernetes.io/projected/c3fbc3ff-fd50-4030-be63-750709b2f3fb-kube-api-access-722bc\") pod \"collect-profiles-29533410-rzrpx\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.508954 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533410-z6bsp" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.529478 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.992758 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx"] Feb 25 07:30:00 crc kubenswrapper[4978]: I0225 07:30:00.997434 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533410-z6bsp"] Feb 25 07:30:01 crc kubenswrapper[4978]: I0225 07:30:01.483476 4978 generic.go:334] "Generic (PLEG): container finished" podID="c3fbc3ff-fd50-4030-be63-750709b2f3fb" containerID="2c0820c4fa4f3842a8ee5ce9b50c8d18b1c11f736387c8957a132224ad5854c1" exitCode=0 Feb 25 07:30:01 crc kubenswrapper[4978]: I0225 07:30:01.483578 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" event={"ID":"c3fbc3ff-fd50-4030-be63-750709b2f3fb","Type":"ContainerDied","Data":"2c0820c4fa4f3842a8ee5ce9b50c8d18b1c11f736387c8957a132224ad5854c1"} Feb 25 07:30:01 crc kubenswrapper[4978]: I0225 07:30:01.483908 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" event={"ID":"c3fbc3ff-fd50-4030-be63-750709b2f3fb","Type":"ContainerStarted","Data":"7729c144232f22a0227cd7e84359ea91dbead19564839ae50dcaf76aea866a68"} Feb 25 07:30:01 crc kubenswrapper[4978]: I0225 07:30:01.485533 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533410-z6bsp" event={"ID":"28650a47-547e-4ce5-b222-7c8ca2b6ebc4","Type":"ContainerStarted","Data":"76d8862cd76bba6ae12d66a51eb0e76f4d54718d5f4bdf5fe33cfa2c8287322d"} Feb 25 07:30:02 crc kubenswrapper[4978]: I0225 07:30:02.771088 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:02 crc kubenswrapper[4978]: I0225 07:30:02.860073 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c3fbc3ff-fd50-4030-be63-750709b2f3fb-secret-volume\") pod \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " Feb 25 07:30:02 crc kubenswrapper[4978]: I0225 07:30:02.860178 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c3fbc3ff-fd50-4030-be63-750709b2f3fb-config-volume\") pod \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " Feb 25 07:30:02 crc kubenswrapper[4978]: I0225 07:30:02.860200 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-722bc\" (UniqueName: \"kubernetes.io/projected/c3fbc3ff-fd50-4030-be63-750709b2f3fb-kube-api-access-722bc\") pod \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\" (UID: \"c3fbc3ff-fd50-4030-be63-750709b2f3fb\") " Feb 25 07:30:02 crc kubenswrapper[4978]: I0225 07:30:02.861419 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3fbc3ff-fd50-4030-be63-750709b2f3fb-config-volume" (OuterVolumeSpecName: "config-volume") pod "c3fbc3ff-fd50-4030-be63-750709b2f3fb" (UID: "c3fbc3ff-fd50-4030-be63-750709b2f3fb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:30:02 crc kubenswrapper[4978]: I0225 07:30:02.866603 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3fbc3ff-fd50-4030-be63-750709b2f3fb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c3fbc3ff-fd50-4030-be63-750709b2f3fb" (UID: "c3fbc3ff-fd50-4030-be63-750709b2f3fb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:30:02 crc kubenswrapper[4978]: I0225 07:30:02.866954 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3fbc3ff-fd50-4030-be63-750709b2f3fb-kube-api-access-722bc" (OuterVolumeSpecName: "kube-api-access-722bc") pod "c3fbc3ff-fd50-4030-be63-750709b2f3fb" (UID: "c3fbc3ff-fd50-4030-be63-750709b2f3fb"). InnerVolumeSpecName "kube-api-access-722bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:30:02 crc kubenswrapper[4978]: I0225 07:30:02.962156 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c3fbc3ff-fd50-4030-be63-750709b2f3fb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 07:30:02 crc kubenswrapper[4978]: I0225 07:30:02.962213 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-722bc\" (UniqueName: \"kubernetes.io/projected/c3fbc3ff-fd50-4030-be63-750709b2f3fb-kube-api-access-722bc\") on node \"crc\" DevicePath \"\"" Feb 25 07:30:02 crc kubenswrapper[4978]: I0225 07:30:02.962236 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c3fbc3ff-fd50-4030-be63-750709b2f3fb-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 07:30:03 crc kubenswrapper[4978]: I0225 07:30:03.506360 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" Feb 25 07:30:03 crc kubenswrapper[4978]: I0225 07:30:03.506408 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx" event={"ID":"c3fbc3ff-fd50-4030-be63-750709b2f3fb","Type":"ContainerDied","Data":"7729c144232f22a0227cd7e84359ea91dbead19564839ae50dcaf76aea866a68"} Feb 25 07:30:03 crc kubenswrapper[4978]: I0225 07:30:03.506971 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7729c144232f22a0227cd7e84359ea91dbead19564839ae50dcaf76aea866a68" Feb 25 07:30:03 crc kubenswrapper[4978]: I0225 07:30:03.509697 4978 generic.go:334] "Generic (PLEG): container finished" podID="28650a47-547e-4ce5-b222-7c8ca2b6ebc4" containerID="277feb2be61c626c93abb9014ae4c52855d3167af06e514dca0f29d7df2f8703" exitCode=0 Feb 25 07:30:03 crc kubenswrapper[4978]: I0225 07:30:03.509752 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533410-z6bsp" event={"ID":"28650a47-547e-4ce5-b222-7c8ca2b6ebc4","Type":"ContainerDied","Data":"277feb2be61c626c93abb9014ae4c52855d3167af06e514dca0f29d7df2f8703"} Feb 25 07:30:03 crc kubenswrapper[4978]: I0225 07:30:03.877043 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk"] Feb 25 07:30:03 crc kubenswrapper[4978]: I0225 07:30:03.887063 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533365-krpxk"] Feb 25 07:30:04 crc kubenswrapper[4978]: I0225 07:30:04.873476 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533410-z6bsp" Feb 25 07:30:04 crc kubenswrapper[4978]: I0225 07:30:04.997216 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxtnf\" (UniqueName: \"kubernetes.io/projected/28650a47-547e-4ce5-b222-7c8ca2b6ebc4-kube-api-access-qxtnf\") pod \"28650a47-547e-4ce5-b222-7c8ca2b6ebc4\" (UID: \"28650a47-547e-4ce5-b222-7c8ca2b6ebc4\") " Feb 25 07:30:05 crc kubenswrapper[4978]: I0225 07:30:05.003192 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28650a47-547e-4ce5-b222-7c8ca2b6ebc4-kube-api-access-qxtnf" (OuterVolumeSpecName: "kube-api-access-qxtnf") pod "28650a47-547e-4ce5-b222-7c8ca2b6ebc4" (UID: "28650a47-547e-4ce5-b222-7c8ca2b6ebc4"). InnerVolumeSpecName "kube-api-access-qxtnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:30:05 crc kubenswrapper[4978]: I0225 07:30:05.099198 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxtnf\" (UniqueName: \"kubernetes.io/projected/28650a47-547e-4ce5-b222-7c8ca2b6ebc4-kube-api-access-qxtnf\") on node \"crc\" DevicePath \"\"" Feb 25 07:30:05 crc kubenswrapper[4978]: I0225 07:30:05.345964 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fd9f3d7-d711-43cb-a896-7c77507621bc" path="/var/lib/kubelet/pods/7fd9f3d7-d711-43cb-a896-7c77507621bc/volumes" Feb 25 07:30:05 crc kubenswrapper[4978]: I0225 07:30:05.530338 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533410-z6bsp" event={"ID":"28650a47-547e-4ce5-b222-7c8ca2b6ebc4","Type":"ContainerDied","Data":"76d8862cd76bba6ae12d66a51eb0e76f4d54718d5f4bdf5fe33cfa2c8287322d"} Feb 25 07:30:05 crc kubenswrapper[4978]: I0225 07:30:05.530470 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76d8862cd76bba6ae12d66a51eb0e76f4d54718d5f4bdf5fe33cfa2c8287322d" Feb 25 07:30:05 crc kubenswrapper[4978]: I0225 07:30:05.530583 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533410-z6bsp" Feb 25 07:30:05 crc kubenswrapper[4978]: I0225 07:30:05.949814 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533404-p6ms8"] Feb 25 07:30:05 crc kubenswrapper[4978]: I0225 07:30:05.956993 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533404-p6ms8"] Feb 25 07:30:07 crc kubenswrapper[4978]: I0225 07:30:07.346108 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e641956c-57e8-4eed-9cb0-7f86c375dba7" path="/var/lib/kubelet/pods/e641956c-57e8-4eed-9cb0-7f86c375dba7/volumes" Feb 25 07:30:16 crc kubenswrapper[4978]: I0225 07:30:16.540991 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:30:16 crc kubenswrapper[4978]: I0225 07:30:16.541608 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:30:21 crc kubenswrapper[4978]: I0225 07:30:21.952868 4978 scope.go:117] "RemoveContainer" containerID="8a9a06d5f426c85882f2a71024311846c9ece01c436f2a16be0cc489eade32d7" Feb 25 07:30:21 crc kubenswrapper[4978]: I0225 07:30:21.987240 4978 scope.go:117] "RemoveContainer" containerID="d86c99f4652655abe18d9b1aff127786d70e31b7a01a59b241f2a0cb90875d68" Feb 25 07:30:46 crc kubenswrapper[4978]: I0225 07:30:46.540465 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:30:46 crc kubenswrapper[4978]: I0225 07:30:46.540958 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:30:50 crc kubenswrapper[4978]: I0225 07:30:50.809651 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kkbdw"] Feb 25 07:30:50 crc kubenswrapper[4978]: E0225 07:30:50.812404 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28650a47-547e-4ce5-b222-7c8ca2b6ebc4" containerName="oc" Feb 25 07:30:50 crc kubenswrapper[4978]: I0225 07:30:50.812434 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="28650a47-547e-4ce5-b222-7c8ca2b6ebc4" containerName="oc" Feb 25 07:30:50 crc kubenswrapper[4978]: E0225 07:30:50.812490 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3fbc3ff-fd50-4030-be63-750709b2f3fb" containerName="collect-profiles" Feb 25 07:30:50 crc kubenswrapper[4978]: I0225 07:30:50.812503 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3fbc3ff-fd50-4030-be63-750709b2f3fb" containerName="collect-profiles" Feb 25 07:30:50 crc kubenswrapper[4978]: I0225 07:30:50.812726 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3fbc3ff-fd50-4030-be63-750709b2f3fb" containerName="collect-profiles" Feb 25 07:30:50 crc kubenswrapper[4978]: I0225 07:30:50.812764 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="28650a47-547e-4ce5-b222-7c8ca2b6ebc4" containerName="oc" Feb 25 07:30:50 crc kubenswrapper[4978]: I0225 07:30:50.814479 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:50 crc kubenswrapper[4978]: I0225 07:30:50.832213 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kkbdw"] Feb 25 07:30:50 crc kubenswrapper[4978]: I0225 07:30:50.976803 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-catalog-content\") pod \"certified-operators-kkbdw\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:50 crc kubenswrapper[4978]: I0225 07:30:50.977157 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpvqt\" (UniqueName: \"kubernetes.io/projected/e185ba6d-5331-4c88-b6fd-734a38e904a4-kube-api-access-hpvqt\") pod \"certified-operators-kkbdw\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:50 crc kubenswrapper[4978]: I0225 07:30:50.977401 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-utilities\") pod \"certified-operators-kkbdw\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.079046 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-utilities\") pod \"certified-operators-kkbdw\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.078408 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-utilities\") pod \"certified-operators-kkbdw\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.079449 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-catalog-content\") pod \"certified-operators-kkbdw\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.079524 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpvqt\" (UniqueName: \"kubernetes.io/projected/e185ba6d-5331-4c88-b6fd-734a38e904a4-kube-api-access-hpvqt\") pod \"certified-operators-kkbdw\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.079956 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-catalog-content\") pod \"certified-operators-kkbdw\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.102017 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpvqt\" (UniqueName: \"kubernetes.io/projected/e185ba6d-5331-4c88-b6fd-734a38e904a4-kube-api-access-hpvqt\") pod \"certified-operators-kkbdw\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.142487 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.612903 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kkbdw"] Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.966350 4978 generic.go:334] "Generic (PLEG): container finished" podID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerID="dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076" exitCode=0 Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.966490 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkbdw" event={"ID":"e185ba6d-5331-4c88-b6fd-734a38e904a4","Type":"ContainerDied","Data":"dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076"} Feb 25 07:30:51 crc kubenswrapper[4978]: I0225 07:30:51.966688 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkbdw" event={"ID":"e185ba6d-5331-4c88-b6fd-734a38e904a4","Type":"ContainerStarted","Data":"64b05645f3b85f09242a9a9c43e8b9abac765e9550e42da78a472a95c7420ddb"} Feb 25 07:30:52 crc kubenswrapper[4978]: I0225 07:30:52.986238 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkbdw" event={"ID":"e185ba6d-5331-4c88-b6fd-734a38e904a4","Type":"ContainerStarted","Data":"5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791"} Feb 25 07:30:53 crc kubenswrapper[4978]: I0225 07:30:53.997464 4978 generic.go:334] "Generic (PLEG): container finished" podID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerID="5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791" exitCode=0 Feb 25 07:30:53 crc kubenswrapper[4978]: I0225 07:30:53.997566 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkbdw" event={"ID":"e185ba6d-5331-4c88-b6fd-734a38e904a4","Type":"ContainerDied","Data":"5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791"} Feb 25 07:30:55 crc kubenswrapper[4978]: I0225 07:30:55.008861 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkbdw" event={"ID":"e185ba6d-5331-4c88-b6fd-734a38e904a4","Type":"ContainerStarted","Data":"5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a"} Feb 25 07:30:55 crc kubenswrapper[4978]: I0225 07:30:55.041920 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kkbdw" podStartSLOduration=2.546666972 podStartE2EDuration="5.041887002s" podCreationTimestamp="2026-02-25 07:30:50 +0000 UTC" firstStartedPulling="2026-02-25 07:30:51.968723929 +0000 UTC m=+2745.407980418" lastFinishedPulling="2026-02-25 07:30:54.463943949 +0000 UTC m=+2747.903200448" observedRunningTime="2026-02-25 07:30:55.029272671 +0000 UTC m=+2748.468529180" watchObservedRunningTime="2026-02-25 07:30:55.041887002 +0000 UTC m=+2748.481143501" Feb 25 07:31:01 crc kubenswrapper[4978]: I0225 07:31:01.143315 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:31:01 crc kubenswrapper[4978]: I0225 07:31:01.143913 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:31:01 crc kubenswrapper[4978]: I0225 07:31:01.186951 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:31:02 crc kubenswrapper[4978]: I0225 07:31:02.132001 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:31:02 crc kubenswrapper[4978]: I0225 07:31:02.204793 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kkbdw"] Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.088486 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kkbdw" podUID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerName="registry-server" containerID="cri-o://5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a" gracePeriod=2 Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.609706 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.693510 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpvqt\" (UniqueName: \"kubernetes.io/projected/e185ba6d-5331-4c88-b6fd-734a38e904a4-kube-api-access-hpvqt\") pod \"e185ba6d-5331-4c88-b6fd-734a38e904a4\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.693625 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-catalog-content\") pod \"e185ba6d-5331-4c88-b6fd-734a38e904a4\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.693704 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-utilities\") pod \"e185ba6d-5331-4c88-b6fd-734a38e904a4\" (UID: \"e185ba6d-5331-4c88-b6fd-734a38e904a4\") " Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.694555 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-utilities" (OuterVolumeSpecName: "utilities") pod "e185ba6d-5331-4c88-b6fd-734a38e904a4" (UID: "e185ba6d-5331-4c88-b6fd-734a38e904a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.698488 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e185ba6d-5331-4c88-b6fd-734a38e904a4-kube-api-access-hpvqt" (OuterVolumeSpecName: "kube-api-access-hpvqt") pod "e185ba6d-5331-4c88-b6fd-734a38e904a4" (UID: "e185ba6d-5331-4c88-b6fd-734a38e904a4"). InnerVolumeSpecName "kube-api-access-hpvqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.751194 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e185ba6d-5331-4c88-b6fd-734a38e904a4" (UID: "e185ba6d-5331-4c88-b6fd-734a38e904a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.795578 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.795607 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e185ba6d-5331-4c88-b6fd-734a38e904a4-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:31:04 crc kubenswrapper[4978]: I0225 07:31:04.795618 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpvqt\" (UniqueName: \"kubernetes.io/projected/e185ba6d-5331-4c88-b6fd-734a38e904a4-kube-api-access-hpvqt\") on node \"crc\" DevicePath \"\"" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.103472 4978 generic.go:334] "Generic (PLEG): container finished" podID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerID="5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a" exitCode=0 Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.103565 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkbdw" event={"ID":"e185ba6d-5331-4c88-b6fd-734a38e904a4","Type":"ContainerDied","Data":"5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a"} Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.103870 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kkbdw" event={"ID":"e185ba6d-5331-4c88-b6fd-734a38e904a4","Type":"ContainerDied","Data":"64b05645f3b85f09242a9a9c43e8b9abac765e9550e42da78a472a95c7420ddb"} Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.103906 4978 scope.go:117] "RemoveContainer" containerID="5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.103620 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kkbdw" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.131188 4978 scope.go:117] "RemoveContainer" containerID="5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.157525 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kkbdw"] Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.165989 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kkbdw"] Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.183905 4978 scope.go:117] "RemoveContainer" containerID="dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.204634 4978 scope.go:117] "RemoveContainer" containerID="5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a" Feb 25 07:31:05 crc kubenswrapper[4978]: E0225 07:31:05.205147 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a\": container with ID starting with 5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a not found: ID does not exist" containerID="5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.205200 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a"} err="failed to get container status \"5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a\": rpc error: code = NotFound desc = could not find container \"5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a\": container with ID starting with 5089142343a989c7ccd381ec2dc99c15bde656fc80470351b518860e3b7fa20a not found: ID does not exist" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.205233 4978 scope.go:117] "RemoveContainer" containerID="5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791" Feb 25 07:31:05 crc kubenswrapper[4978]: E0225 07:31:05.205641 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791\": container with ID starting with 5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791 not found: ID does not exist" containerID="5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.205686 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791"} err="failed to get container status \"5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791\": rpc error: code = NotFound desc = could not find container \"5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791\": container with ID starting with 5222b065d4281f982423fddf01cfac9c7837e57e997bda55c42e382eb9969791 not found: ID does not exist" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.205716 4978 scope.go:117] "RemoveContainer" containerID="dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076" Feb 25 07:31:05 crc kubenswrapper[4978]: E0225 07:31:05.206243 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076\": container with ID starting with dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076 not found: ID does not exist" containerID="dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.206283 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076"} err="failed to get container status \"dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076\": rpc error: code = NotFound desc = could not find container \"dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076\": container with ID starting with dedb304e1b8d8567f10d794c363ff08f71098d7755a45ca1257c36d0f40c0076 not found: ID does not exist" Feb 25 07:31:05 crc kubenswrapper[4978]: I0225 07:31:05.342277 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e185ba6d-5331-4c88-b6fd-734a38e904a4" path="/var/lib/kubelet/pods/e185ba6d-5331-4c88-b6fd-734a38e904a4/volumes" Feb 25 07:31:16 crc kubenswrapper[4978]: I0225 07:31:16.540508 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:31:16 crc kubenswrapper[4978]: I0225 07:31:16.541248 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:31:16 crc kubenswrapper[4978]: I0225 07:31:16.541338 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:31:16 crc kubenswrapper[4978]: I0225 07:31:16.542211 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"164f20c4dafdd6bdb0816971bff3c38ceeea3c62076bcdad2e26cfca1cfb00a2"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:31:16 crc kubenswrapper[4978]: I0225 07:31:16.542296 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://164f20c4dafdd6bdb0816971bff3c38ceeea3c62076bcdad2e26cfca1cfb00a2" gracePeriod=600 Feb 25 07:31:17 crc kubenswrapper[4978]: I0225 07:31:17.219592 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="164f20c4dafdd6bdb0816971bff3c38ceeea3c62076bcdad2e26cfca1cfb00a2" exitCode=0 Feb 25 07:31:17 crc kubenswrapper[4978]: I0225 07:31:17.220040 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"164f20c4dafdd6bdb0816971bff3c38ceeea3c62076bcdad2e26cfca1cfb00a2"} Feb 25 07:31:17 crc kubenswrapper[4978]: I0225 07:31:17.220065 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754"} Feb 25 07:31:17 crc kubenswrapper[4978]: I0225 07:31:17.220083 4978 scope.go:117] "RemoveContainer" containerID="0f73313c83d15f830f5463ec167590dbe21f665afca6b429ec96e29a034a877c" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.188616 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tf6xt"] Feb 25 07:31:29 crc kubenswrapper[4978]: E0225 07:31:29.189985 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerName="extract-utilities" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.190017 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerName="extract-utilities" Feb 25 07:31:29 crc kubenswrapper[4978]: E0225 07:31:29.190063 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerName="registry-server" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.190080 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerName="registry-server" Feb 25 07:31:29 crc kubenswrapper[4978]: E0225 07:31:29.190107 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerName="extract-content" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.190123 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerName="extract-content" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.190534 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e185ba6d-5331-4c88-b6fd-734a38e904a4" containerName="registry-server" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.193097 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.207537 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tf6xt"] Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.291502 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-catalog-content\") pod \"redhat-operators-tf6xt\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.291590 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-utilities\") pod \"redhat-operators-tf6xt\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.291622 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfzx4\" (UniqueName: \"kubernetes.io/projected/b180623e-c84d-41c4-be6e-8c38753a15a2-kube-api-access-qfzx4\") pod \"redhat-operators-tf6xt\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.392569 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-utilities\") pod \"redhat-operators-tf6xt\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.392911 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfzx4\" (UniqueName: \"kubernetes.io/projected/b180623e-c84d-41c4-be6e-8c38753a15a2-kube-api-access-qfzx4\") pod \"redhat-operators-tf6xt\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.393063 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-catalog-content\") pod \"redhat-operators-tf6xt\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.393628 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-catalog-content\") pod \"redhat-operators-tf6xt\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.393976 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-utilities\") pod \"redhat-operators-tf6xt\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.420730 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfzx4\" (UniqueName: \"kubernetes.io/projected/b180623e-c84d-41c4-be6e-8c38753a15a2-kube-api-access-qfzx4\") pod \"redhat-operators-tf6xt\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.523784 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:29 crc kubenswrapper[4978]: I0225 07:31:29.977064 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tf6xt"] Feb 25 07:31:29 crc kubenswrapper[4978]: W0225 07:31:29.978985 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb180623e_c84d_41c4_be6e_8c38753a15a2.slice/crio-43f0b2c996c223cf893152ba139103d3349fbcbeb68a76882b70654f3dcc13e2 WatchSource:0}: Error finding container 43f0b2c996c223cf893152ba139103d3349fbcbeb68a76882b70654f3dcc13e2: Status 404 returned error can't find the container with id 43f0b2c996c223cf893152ba139103d3349fbcbeb68a76882b70654f3dcc13e2 Feb 25 07:31:30 crc kubenswrapper[4978]: I0225 07:31:30.351922 4978 generic.go:334] "Generic (PLEG): container finished" podID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerID="48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8" exitCode=0 Feb 25 07:31:30 crc kubenswrapper[4978]: I0225 07:31:30.352037 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tf6xt" event={"ID":"b180623e-c84d-41c4-be6e-8c38753a15a2","Type":"ContainerDied","Data":"48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8"} Feb 25 07:31:30 crc kubenswrapper[4978]: I0225 07:31:30.352202 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tf6xt" event={"ID":"b180623e-c84d-41c4-be6e-8c38753a15a2","Type":"ContainerStarted","Data":"43f0b2c996c223cf893152ba139103d3349fbcbeb68a76882b70654f3dcc13e2"} Feb 25 07:31:31 crc kubenswrapper[4978]: I0225 07:31:31.373741 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tf6xt" event={"ID":"b180623e-c84d-41c4-be6e-8c38753a15a2","Type":"ContainerStarted","Data":"6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b"} Feb 25 07:31:32 crc kubenswrapper[4978]: I0225 07:31:32.387631 4978 generic.go:334] "Generic (PLEG): container finished" podID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerID="6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b" exitCode=0 Feb 25 07:31:32 crc kubenswrapper[4978]: I0225 07:31:32.387767 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tf6xt" event={"ID":"b180623e-c84d-41c4-be6e-8c38753a15a2","Type":"ContainerDied","Data":"6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b"} Feb 25 07:31:33 crc kubenswrapper[4978]: I0225 07:31:33.400349 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tf6xt" event={"ID":"b180623e-c84d-41c4-be6e-8c38753a15a2","Type":"ContainerStarted","Data":"c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c"} Feb 25 07:31:33 crc kubenswrapper[4978]: I0225 07:31:33.426244 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tf6xt" podStartSLOduration=1.834267492 podStartE2EDuration="4.42622884s" podCreationTimestamp="2026-02-25 07:31:29 +0000 UTC" firstStartedPulling="2026-02-25 07:31:30.353947964 +0000 UTC m=+2783.793204463" lastFinishedPulling="2026-02-25 07:31:32.945909322 +0000 UTC m=+2786.385165811" observedRunningTime="2026-02-25 07:31:33.421693639 +0000 UTC m=+2786.860950108" watchObservedRunningTime="2026-02-25 07:31:33.42622884 +0000 UTC m=+2786.865485289" Feb 25 07:31:39 crc kubenswrapper[4978]: I0225 07:31:39.523993 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:39 crc kubenswrapper[4978]: I0225 07:31:39.524681 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:40 crc kubenswrapper[4978]: I0225 07:31:40.570182 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tf6xt" podUID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerName="registry-server" probeResult="failure" output=< Feb 25 07:31:40 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 07:31:40 crc kubenswrapper[4978]: > Feb 25 07:31:49 crc kubenswrapper[4978]: I0225 07:31:49.596722 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:49 crc kubenswrapper[4978]: I0225 07:31:49.672227 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:49 crc kubenswrapper[4978]: I0225 07:31:49.861158 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tf6xt"] Feb 25 07:31:51 crc kubenswrapper[4978]: I0225 07:31:51.567160 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tf6xt" podUID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerName="registry-server" containerID="cri-o://c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c" gracePeriod=2 Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.044169 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.165750 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-utilities\") pod \"b180623e-c84d-41c4-be6e-8c38753a15a2\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.165815 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfzx4\" (UniqueName: \"kubernetes.io/projected/b180623e-c84d-41c4-be6e-8c38753a15a2-kube-api-access-qfzx4\") pod \"b180623e-c84d-41c4-be6e-8c38753a15a2\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.165896 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-catalog-content\") pod \"b180623e-c84d-41c4-be6e-8c38753a15a2\" (UID: \"b180623e-c84d-41c4-be6e-8c38753a15a2\") " Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.167264 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-utilities" (OuterVolumeSpecName: "utilities") pod "b180623e-c84d-41c4-be6e-8c38753a15a2" (UID: "b180623e-c84d-41c4-be6e-8c38753a15a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.170995 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b180623e-c84d-41c4-be6e-8c38753a15a2-kube-api-access-qfzx4" (OuterVolumeSpecName: "kube-api-access-qfzx4") pod "b180623e-c84d-41c4-be6e-8c38753a15a2" (UID: "b180623e-c84d-41c4-be6e-8c38753a15a2"). InnerVolumeSpecName "kube-api-access-qfzx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.267816 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.267859 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfzx4\" (UniqueName: \"kubernetes.io/projected/b180623e-c84d-41c4-be6e-8c38753a15a2-kube-api-access-qfzx4\") on node \"crc\" DevicePath \"\"" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.313133 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b180623e-c84d-41c4-be6e-8c38753a15a2" (UID: "b180623e-c84d-41c4-be6e-8c38753a15a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.369309 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b180623e-c84d-41c4-be6e-8c38753a15a2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.575280 4978 generic.go:334] "Generic (PLEG): container finished" podID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerID="c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c" exitCode=0 Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.575312 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tf6xt" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.575457 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tf6xt" event={"ID":"b180623e-c84d-41c4-be6e-8c38753a15a2","Type":"ContainerDied","Data":"c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c"} Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.575541 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tf6xt" event={"ID":"b180623e-c84d-41c4-be6e-8c38753a15a2","Type":"ContainerDied","Data":"43f0b2c996c223cf893152ba139103d3349fbcbeb68a76882b70654f3dcc13e2"} Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.575573 4978 scope.go:117] "RemoveContainer" containerID="c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.599784 4978 scope.go:117] "RemoveContainer" containerID="6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.613281 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tf6xt"] Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.618104 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tf6xt"] Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.639813 4978 scope.go:117] "RemoveContainer" containerID="48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.672121 4978 scope.go:117] "RemoveContainer" containerID="c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c" Feb 25 07:31:52 crc kubenswrapper[4978]: E0225 07:31:52.672502 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c\": container with ID starting with c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c not found: ID does not exist" containerID="c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.672544 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c"} err="failed to get container status \"c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c\": rpc error: code = NotFound desc = could not find container \"c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c\": container with ID starting with c0824d2ae3d295a6559934549008c82d4b4a9293d37492daa96a499187793e1c not found: ID does not exist" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.672571 4978 scope.go:117] "RemoveContainer" containerID="6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b" Feb 25 07:31:52 crc kubenswrapper[4978]: E0225 07:31:52.672771 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b\": container with ID starting with 6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b not found: ID does not exist" containerID="6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.672797 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b"} err="failed to get container status \"6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b\": rpc error: code = NotFound desc = could not find container \"6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b\": container with ID starting with 6c061795ed62157391bede046f18fd2288af317e95f093d1c7a688322f584a3b not found: ID does not exist" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.672820 4978 scope.go:117] "RemoveContainer" containerID="48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8" Feb 25 07:31:52 crc kubenswrapper[4978]: E0225 07:31:52.673031 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8\": container with ID starting with 48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8 not found: ID does not exist" containerID="48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8" Feb 25 07:31:52 crc kubenswrapper[4978]: I0225 07:31:52.673053 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8"} err="failed to get container status \"48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8\": rpc error: code = NotFound desc = could not find container \"48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8\": container with ID starting with 48e3411f3346a9949897240c32c88ba317cc65ab9585aa406719347161526aa8 not found: ID does not exist" Feb 25 07:31:53 crc kubenswrapper[4978]: I0225 07:31:53.335252 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b180623e-c84d-41c4-be6e-8c38753a15a2" path="/var/lib/kubelet/pods/b180623e-c84d-41c4-be6e-8c38753a15a2/volumes" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.153953 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533412-25jfm"] Feb 25 07:32:00 crc kubenswrapper[4978]: E0225 07:32:00.160225 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerName="registry-server" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.160326 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerName="registry-server" Feb 25 07:32:00 crc kubenswrapper[4978]: E0225 07:32:00.160450 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerName="extract-content" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.160542 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerName="extract-content" Feb 25 07:32:00 crc kubenswrapper[4978]: E0225 07:32:00.160631 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerName="extract-utilities" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.160695 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerName="extract-utilities" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.160915 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b180623e-c84d-41c4-be6e-8c38753a15a2" containerName="registry-server" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.161582 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533412-25jfm" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.162779 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533412-25jfm"] Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.163590 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.165233 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.167409 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.182132 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7x6v\" (UniqueName: \"kubernetes.io/projected/9ee093ca-f815-46cd-bcaf-cb1470142811-kube-api-access-d7x6v\") pod \"auto-csr-approver-29533412-25jfm\" (UID: \"9ee093ca-f815-46cd-bcaf-cb1470142811\") " pod="openshift-infra/auto-csr-approver-29533412-25jfm" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.284008 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7x6v\" (UniqueName: \"kubernetes.io/projected/9ee093ca-f815-46cd-bcaf-cb1470142811-kube-api-access-d7x6v\") pod \"auto-csr-approver-29533412-25jfm\" (UID: \"9ee093ca-f815-46cd-bcaf-cb1470142811\") " pod="openshift-infra/auto-csr-approver-29533412-25jfm" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.304128 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7x6v\" (UniqueName: \"kubernetes.io/projected/9ee093ca-f815-46cd-bcaf-cb1470142811-kube-api-access-d7x6v\") pod \"auto-csr-approver-29533412-25jfm\" (UID: \"9ee093ca-f815-46cd-bcaf-cb1470142811\") " pod="openshift-infra/auto-csr-approver-29533412-25jfm" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.484035 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533412-25jfm" Feb 25 07:32:00 crc kubenswrapper[4978]: I0225 07:32:00.851228 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533412-25jfm"] Feb 25 07:32:01 crc kubenswrapper[4978]: I0225 07:32:01.659169 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533412-25jfm" event={"ID":"9ee093ca-f815-46cd-bcaf-cb1470142811","Type":"ContainerStarted","Data":"f3dc9f302852a6d3ef2a5aad62f7477dfd3bc1f6b9ed7b41c009191339daa40c"} Feb 25 07:32:02 crc kubenswrapper[4978]: I0225 07:32:02.675518 4978 generic.go:334] "Generic (PLEG): container finished" podID="9ee093ca-f815-46cd-bcaf-cb1470142811" containerID="d67488c746b4786725fc9d7a030eedec62857adb971ed0e547d882635e7ff714" exitCode=0 Feb 25 07:32:02 crc kubenswrapper[4978]: I0225 07:32:02.675564 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533412-25jfm" event={"ID":"9ee093ca-f815-46cd-bcaf-cb1470142811","Type":"ContainerDied","Data":"d67488c746b4786725fc9d7a030eedec62857adb971ed0e547d882635e7ff714"} Feb 25 07:32:04 crc kubenswrapper[4978]: I0225 07:32:04.058671 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533412-25jfm" Feb 25 07:32:04 crc kubenswrapper[4978]: I0225 07:32:04.242113 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7x6v\" (UniqueName: \"kubernetes.io/projected/9ee093ca-f815-46cd-bcaf-cb1470142811-kube-api-access-d7x6v\") pod \"9ee093ca-f815-46cd-bcaf-cb1470142811\" (UID: \"9ee093ca-f815-46cd-bcaf-cb1470142811\") " Feb 25 07:32:04 crc kubenswrapper[4978]: I0225 07:32:04.248574 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ee093ca-f815-46cd-bcaf-cb1470142811-kube-api-access-d7x6v" (OuterVolumeSpecName: "kube-api-access-d7x6v") pod "9ee093ca-f815-46cd-bcaf-cb1470142811" (UID: "9ee093ca-f815-46cd-bcaf-cb1470142811"). InnerVolumeSpecName "kube-api-access-d7x6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:32:04 crc kubenswrapper[4978]: I0225 07:32:04.344483 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7x6v\" (UniqueName: \"kubernetes.io/projected/9ee093ca-f815-46cd-bcaf-cb1470142811-kube-api-access-d7x6v\") on node \"crc\" DevicePath \"\"" Feb 25 07:32:04 crc kubenswrapper[4978]: I0225 07:32:04.697298 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533412-25jfm" event={"ID":"9ee093ca-f815-46cd-bcaf-cb1470142811","Type":"ContainerDied","Data":"f3dc9f302852a6d3ef2a5aad62f7477dfd3bc1f6b9ed7b41c009191339daa40c"} Feb 25 07:32:04 crc kubenswrapper[4978]: I0225 07:32:04.697356 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3dc9f302852a6d3ef2a5aad62f7477dfd3bc1f6b9ed7b41c009191339daa40c" Feb 25 07:32:04 crc kubenswrapper[4978]: I0225 07:32:04.697475 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533412-25jfm" Feb 25 07:32:04 crc kubenswrapper[4978]: E0225 07:32:04.920422 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ee093ca_f815_46cd_bcaf_cb1470142811.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ee093ca_f815_46cd_bcaf_cb1470142811.slice/crio-f3dc9f302852a6d3ef2a5aad62f7477dfd3bc1f6b9ed7b41c009191339daa40c\": RecentStats: unable to find data in memory cache]" Feb 25 07:32:05 crc kubenswrapper[4978]: I0225 07:32:05.138301 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533406-hpwgx"] Feb 25 07:32:05 crc kubenswrapper[4978]: I0225 07:32:05.147814 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533406-hpwgx"] Feb 25 07:32:05 crc kubenswrapper[4978]: I0225 07:32:05.349931 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b116345-90ab-4ff1-a394-c50059ec64f9" path="/var/lib/kubelet/pods/6b116345-90ab-4ff1-a394-c50059ec64f9/volumes" Feb 25 07:32:22 crc kubenswrapper[4978]: I0225 07:32:22.150029 4978 scope.go:117] "RemoveContainer" containerID="035bceb27599e2dfaef5d6a3b5b2a5bd925498e17fda1d594b7f6929967d93a5" Feb 25 07:33:10 crc kubenswrapper[4978]: I0225 07:33:10.867111 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hjdrp"] Feb 25 07:33:10 crc kubenswrapper[4978]: E0225 07:33:10.868244 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ee093ca-f815-46cd-bcaf-cb1470142811" containerName="oc" Feb 25 07:33:10 crc kubenswrapper[4978]: I0225 07:33:10.868260 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ee093ca-f815-46cd-bcaf-cb1470142811" containerName="oc" Feb 25 07:33:10 crc kubenswrapper[4978]: I0225 07:33:10.868401 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ee093ca-f815-46cd-bcaf-cb1470142811" containerName="oc" Feb 25 07:33:10 crc kubenswrapper[4978]: I0225 07:33:10.870630 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:10 crc kubenswrapper[4978]: I0225 07:33:10.875052 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hjdrp"] Feb 25 07:33:10 crc kubenswrapper[4978]: I0225 07:33:10.919229 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcpm5\" (UniqueName: \"kubernetes.io/projected/5e386129-0cac-4f70-860f-70d9a3b26bdb-kube-api-access-vcpm5\") pod \"community-operators-hjdrp\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:10 crc kubenswrapper[4978]: I0225 07:33:10.919694 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-utilities\") pod \"community-operators-hjdrp\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:10 crc kubenswrapper[4978]: I0225 07:33:10.919721 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-catalog-content\") pod \"community-operators-hjdrp\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:11 crc kubenswrapper[4978]: I0225 07:33:11.021661 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-utilities\") pod \"community-operators-hjdrp\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:11 crc kubenswrapper[4978]: I0225 07:33:11.021719 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-catalog-content\") pod \"community-operators-hjdrp\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:11 crc kubenswrapper[4978]: I0225 07:33:11.021816 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcpm5\" (UniqueName: \"kubernetes.io/projected/5e386129-0cac-4f70-860f-70d9a3b26bdb-kube-api-access-vcpm5\") pod \"community-operators-hjdrp\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:11 crc kubenswrapper[4978]: I0225 07:33:11.022827 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-catalog-content\") pod \"community-operators-hjdrp\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:11 crc kubenswrapper[4978]: I0225 07:33:11.022919 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-utilities\") pod \"community-operators-hjdrp\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:11 crc kubenswrapper[4978]: I0225 07:33:11.061832 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcpm5\" (UniqueName: \"kubernetes.io/projected/5e386129-0cac-4f70-860f-70d9a3b26bdb-kube-api-access-vcpm5\") pod \"community-operators-hjdrp\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:11 crc kubenswrapper[4978]: I0225 07:33:11.218660 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:11 crc kubenswrapper[4978]: I0225 07:33:11.715041 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hjdrp"] Feb 25 07:33:11 crc kubenswrapper[4978]: W0225 07:33:11.719951 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e386129_0cac_4f70_860f_70d9a3b26bdb.slice/crio-0550ab5620aca2c32649de82a8b8f8a7acf534e95ff8c05d747d3b882cfa8cd7 WatchSource:0}: Error finding container 0550ab5620aca2c32649de82a8b8f8a7acf534e95ff8c05d747d3b882cfa8cd7: Status 404 returned error can't find the container with id 0550ab5620aca2c32649de82a8b8f8a7acf534e95ff8c05d747d3b882cfa8cd7 Feb 25 07:33:12 crc kubenswrapper[4978]: I0225 07:33:12.313920 4978 generic.go:334] "Generic (PLEG): container finished" podID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerID="76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c" exitCode=0 Feb 25 07:33:12 crc kubenswrapper[4978]: I0225 07:33:12.313992 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjdrp" event={"ID":"5e386129-0cac-4f70-860f-70d9a3b26bdb","Type":"ContainerDied","Data":"76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c"} Feb 25 07:33:12 crc kubenswrapper[4978]: I0225 07:33:12.314285 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjdrp" event={"ID":"5e386129-0cac-4f70-860f-70d9a3b26bdb","Type":"ContainerStarted","Data":"0550ab5620aca2c32649de82a8b8f8a7acf534e95ff8c05d747d3b882cfa8cd7"} Feb 25 07:33:12 crc kubenswrapper[4978]: I0225 07:33:12.317823 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 07:33:14 crc kubenswrapper[4978]: I0225 07:33:14.674391 4978 generic.go:334] "Generic (PLEG): container finished" podID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerID="1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668" exitCode=0 Feb 25 07:33:14 crc kubenswrapper[4978]: I0225 07:33:14.674523 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjdrp" event={"ID":"5e386129-0cac-4f70-860f-70d9a3b26bdb","Type":"ContainerDied","Data":"1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668"} Feb 25 07:33:15 crc kubenswrapper[4978]: I0225 07:33:15.686341 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjdrp" event={"ID":"5e386129-0cac-4f70-860f-70d9a3b26bdb","Type":"ContainerStarted","Data":"da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134"} Feb 25 07:33:15 crc kubenswrapper[4978]: I0225 07:33:15.717544 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hjdrp" podStartSLOduration=2.9818509779999998 podStartE2EDuration="5.71751496s" podCreationTimestamp="2026-02-25 07:33:10 +0000 UTC" firstStartedPulling="2026-02-25 07:33:12.31757761 +0000 UTC m=+2885.756834079" lastFinishedPulling="2026-02-25 07:33:15.053241592 +0000 UTC m=+2888.492498061" observedRunningTime="2026-02-25 07:33:15.710676039 +0000 UTC m=+2889.149932558" watchObservedRunningTime="2026-02-25 07:33:15.71751496 +0000 UTC m=+2889.156771459" Feb 25 07:33:16 crc kubenswrapper[4978]: I0225 07:33:16.540280 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:33:16 crc kubenswrapper[4978]: I0225 07:33:16.540784 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:33:21 crc kubenswrapper[4978]: I0225 07:33:21.220442 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:21 crc kubenswrapper[4978]: I0225 07:33:21.221127 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:21 crc kubenswrapper[4978]: I0225 07:33:21.307700 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:21 crc kubenswrapper[4978]: I0225 07:33:21.818671 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:21 crc kubenswrapper[4978]: I0225 07:33:21.926146 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hjdrp"] Feb 25 07:33:23 crc kubenswrapper[4978]: I0225 07:33:23.762941 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hjdrp" podUID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerName="registry-server" containerID="cri-o://da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134" gracePeriod=2 Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.275351 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.419200 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-catalog-content\") pod \"5e386129-0cac-4f70-860f-70d9a3b26bdb\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.419359 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcpm5\" (UniqueName: \"kubernetes.io/projected/5e386129-0cac-4f70-860f-70d9a3b26bdb-kube-api-access-vcpm5\") pod \"5e386129-0cac-4f70-860f-70d9a3b26bdb\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.419485 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-utilities\") pod \"5e386129-0cac-4f70-860f-70d9a3b26bdb\" (UID: \"5e386129-0cac-4f70-860f-70d9a3b26bdb\") " Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.420452 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-utilities" (OuterVolumeSpecName: "utilities") pod "5e386129-0cac-4f70-860f-70d9a3b26bdb" (UID: "5e386129-0cac-4f70-860f-70d9a3b26bdb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.425793 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e386129-0cac-4f70-860f-70d9a3b26bdb-kube-api-access-vcpm5" (OuterVolumeSpecName: "kube-api-access-vcpm5") pod "5e386129-0cac-4f70-860f-70d9a3b26bdb" (UID: "5e386129-0cac-4f70-860f-70d9a3b26bdb"). InnerVolumeSpecName "kube-api-access-vcpm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.522339 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcpm5\" (UniqueName: \"kubernetes.io/projected/5e386129-0cac-4f70-860f-70d9a3b26bdb-kube-api-access-vcpm5\") on node \"crc\" DevicePath \"\"" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.522393 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.586096 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e386129-0cac-4f70-860f-70d9a3b26bdb" (UID: "5e386129-0cac-4f70-860f-70d9a3b26bdb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.624053 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e386129-0cac-4f70-860f-70d9a3b26bdb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.778611 4978 generic.go:334] "Generic (PLEG): container finished" podID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerID="da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134" exitCode=0 Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.778681 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjdrp" event={"ID":"5e386129-0cac-4f70-860f-70d9a3b26bdb","Type":"ContainerDied","Data":"da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134"} Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.778721 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hjdrp" event={"ID":"5e386129-0cac-4f70-860f-70d9a3b26bdb","Type":"ContainerDied","Data":"0550ab5620aca2c32649de82a8b8f8a7acf534e95ff8c05d747d3b882cfa8cd7"} Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.778750 4978 scope.go:117] "RemoveContainer" containerID="da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.778998 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hjdrp" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.814480 4978 scope.go:117] "RemoveContainer" containerID="1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.831923 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hjdrp"] Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.846256 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hjdrp"] Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.864803 4978 scope.go:117] "RemoveContainer" containerID="76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.894831 4978 scope.go:117] "RemoveContainer" containerID="da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134" Feb 25 07:33:24 crc kubenswrapper[4978]: E0225 07:33:24.895299 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134\": container with ID starting with da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134 not found: ID does not exist" containerID="da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.895334 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134"} err="failed to get container status \"da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134\": rpc error: code = NotFound desc = could not find container \"da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134\": container with ID starting with da529b48e89811cb643fea4945eff73c0aef5a4c2cb63887003bc71135cfc134 not found: ID does not exist" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.895429 4978 scope.go:117] "RemoveContainer" containerID="1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668" Feb 25 07:33:24 crc kubenswrapper[4978]: E0225 07:33:24.895855 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668\": container with ID starting with 1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668 not found: ID does not exist" containerID="1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.895881 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668"} err="failed to get container status \"1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668\": rpc error: code = NotFound desc = could not find container \"1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668\": container with ID starting with 1f0e083787dd126e65896076dcd477ec625cebae7c7a1ab6e6bb6d41d0979668 not found: ID does not exist" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.895897 4978 scope.go:117] "RemoveContainer" containerID="76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c" Feb 25 07:33:24 crc kubenswrapper[4978]: E0225 07:33:24.896167 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c\": container with ID starting with 76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c not found: ID does not exist" containerID="76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c" Feb 25 07:33:24 crc kubenswrapper[4978]: I0225 07:33:24.896194 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c"} err="failed to get container status \"76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c\": rpc error: code = NotFound desc = could not find container \"76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c\": container with ID starting with 76577b8d88983687ddff689f6406484bce6860cbaf47e8545c951ab177d8672c not found: ID does not exist" Feb 25 07:33:25 crc kubenswrapper[4978]: I0225 07:33:25.340256 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e386129-0cac-4f70-860f-70d9a3b26bdb" path="/var/lib/kubelet/pods/5e386129-0cac-4f70-860f-70d9a3b26bdb/volumes" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.228544 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h5p9h"] Feb 25 07:33:40 crc kubenswrapper[4978]: E0225 07:33:40.230101 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerName="registry-server" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.230138 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerName="registry-server" Feb 25 07:33:40 crc kubenswrapper[4978]: E0225 07:33:40.230204 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerName="extract-utilities" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.230231 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerName="extract-utilities" Feb 25 07:33:40 crc kubenswrapper[4978]: E0225 07:33:40.230260 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerName="extract-content" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.230278 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerName="extract-content" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.231127 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e386129-0cac-4f70-860f-70d9a3b26bdb" containerName="registry-server" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.234447 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.245834 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5p9h"] Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.389086 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dmvf\" (UniqueName: \"kubernetes.io/projected/0fbf5944-e447-488d-9570-bac2572f463c-kube-api-access-4dmvf\") pod \"redhat-marketplace-h5p9h\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.390073 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-utilities\") pod \"redhat-marketplace-h5p9h\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.390331 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-catalog-content\") pod \"redhat-marketplace-h5p9h\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.492142 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-catalog-content\") pod \"redhat-marketplace-h5p9h\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.492702 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dmvf\" (UniqueName: \"kubernetes.io/projected/0fbf5944-e447-488d-9570-bac2572f463c-kube-api-access-4dmvf\") pod \"redhat-marketplace-h5p9h\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.493012 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-utilities\") pod \"redhat-marketplace-h5p9h\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.493008 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-catalog-content\") pod \"redhat-marketplace-h5p9h\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.493472 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-utilities\") pod \"redhat-marketplace-h5p9h\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.517281 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dmvf\" (UniqueName: \"kubernetes.io/projected/0fbf5944-e447-488d-9570-bac2572f463c-kube-api-access-4dmvf\") pod \"redhat-marketplace-h5p9h\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:40 crc kubenswrapper[4978]: I0225 07:33:40.559270 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:41 crc kubenswrapper[4978]: I0225 07:33:41.032776 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5p9h"] Feb 25 07:33:41 crc kubenswrapper[4978]: I0225 07:33:41.949661 4978 generic.go:334] "Generic (PLEG): container finished" podID="0fbf5944-e447-488d-9570-bac2572f463c" containerID="d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8" exitCode=0 Feb 25 07:33:41 crc kubenswrapper[4978]: I0225 07:33:41.949733 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p9h" event={"ID":"0fbf5944-e447-488d-9570-bac2572f463c","Type":"ContainerDied","Data":"d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8"} Feb 25 07:33:41 crc kubenswrapper[4978]: I0225 07:33:41.949775 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p9h" event={"ID":"0fbf5944-e447-488d-9570-bac2572f463c","Type":"ContainerStarted","Data":"a7995cf2f9d8388c439dfd972863fa7f3aa0d80c7205fb4f671abfbbf4c438e3"} Feb 25 07:33:42 crc kubenswrapper[4978]: I0225 07:33:42.957482 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p9h" event={"ID":"0fbf5944-e447-488d-9570-bac2572f463c","Type":"ContainerStarted","Data":"124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2"} Feb 25 07:33:43 crc kubenswrapper[4978]: I0225 07:33:43.968437 4978 generic.go:334] "Generic (PLEG): container finished" podID="0fbf5944-e447-488d-9570-bac2572f463c" containerID="124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2" exitCode=0 Feb 25 07:33:43 crc kubenswrapper[4978]: I0225 07:33:43.968576 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p9h" event={"ID":"0fbf5944-e447-488d-9570-bac2572f463c","Type":"ContainerDied","Data":"124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2"} Feb 25 07:33:44 crc kubenswrapper[4978]: I0225 07:33:44.983794 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p9h" event={"ID":"0fbf5944-e447-488d-9570-bac2572f463c","Type":"ContainerStarted","Data":"df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80"} Feb 25 07:33:45 crc kubenswrapper[4978]: I0225 07:33:45.025591 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h5p9h" podStartSLOduration=2.539244633 podStartE2EDuration="5.025563917s" podCreationTimestamp="2026-02-25 07:33:40 +0000 UTC" firstStartedPulling="2026-02-25 07:33:41.954361205 +0000 UTC m=+2915.393617704" lastFinishedPulling="2026-02-25 07:33:44.440680519 +0000 UTC m=+2917.879936988" observedRunningTime="2026-02-25 07:33:45.012612445 +0000 UTC m=+2918.451868934" watchObservedRunningTime="2026-02-25 07:33:45.025563917 +0000 UTC m=+2918.464820416" Feb 25 07:33:46 crc kubenswrapper[4978]: I0225 07:33:46.541026 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:33:46 crc kubenswrapper[4978]: I0225 07:33:46.541114 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:33:50 crc kubenswrapper[4978]: I0225 07:33:50.559759 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:50 crc kubenswrapper[4978]: I0225 07:33:50.560249 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:50 crc kubenswrapper[4978]: I0225 07:33:50.637097 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:51 crc kubenswrapper[4978]: I0225 07:33:51.127096 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:51 crc kubenswrapper[4978]: I0225 07:33:51.197632 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5p9h"] Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.073150 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h5p9h" podUID="0fbf5944-e447-488d-9570-bac2572f463c" containerName="registry-server" containerID="cri-o://df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80" gracePeriod=2 Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.497300 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.610057 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dmvf\" (UniqueName: \"kubernetes.io/projected/0fbf5944-e447-488d-9570-bac2572f463c-kube-api-access-4dmvf\") pod \"0fbf5944-e447-488d-9570-bac2572f463c\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.610137 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-catalog-content\") pod \"0fbf5944-e447-488d-9570-bac2572f463c\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.610313 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-utilities\") pod \"0fbf5944-e447-488d-9570-bac2572f463c\" (UID: \"0fbf5944-e447-488d-9570-bac2572f463c\") " Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.611279 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-utilities" (OuterVolumeSpecName: "utilities") pod "0fbf5944-e447-488d-9570-bac2572f463c" (UID: "0fbf5944-e447-488d-9570-bac2572f463c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.611741 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.616080 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fbf5944-e447-488d-9570-bac2572f463c-kube-api-access-4dmvf" (OuterVolumeSpecName: "kube-api-access-4dmvf") pod "0fbf5944-e447-488d-9570-bac2572f463c" (UID: "0fbf5944-e447-488d-9570-bac2572f463c"). InnerVolumeSpecName "kube-api-access-4dmvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.651877 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fbf5944-e447-488d-9570-bac2572f463c" (UID: "0fbf5944-e447-488d-9570-bac2572f463c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.713264 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dmvf\" (UniqueName: \"kubernetes.io/projected/0fbf5944-e447-488d-9570-bac2572f463c-kube-api-access-4dmvf\") on node \"crc\" DevicePath \"\"" Feb 25 07:33:53 crc kubenswrapper[4978]: I0225 07:33:53.713333 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fbf5944-e447-488d-9570-bac2572f463c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.085532 4978 generic.go:334] "Generic (PLEG): container finished" podID="0fbf5944-e447-488d-9570-bac2572f463c" containerID="df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80" exitCode=0 Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.085612 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p9h" event={"ID":"0fbf5944-e447-488d-9570-bac2572f463c","Type":"ContainerDied","Data":"df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80"} Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.085635 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h5p9h" Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.085676 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h5p9h" event={"ID":"0fbf5944-e447-488d-9570-bac2572f463c","Type":"ContainerDied","Data":"a7995cf2f9d8388c439dfd972863fa7f3aa0d80c7205fb4f671abfbbf4c438e3"} Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.085724 4978 scope.go:117] "RemoveContainer" containerID="df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80" Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.128844 4978 scope.go:117] "RemoveContainer" containerID="124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2" Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.147777 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5p9h"] Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.158145 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h5p9h"] Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.170154 4978 scope.go:117] "RemoveContainer" containerID="d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8" Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.208912 4978 scope.go:117] "RemoveContainer" containerID="df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80" Feb 25 07:33:54 crc kubenswrapper[4978]: E0225 07:33:54.209569 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80\": container with ID starting with df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80 not found: ID does not exist" containerID="df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80" Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.209618 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80"} err="failed to get container status \"df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80\": rpc error: code = NotFound desc = could not find container \"df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80\": container with ID starting with df0e4cce62b7fc87b150b63038bb9b7e9dcd64db86ee7107179a3eda0c7f6b80 not found: ID does not exist" Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.209649 4978 scope.go:117] "RemoveContainer" containerID="124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2" Feb 25 07:33:54 crc kubenswrapper[4978]: E0225 07:33:54.210582 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2\": container with ID starting with 124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2 not found: ID does not exist" containerID="124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2" Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.210648 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2"} err="failed to get container status \"124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2\": rpc error: code = NotFound desc = could not find container \"124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2\": container with ID starting with 124aa95a3e5ec01212948749c4cacb08878db5a286253825c981b099e12cf8c2 not found: ID does not exist" Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.210696 4978 scope.go:117] "RemoveContainer" containerID="d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8" Feb 25 07:33:54 crc kubenswrapper[4978]: E0225 07:33:54.211184 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8\": container with ID starting with d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8 not found: ID does not exist" containerID="d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8" Feb 25 07:33:54 crc kubenswrapper[4978]: I0225 07:33:54.211234 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8"} err="failed to get container status \"d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8\": rpc error: code = NotFound desc = could not find container \"d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8\": container with ID starting with d7a57fe3887ec7de6b30102a814e496f666d646c855c4864f409f7d19dfbdec8 not found: ID does not exist" Feb 25 07:33:55 crc kubenswrapper[4978]: I0225 07:33:55.344648 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fbf5944-e447-488d-9570-bac2572f463c" path="/var/lib/kubelet/pods/0fbf5944-e447-488d-9570-bac2572f463c/volumes" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.159689 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533414-trxd4"] Feb 25 07:34:00 crc kubenswrapper[4978]: E0225 07:34:00.160187 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbf5944-e447-488d-9570-bac2572f463c" containerName="registry-server" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.160209 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbf5944-e447-488d-9570-bac2572f463c" containerName="registry-server" Feb 25 07:34:00 crc kubenswrapper[4978]: E0225 07:34:00.160242 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbf5944-e447-488d-9570-bac2572f463c" containerName="extract-utilities" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.160254 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbf5944-e447-488d-9570-bac2572f463c" containerName="extract-utilities" Feb 25 07:34:00 crc kubenswrapper[4978]: E0225 07:34:00.160275 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fbf5944-e447-488d-9570-bac2572f463c" containerName="extract-content" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.160288 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fbf5944-e447-488d-9570-bac2572f463c" containerName="extract-content" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.160554 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fbf5944-e447-488d-9570-bac2572f463c" containerName="registry-server" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.161275 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533414-trxd4" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.163816 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.164151 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.165726 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.173989 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533414-trxd4"] Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.316590 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw9jc\" (UniqueName: \"kubernetes.io/projected/5be6de53-f11d-432f-a713-f5b236ecdd12-kube-api-access-qw9jc\") pod \"auto-csr-approver-29533414-trxd4\" (UID: \"5be6de53-f11d-432f-a713-f5b236ecdd12\") " pod="openshift-infra/auto-csr-approver-29533414-trxd4" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.419071 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw9jc\" (UniqueName: \"kubernetes.io/projected/5be6de53-f11d-432f-a713-f5b236ecdd12-kube-api-access-qw9jc\") pod \"auto-csr-approver-29533414-trxd4\" (UID: \"5be6de53-f11d-432f-a713-f5b236ecdd12\") " pod="openshift-infra/auto-csr-approver-29533414-trxd4" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.450391 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw9jc\" (UniqueName: \"kubernetes.io/projected/5be6de53-f11d-432f-a713-f5b236ecdd12-kube-api-access-qw9jc\") pod \"auto-csr-approver-29533414-trxd4\" (UID: \"5be6de53-f11d-432f-a713-f5b236ecdd12\") " pod="openshift-infra/auto-csr-approver-29533414-trxd4" Feb 25 07:34:00 crc kubenswrapper[4978]: I0225 07:34:00.484786 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533414-trxd4" Feb 25 07:34:01 crc kubenswrapper[4978]: I0225 07:34:01.041987 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533414-trxd4"] Feb 25 07:34:01 crc kubenswrapper[4978]: I0225 07:34:01.160543 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533414-trxd4" event={"ID":"5be6de53-f11d-432f-a713-f5b236ecdd12","Type":"ContainerStarted","Data":"b89a123c7b6f8dde7943a360bb6d672bdb550657dd7f94bc31456e3ecbedb7a5"} Feb 25 07:34:03 crc kubenswrapper[4978]: I0225 07:34:03.180105 4978 generic.go:334] "Generic (PLEG): container finished" podID="5be6de53-f11d-432f-a713-f5b236ecdd12" containerID="e5dce8b8fa2e1c087d67cc08b7da54fb4060e2a715d917e2d379845357467cf1" exitCode=0 Feb 25 07:34:03 crc kubenswrapper[4978]: I0225 07:34:03.180192 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533414-trxd4" event={"ID":"5be6de53-f11d-432f-a713-f5b236ecdd12","Type":"ContainerDied","Data":"e5dce8b8fa2e1c087d67cc08b7da54fb4060e2a715d917e2d379845357467cf1"} Feb 25 07:34:04 crc kubenswrapper[4978]: I0225 07:34:04.578793 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533414-trxd4" Feb 25 07:34:04 crc kubenswrapper[4978]: I0225 07:34:04.695685 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw9jc\" (UniqueName: \"kubernetes.io/projected/5be6de53-f11d-432f-a713-f5b236ecdd12-kube-api-access-qw9jc\") pod \"5be6de53-f11d-432f-a713-f5b236ecdd12\" (UID: \"5be6de53-f11d-432f-a713-f5b236ecdd12\") " Feb 25 07:34:04 crc kubenswrapper[4978]: I0225 07:34:04.703215 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5be6de53-f11d-432f-a713-f5b236ecdd12-kube-api-access-qw9jc" (OuterVolumeSpecName: "kube-api-access-qw9jc") pod "5be6de53-f11d-432f-a713-f5b236ecdd12" (UID: "5be6de53-f11d-432f-a713-f5b236ecdd12"). InnerVolumeSpecName "kube-api-access-qw9jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:34:04 crc kubenswrapper[4978]: I0225 07:34:04.797452 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw9jc\" (UniqueName: \"kubernetes.io/projected/5be6de53-f11d-432f-a713-f5b236ecdd12-kube-api-access-qw9jc\") on node \"crc\" DevicePath \"\"" Feb 25 07:34:05 crc kubenswrapper[4978]: I0225 07:34:05.204587 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533414-trxd4" event={"ID":"5be6de53-f11d-432f-a713-f5b236ecdd12","Type":"ContainerDied","Data":"b89a123c7b6f8dde7943a360bb6d672bdb550657dd7f94bc31456e3ecbedb7a5"} Feb 25 07:34:05 crc kubenswrapper[4978]: I0225 07:34:05.204647 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b89a123c7b6f8dde7943a360bb6d672bdb550657dd7f94bc31456e3ecbedb7a5" Feb 25 07:34:05 crc kubenswrapper[4978]: I0225 07:34:05.204674 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533414-trxd4" Feb 25 07:34:05 crc kubenswrapper[4978]: I0225 07:34:05.673196 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533408-lr9bq"] Feb 25 07:34:05 crc kubenswrapper[4978]: I0225 07:34:05.682850 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533408-lr9bq"] Feb 25 07:34:07 crc kubenswrapper[4978]: I0225 07:34:07.343313 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a072887d-45d9-429c-866e-71d1dd6d726c" path="/var/lib/kubelet/pods/a072887d-45d9-429c-866e-71d1dd6d726c/volumes" Feb 25 07:34:16 crc kubenswrapper[4978]: I0225 07:34:16.541040 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:34:16 crc kubenswrapper[4978]: I0225 07:34:16.542542 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:34:16 crc kubenswrapper[4978]: I0225 07:34:16.542623 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:34:16 crc kubenswrapper[4978]: I0225 07:34:16.543544 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:34:16 crc kubenswrapper[4978]: I0225 07:34:16.543637 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" gracePeriod=600 Feb 25 07:34:16 crc kubenswrapper[4978]: E0225 07:34:16.675631 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:34:17 crc kubenswrapper[4978]: I0225 07:34:17.351054 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" exitCode=0 Feb 25 07:34:17 crc kubenswrapper[4978]: I0225 07:34:17.351133 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754"} Feb 25 07:34:17 crc kubenswrapper[4978]: I0225 07:34:17.351202 4978 scope.go:117] "RemoveContainer" containerID="164f20c4dafdd6bdb0816971bff3c38ceeea3c62076bcdad2e26cfca1cfb00a2" Feb 25 07:34:17 crc kubenswrapper[4978]: I0225 07:34:17.352408 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:34:17 crc kubenswrapper[4978]: E0225 07:34:17.353103 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:34:22 crc kubenswrapper[4978]: I0225 07:34:22.304429 4978 scope.go:117] "RemoveContainer" containerID="355baa64dfffcefe6b7012720bc7447b6d9bf2ce88e72b6486e64dd219961ea2" Feb 25 07:34:31 crc kubenswrapper[4978]: I0225 07:34:31.327514 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:34:31 crc kubenswrapper[4978]: E0225 07:34:31.328642 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:34:45 crc kubenswrapper[4978]: I0225 07:34:45.327656 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:34:45 crc kubenswrapper[4978]: E0225 07:34:45.328458 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:35:00 crc kubenswrapper[4978]: I0225 07:35:00.327727 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:35:00 crc kubenswrapper[4978]: E0225 07:35:00.328548 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:35:13 crc kubenswrapper[4978]: I0225 07:35:13.506570 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:35:13 crc kubenswrapper[4978]: E0225 07:35:13.507805 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:35:25 crc kubenswrapper[4978]: I0225 07:35:25.328412 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:35:25 crc kubenswrapper[4978]: E0225 07:35:25.329524 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:35:40 crc kubenswrapper[4978]: I0225 07:35:40.328356 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:35:40 crc kubenswrapper[4978]: E0225 07:35:40.329478 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:35:53 crc kubenswrapper[4978]: I0225 07:35:53.328143 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:35:53 crc kubenswrapper[4978]: E0225 07:35:53.328910 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.151155 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533416-vbpj6"] Feb 25 07:36:00 crc kubenswrapper[4978]: E0225 07:36:00.151943 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5be6de53-f11d-432f-a713-f5b236ecdd12" containerName="oc" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.151965 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5be6de53-f11d-432f-a713-f5b236ecdd12" containerName="oc" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.152247 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5be6de53-f11d-432f-a713-f5b236ecdd12" containerName="oc" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.153156 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533416-vbpj6" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.155767 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.156142 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.155905 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.170545 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533416-vbpj6"] Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.218971 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz4wz\" (UniqueName: \"kubernetes.io/projected/f1806265-d5a0-4e54-bfe8-af444130563a-kube-api-access-mz4wz\") pod \"auto-csr-approver-29533416-vbpj6\" (UID: \"f1806265-d5a0-4e54-bfe8-af444130563a\") " pod="openshift-infra/auto-csr-approver-29533416-vbpj6" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.320232 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz4wz\" (UniqueName: \"kubernetes.io/projected/f1806265-d5a0-4e54-bfe8-af444130563a-kube-api-access-mz4wz\") pod \"auto-csr-approver-29533416-vbpj6\" (UID: \"f1806265-d5a0-4e54-bfe8-af444130563a\") " pod="openshift-infra/auto-csr-approver-29533416-vbpj6" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.353560 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz4wz\" (UniqueName: \"kubernetes.io/projected/f1806265-d5a0-4e54-bfe8-af444130563a-kube-api-access-mz4wz\") pod \"auto-csr-approver-29533416-vbpj6\" (UID: \"f1806265-d5a0-4e54-bfe8-af444130563a\") " pod="openshift-infra/auto-csr-approver-29533416-vbpj6" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.486021 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533416-vbpj6" Feb 25 07:36:00 crc kubenswrapper[4978]: I0225 07:36:00.996550 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533416-vbpj6"] Feb 25 07:36:01 crc kubenswrapper[4978]: I0225 07:36:01.343527 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533416-vbpj6" event={"ID":"f1806265-d5a0-4e54-bfe8-af444130563a","Type":"ContainerStarted","Data":"6908d26df26ce8cff6821c3889d9d2c5387a8dc81b58a99234904387ffc178e3"} Feb 25 07:36:03 crc kubenswrapper[4978]: I0225 07:36:03.355558 4978 generic.go:334] "Generic (PLEG): container finished" podID="f1806265-d5a0-4e54-bfe8-af444130563a" containerID="a4366b5ae1c491ae5c1705deb71299a71c119d286e9dff78a6f6b13f7042a1a7" exitCode=0 Feb 25 07:36:03 crc kubenswrapper[4978]: I0225 07:36:03.355640 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533416-vbpj6" event={"ID":"f1806265-d5a0-4e54-bfe8-af444130563a","Type":"ContainerDied","Data":"a4366b5ae1c491ae5c1705deb71299a71c119d286e9dff78a6f6b13f7042a1a7"} Feb 25 07:36:04 crc kubenswrapper[4978]: I0225 07:36:04.704558 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533416-vbpj6" Feb 25 07:36:04 crc kubenswrapper[4978]: I0225 07:36:04.794153 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mz4wz\" (UniqueName: \"kubernetes.io/projected/f1806265-d5a0-4e54-bfe8-af444130563a-kube-api-access-mz4wz\") pod \"f1806265-d5a0-4e54-bfe8-af444130563a\" (UID: \"f1806265-d5a0-4e54-bfe8-af444130563a\") " Feb 25 07:36:04 crc kubenswrapper[4978]: I0225 07:36:04.809706 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1806265-d5a0-4e54-bfe8-af444130563a-kube-api-access-mz4wz" (OuterVolumeSpecName: "kube-api-access-mz4wz") pod "f1806265-d5a0-4e54-bfe8-af444130563a" (UID: "f1806265-d5a0-4e54-bfe8-af444130563a"). InnerVolumeSpecName "kube-api-access-mz4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:36:04 crc kubenswrapper[4978]: I0225 07:36:04.896433 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mz4wz\" (UniqueName: \"kubernetes.io/projected/f1806265-d5a0-4e54-bfe8-af444130563a-kube-api-access-mz4wz\") on node \"crc\" DevicePath \"\"" Feb 25 07:36:05 crc kubenswrapper[4978]: I0225 07:36:05.327651 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:36:05 crc kubenswrapper[4978]: E0225 07:36:05.328063 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:36:05 crc kubenswrapper[4978]: I0225 07:36:05.375033 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533416-vbpj6" event={"ID":"f1806265-d5a0-4e54-bfe8-af444130563a","Type":"ContainerDied","Data":"6908d26df26ce8cff6821c3889d9d2c5387a8dc81b58a99234904387ffc178e3"} Feb 25 07:36:05 crc kubenswrapper[4978]: I0225 07:36:05.375073 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6908d26df26ce8cff6821c3889d9d2c5387a8dc81b58a99234904387ffc178e3" Feb 25 07:36:05 crc kubenswrapper[4978]: I0225 07:36:05.375448 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533416-vbpj6" Feb 25 07:36:05 crc kubenswrapper[4978]: I0225 07:36:05.807769 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533410-z6bsp"] Feb 25 07:36:05 crc kubenswrapper[4978]: I0225 07:36:05.818887 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533410-z6bsp"] Feb 25 07:36:07 crc kubenswrapper[4978]: I0225 07:36:07.340512 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28650a47-547e-4ce5-b222-7c8ca2b6ebc4" path="/var/lib/kubelet/pods/28650a47-547e-4ce5-b222-7c8ca2b6ebc4/volumes" Feb 25 07:36:17 crc kubenswrapper[4978]: I0225 07:36:17.333444 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:36:17 crc kubenswrapper[4978]: E0225 07:36:17.334307 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:36:22 crc kubenswrapper[4978]: I0225 07:36:22.497069 4978 scope.go:117] "RemoveContainer" containerID="277feb2be61c626c93abb9014ae4c52855d3167af06e514dca0f29d7df2f8703" Feb 25 07:36:32 crc kubenswrapper[4978]: I0225 07:36:32.327479 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:36:32 crc kubenswrapper[4978]: E0225 07:36:32.328575 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:36:45 crc kubenswrapper[4978]: I0225 07:36:45.329804 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:36:45 crc kubenswrapper[4978]: E0225 07:36:45.331182 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:37:00 crc kubenswrapper[4978]: I0225 07:37:00.328151 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:37:00 crc kubenswrapper[4978]: E0225 07:37:00.329243 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:37:11 crc kubenswrapper[4978]: I0225 07:37:11.328276 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:37:11 crc kubenswrapper[4978]: E0225 07:37:11.329290 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:37:23 crc kubenswrapper[4978]: I0225 07:37:23.328461 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:37:23 crc kubenswrapper[4978]: E0225 07:37:23.329332 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:37:36 crc kubenswrapper[4978]: I0225 07:37:36.327325 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:37:36 crc kubenswrapper[4978]: E0225 07:37:36.328308 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:37:50 crc kubenswrapper[4978]: I0225 07:37:50.327039 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:37:50 crc kubenswrapper[4978]: E0225 07:37:50.327831 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.195340 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533418-42npc"] Feb 25 07:38:00 crc kubenswrapper[4978]: E0225 07:38:00.197286 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1806265-d5a0-4e54-bfe8-af444130563a" containerName="oc" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.197419 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1806265-d5a0-4e54-bfe8-af444130563a" containerName="oc" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.197688 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1806265-d5a0-4e54-bfe8-af444130563a" containerName="oc" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.198302 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533418-42npc" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.204205 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.205345 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.208766 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533418-42npc"] Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.211575 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.338689 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqtz8\" (UniqueName: \"kubernetes.io/projected/325ca167-ab03-4703-b1f3-ce57e98dfdbb-kube-api-access-nqtz8\") pod \"auto-csr-approver-29533418-42npc\" (UID: \"325ca167-ab03-4703-b1f3-ce57e98dfdbb\") " pod="openshift-infra/auto-csr-approver-29533418-42npc" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.439984 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqtz8\" (UniqueName: \"kubernetes.io/projected/325ca167-ab03-4703-b1f3-ce57e98dfdbb-kube-api-access-nqtz8\") pod \"auto-csr-approver-29533418-42npc\" (UID: \"325ca167-ab03-4703-b1f3-ce57e98dfdbb\") " pod="openshift-infra/auto-csr-approver-29533418-42npc" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.460284 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqtz8\" (UniqueName: \"kubernetes.io/projected/325ca167-ab03-4703-b1f3-ce57e98dfdbb-kube-api-access-nqtz8\") pod \"auto-csr-approver-29533418-42npc\" (UID: \"325ca167-ab03-4703-b1f3-ce57e98dfdbb\") " pod="openshift-infra/auto-csr-approver-29533418-42npc" Feb 25 07:38:00 crc kubenswrapper[4978]: I0225 07:38:00.524974 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533418-42npc" Feb 25 07:38:01 crc kubenswrapper[4978]: I0225 07:38:01.021394 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533418-42npc"] Feb 25 07:38:01 crc kubenswrapper[4978]: W0225 07:38:01.023241 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod325ca167_ab03_4703_b1f3_ce57e98dfdbb.slice/crio-da61462b25e1cbace68e72dec164ce45f38af103c71b0d3c50118549cd7f2943 WatchSource:0}: Error finding container da61462b25e1cbace68e72dec164ce45f38af103c71b0d3c50118549cd7f2943: Status 404 returned error can't find the container with id da61462b25e1cbace68e72dec164ce45f38af103c71b0d3c50118549cd7f2943 Feb 25 07:38:01 crc kubenswrapper[4978]: I0225 07:38:01.538332 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533418-42npc" event={"ID":"325ca167-ab03-4703-b1f3-ce57e98dfdbb","Type":"ContainerStarted","Data":"da61462b25e1cbace68e72dec164ce45f38af103c71b0d3c50118549cd7f2943"} Feb 25 07:38:02 crc kubenswrapper[4978]: I0225 07:38:02.548625 4978 generic.go:334] "Generic (PLEG): container finished" podID="325ca167-ab03-4703-b1f3-ce57e98dfdbb" containerID="0c412716ec71b35a9ef60e3ca82a45fef3671679b2456fbc4c16354d5858a5d1" exitCode=0 Feb 25 07:38:02 crc kubenswrapper[4978]: I0225 07:38:02.548694 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533418-42npc" event={"ID":"325ca167-ab03-4703-b1f3-ce57e98dfdbb","Type":"ContainerDied","Data":"0c412716ec71b35a9ef60e3ca82a45fef3671679b2456fbc4c16354d5858a5d1"} Feb 25 07:38:03 crc kubenswrapper[4978]: I0225 07:38:03.956018 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533418-42npc" Feb 25 07:38:04 crc kubenswrapper[4978]: I0225 07:38:04.113082 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqtz8\" (UniqueName: \"kubernetes.io/projected/325ca167-ab03-4703-b1f3-ce57e98dfdbb-kube-api-access-nqtz8\") pod \"325ca167-ab03-4703-b1f3-ce57e98dfdbb\" (UID: \"325ca167-ab03-4703-b1f3-ce57e98dfdbb\") " Feb 25 07:38:04 crc kubenswrapper[4978]: I0225 07:38:04.126804 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/325ca167-ab03-4703-b1f3-ce57e98dfdbb-kube-api-access-nqtz8" (OuterVolumeSpecName: "kube-api-access-nqtz8") pod "325ca167-ab03-4703-b1f3-ce57e98dfdbb" (UID: "325ca167-ab03-4703-b1f3-ce57e98dfdbb"). InnerVolumeSpecName "kube-api-access-nqtz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:38:04 crc kubenswrapper[4978]: I0225 07:38:04.216530 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqtz8\" (UniqueName: \"kubernetes.io/projected/325ca167-ab03-4703-b1f3-ce57e98dfdbb-kube-api-access-nqtz8\") on node \"crc\" DevicePath \"\"" Feb 25 07:38:04 crc kubenswrapper[4978]: I0225 07:38:04.327846 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:38:04 crc kubenswrapper[4978]: E0225 07:38:04.328216 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:38:04 crc kubenswrapper[4978]: I0225 07:38:04.571828 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533418-42npc" event={"ID":"325ca167-ab03-4703-b1f3-ce57e98dfdbb","Type":"ContainerDied","Data":"da61462b25e1cbace68e72dec164ce45f38af103c71b0d3c50118549cd7f2943"} Feb 25 07:38:04 crc kubenswrapper[4978]: I0225 07:38:04.571889 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da61462b25e1cbace68e72dec164ce45f38af103c71b0d3c50118549cd7f2943" Feb 25 07:38:04 crc kubenswrapper[4978]: I0225 07:38:04.571893 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533418-42npc" Feb 25 07:38:05 crc kubenswrapper[4978]: I0225 07:38:05.057210 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533412-25jfm"] Feb 25 07:38:05 crc kubenswrapper[4978]: I0225 07:38:05.064251 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533412-25jfm"] Feb 25 07:38:05 crc kubenswrapper[4978]: I0225 07:38:05.336606 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ee093ca-f815-46cd-bcaf-cb1470142811" path="/var/lib/kubelet/pods/9ee093ca-f815-46cd-bcaf-cb1470142811/volumes" Feb 25 07:38:16 crc kubenswrapper[4978]: I0225 07:38:16.327621 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:38:16 crc kubenswrapper[4978]: E0225 07:38:16.328700 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:38:22 crc kubenswrapper[4978]: I0225 07:38:22.630185 4978 scope.go:117] "RemoveContainer" containerID="d67488c746b4786725fc9d7a030eedec62857adb971ed0e547d882635e7ff714" Feb 25 07:38:27 crc kubenswrapper[4978]: I0225 07:38:27.335046 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:38:27 crc kubenswrapper[4978]: E0225 07:38:27.335847 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:38:41 crc kubenswrapper[4978]: I0225 07:38:41.328690 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:38:41 crc kubenswrapper[4978]: E0225 07:38:41.329709 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:38:55 crc kubenswrapper[4978]: I0225 07:38:55.328683 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:38:55 crc kubenswrapper[4978]: E0225 07:38:55.330013 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:39:08 crc kubenswrapper[4978]: I0225 07:39:08.328079 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:39:08 crc kubenswrapper[4978]: E0225 07:39:08.328957 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:39:23 crc kubenswrapper[4978]: I0225 07:39:23.328326 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:39:24 crc kubenswrapper[4978]: I0225 07:39:24.342945 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"4e0615218c5cf39b6b1cde753ee4d72d39432f0b0ffecbd6b54f3614eaa3d0ab"} Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.168217 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533420-zwpxs"] Feb 25 07:40:00 crc kubenswrapper[4978]: E0225 07:40:00.170121 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="325ca167-ab03-4703-b1f3-ce57e98dfdbb" containerName="oc" Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.170174 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="325ca167-ab03-4703-b1f3-ce57e98dfdbb" containerName="oc" Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.171103 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="325ca167-ab03-4703-b1f3-ce57e98dfdbb" containerName="oc" Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.172020 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533420-zwpxs" Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.178981 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.180307 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.180898 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.194628 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533420-zwpxs"] Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.195710 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4kqx\" (UniqueName: \"kubernetes.io/projected/4b9eb274-2592-4593-bb61-4387f8416f36-kube-api-access-q4kqx\") pod \"auto-csr-approver-29533420-zwpxs\" (UID: \"4b9eb274-2592-4593-bb61-4387f8416f36\") " pod="openshift-infra/auto-csr-approver-29533420-zwpxs" Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.296938 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4kqx\" (UniqueName: \"kubernetes.io/projected/4b9eb274-2592-4593-bb61-4387f8416f36-kube-api-access-q4kqx\") pod \"auto-csr-approver-29533420-zwpxs\" (UID: \"4b9eb274-2592-4593-bb61-4387f8416f36\") " pod="openshift-infra/auto-csr-approver-29533420-zwpxs" Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.326841 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4kqx\" (UniqueName: \"kubernetes.io/projected/4b9eb274-2592-4593-bb61-4387f8416f36-kube-api-access-q4kqx\") pod \"auto-csr-approver-29533420-zwpxs\" (UID: \"4b9eb274-2592-4593-bb61-4387f8416f36\") " pod="openshift-infra/auto-csr-approver-29533420-zwpxs" Feb 25 07:40:00 crc kubenswrapper[4978]: I0225 07:40:00.507244 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533420-zwpxs" Feb 25 07:40:01 crc kubenswrapper[4978]: I0225 07:40:01.068759 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533420-zwpxs"] Feb 25 07:40:01 crc kubenswrapper[4978]: I0225 07:40:01.080129 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 07:40:01 crc kubenswrapper[4978]: I0225 07:40:01.687331 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533420-zwpxs" event={"ID":"4b9eb274-2592-4593-bb61-4387f8416f36","Type":"ContainerStarted","Data":"10a7de6184f24c6382b83e8a67147ab2e020dfd33003d5d626cbc2c056fa8041"} Feb 25 07:40:03 crc kubenswrapper[4978]: I0225 07:40:03.705004 4978 generic.go:334] "Generic (PLEG): container finished" podID="4b9eb274-2592-4593-bb61-4387f8416f36" containerID="b523bc6d2b5aaa6355d9541669197aae9f484b45984acc20e5a8b7800b13359c" exitCode=0 Feb 25 07:40:03 crc kubenswrapper[4978]: I0225 07:40:03.705115 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533420-zwpxs" event={"ID":"4b9eb274-2592-4593-bb61-4387f8416f36","Type":"ContainerDied","Data":"b523bc6d2b5aaa6355d9541669197aae9f484b45984acc20e5a8b7800b13359c"} Feb 25 07:40:05 crc kubenswrapper[4978]: I0225 07:40:05.070813 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533420-zwpxs" Feb 25 07:40:05 crc kubenswrapper[4978]: I0225 07:40:05.075032 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4kqx\" (UniqueName: \"kubernetes.io/projected/4b9eb274-2592-4593-bb61-4387f8416f36-kube-api-access-q4kqx\") pod \"4b9eb274-2592-4593-bb61-4387f8416f36\" (UID: \"4b9eb274-2592-4593-bb61-4387f8416f36\") " Feb 25 07:40:05 crc kubenswrapper[4978]: I0225 07:40:05.082242 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b9eb274-2592-4593-bb61-4387f8416f36-kube-api-access-q4kqx" (OuterVolumeSpecName: "kube-api-access-q4kqx") pod "4b9eb274-2592-4593-bb61-4387f8416f36" (UID: "4b9eb274-2592-4593-bb61-4387f8416f36"). InnerVolumeSpecName "kube-api-access-q4kqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:40:05 crc kubenswrapper[4978]: I0225 07:40:05.176222 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4kqx\" (UniqueName: \"kubernetes.io/projected/4b9eb274-2592-4593-bb61-4387f8416f36-kube-api-access-q4kqx\") on node \"crc\" DevicePath \"\"" Feb 25 07:40:05 crc kubenswrapper[4978]: I0225 07:40:05.725178 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533420-zwpxs" event={"ID":"4b9eb274-2592-4593-bb61-4387f8416f36","Type":"ContainerDied","Data":"10a7de6184f24c6382b83e8a67147ab2e020dfd33003d5d626cbc2c056fa8041"} Feb 25 07:40:05 crc kubenswrapper[4978]: I0225 07:40:05.725225 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10a7de6184f24c6382b83e8a67147ab2e020dfd33003d5d626cbc2c056fa8041" Feb 25 07:40:05 crc kubenswrapper[4978]: I0225 07:40:05.725255 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533420-zwpxs" Feb 25 07:40:06 crc kubenswrapper[4978]: I0225 07:40:06.163450 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533414-trxd4"] Feb 25 07:40:06 crc kubenswrapper[4978]: I0225 07:40:06.175322 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533414-trxd4"] Feb 25 07:40:07 crc kubenswrapper[4978]: I0225 07:40:07.344448 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5be6de53-f11d-432f-a713-f5b236ecdd12" path="/var/lib/kubelet/pods/5be6de53-f11d-432f-a713-f5b236ecdd12/volumes" Feb 25 07:40:22 crc kubenswrapper[4978]: I0225 07:40:22.745955 4978 scope.go:117] "RemoveContainer" containerID="e5dce8b8fa2e1c087d67cc08b7da54fb4060e2a715d917e2d379845357467cf1" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.224868 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4d29q"] Feb 25 07:41:25 crc kubenswrapper[4978]: E0225 07:41:25.229448 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b9eb274-2592-4593-bb61-4387f8416f36" containerName="oc" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.229636 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b9eb274-2592-4593-bb61-4387f8416f36" containerName="oc" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.230031 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b9eb274-2592-4593-bb61-4387f8416f36" containerName="oc" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.232289 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.245723 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4d29q"] Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.286153 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-catalog-content\") pod \"certified-operators-4d29q\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.286610 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv2pb\" (UniqueName: \"kubernetes.io/projected/4cb77d13-cade-4a62-bf2a-d938ed042cc8-kube-api-access-sv2pb\") pod \"certified-operators-4d29q\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.286653 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-utilities\") pod \"certified-operators-4d29q\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.387895 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv2pb\" (UniqueName: \"kubernetes.io/projected/4cb77d13-cade-4a62-bf2a-d938ed042cc8-kube-api-access-sv2pb\") pod \"certified-operators-4d29q\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.388005 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-utilities\") pod \"certified-operators-4d29q\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.388251 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-catalog-content\") pod \"certified-operators-4d29q\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.390073 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-utilities\") pod \"certified-operators-4d29q\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.390209 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-catalog-content\") pod \"certified-operators-4d29q\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.423747 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv2pb\" (UniqueName: \"kubernetes.io/projected/4cb77d13-cade-4a62-bf2a-d938ed042cc8-kube-api-access-sv2pb\") pod \"certified-operators-4d29q\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:25 crc kubenswrapper[4978]: I0225 07:41:25.569764 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:26 crc kubenswrapper[4978]: I0225 07:41:26.075210 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4d29q"] Feb 25 07:41:26 crc kubenswrapper[4978]: I0225 07:41:26.480741 4978 generic.go:334] "Generic (PLEG): container finished" podID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerID="7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b" exitCode=0 Feb 25 07:41:26 crc kubenswrapper[4978]: I0225 07:41:26.481020 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d29q" event={"ID":"4cb77d13-cade-4a62-bf2a-d938ed042cc8","Type":"ContainerDied","Data":"7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b"} Feb 25 07:41:26 crc kubenswrapper[4978]: I0225 07:41:26.482165 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d29q" event={"ID":"4cb77d13-cade-4a62-bf2a-d938ed042cc8","Type":"ContainerStarted","Data":"082a8c09062c1890f20014d79352e18a8629df145683373913f9baa33ca831e5"} Feb 25 07:41:28 crc kubenswrapper[4978]: I0225 07:41:28.506421 4978 generic.go:334] "Generic (PLEG): container finished" podID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerID="23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2" exitCode=0 Feb 25 07:41:28 crc kubenswrapper[4978]: I0225 07:41:28.506484 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d29q" event={"ID":"4cb77d13-cade-4a62-bf2a-d938ed042cc8","Type":"ContainerDied","Data":"23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2"} Feb 25 07:41:29 crc kubenswrapper[4978]: I0225 07:41:29.515255 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d29q" event={"ID":"4cb77d13-cade-4a62-bf2a-d938ed042cc8","Type":"ContainerStarted","Data":"9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a"} Feb 25 07:41:29 crc kubenswrapper[4978]: I0225 07:41:29.538961 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4d29q" podStartSLOduration=2.082836847 podStartE2EDuration="4.538947476s" podCreationTimestamp="2026-02-25 07:41:25 +0000 UTC" firstStartedPulling="2026-02-25 07:41:26.483898148 +0000 UTC m=+3379.923154617" lastFinishedPulling="2026-02-25 07:41:28.940008777 +0000 UTC m=+3382.379265246" observedRunningTime="2026-02-25 07:41:29.535294322 +0000 UTC m=+3382.974550771" watchObservedRunningTime="2026-02-25 07:41:29.538947476 +0000 UTC m=+3382.978203935" Feb 25 07:41:35 crc kubenswrapper[4978]: I0225 07:41:35.570850 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:35 crc kubenswrapper[4978]: I0225 07:41:35.571200 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:35 crc kubenswrapper[4978]: I0225 07:41:35.639673 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:36 crc kubenswrapper[4978]: I0225 07:41:36.659733 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:36 crc kubenswrapper[4978]: I0225 07:41:36.733502 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4d29q"] Feb 25 07:41:38 crc kubenswrapper[4978]: I0225 07:41:38.607707 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4d29q" podUID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerName="registry-server" containerID="cri-o://9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a" gracePeriod=2 Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.113587 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.242583 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-catalog-content\") pod \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.242681 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv2pb\" (UniqueName: \"kubernetes.io/projected/4cb77d13-cade-4a62-bf2a-d938ed042cc8-kube-api-access-sv2pb\") pod \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.242755 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-utilities\") pod \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\" (UID: \"4cb77d13-cade-4a62-bf2a-d938ed042cc8\") " Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.243608 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-utilities" (OuterVolumeSpecName: "utilities") pod "4cb77d13-cade-4a62-bf2a-d938ed042cc8" (UID: "4cb77d13-cade-4a62-bf2a-d938ed042cc8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.248255 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cb77d13-cade-4a62-bf2a-d938ed042cc8-kube-api-access-sv2pb" (OuterVolumeSpecName: "kube-api-access-sv2pb") pod "4cb77d13-cade-4a62-bf2a-d938ed042cc8" (UID: "4cb77d13-cade-4a62-bf2a-d938ed042cc8"). InnerVolumeSpecName "kube-api-access-sv2pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.311858 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cb77d13-cade-4a62-bf2a-d938ed042cc8" (UID: "4cb77d13-cade-4a62-bf2a-d938ed042cc8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.344227 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.344448 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cb77d13-cade-4a62-bf2a-d938ed042cc8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.344547 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv2pb\" (UniqueName: \"kubernetes.io/projected/4cb77d13-cade-4a62-bf2a-d938ed042cc8-kube-api-access-sv2pb\") on node \"crc\" DevicePath \"\"" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.621092 4978 generic.go:334] "Generic (PLEG): container finished" podID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerID="9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a" exitCode=0 Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.621160 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d29q" event={"ID":"4cb77d13-cade-4a62-bf2a-d938ed042cc8","Type":"ContainerDied","Data":"9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a"} Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.621212 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4d29q" event={"ID":"4cb77d13-cade-4a62-bf2a-d938ed042cc8","Type":"ContainerDied","Data":"082a8c09062c1890f20014d79352e18a8629df145683373913f9baa33ca831e5"} Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.621241 4978 scope.go:117] "RemoveContainer" containerID="9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.621279 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4d29q" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.664767 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4d29q"] Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.665034 4978 scope.go:117] "RemoveContainer" containerID="23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.672265 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4d29q"] Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.721275 4978 scope.go:117] "RemoveContainer" containerID="7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.749286 4978 scope.go:117] "RemoveContainer" containerID="9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a" Feb 25 07:41:39 crc kubenswrapper[4978]: E0225 07:41:39.749941 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a\": container with ID starting with 9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a not found: ID does not exist" containerID="9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.749979 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a"} err="failed to get container status \"9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a\": rpc error: code = NotFound desc = could not find container \"9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a\": container with ID starting with 9b2794c0330c177c0a0841a914f5647d23b53cf7b8ea5c706bdd7d22ea5d450a not found: ID does not exist" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.750003 4978 scope.go:117] "RemoveContainer" containerID="23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2" Feb 25 07:41:39 crc kubenswrapper[4978]: E0225 07:41:39.750445 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2\": container with ID starting with 23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2 not found: ID does not exist" containerID="23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.750475 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2"} err="failed to get container status \"23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2\": rpc error: code = NotFound desc = could not find container \"23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2\": container with ID starting with 23139c03738f3a3dc9b866c677960dfba8de797f6e3c9b1863015557d8365ba2 not found: ID does not exist" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.750493 4978 scope.go:117] "RemoveContainer" containerID="7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b" Feb 25 07:41:39 crc kubenswrapper[4978]: E0225 07:41:39.751600 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b\": container with ID starting with 7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b not found: ID does not exist" containerID="7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b" Feb 25 07:41:39 crc kubenswrapper[4978]: I0225 07:41:39.751671 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b"} err="failed to get container status \"7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b\": rpc error: code = NotFound desc = could not find container \"7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b\": container with ID starting with 7043dee024ba458b53e5d4e9c486929c17bacd4db3246b24313662a476be5d2b not found: ID does not exist" Feb 25 07:41:41 crc kubenswrapper[4978]: I0225 07:41:41.342346 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" path="/var/lib/kubelet/pods/4cb77d13-cade-4a62-bf2a-d938ed042cc8/volumes" Feb 25 07:41:46 crc kubenswrapper[4978]: I0225 07:41:46.541131 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:41:46 crc kubenswrapper[4978]: I0225 07:41:46.541221 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.148931 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533422-t27jr"] Feb 25 07:42:00 crc kubenswrapper[4978]: E0225 07:42:00.149772 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerName="extract-utilities" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.149786 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerName="extract-utilities" Feb 25 07:42:00 crc kubenswrapper[4978]: E0225 07:42:00.149796 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerName="extract-content" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.149802 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerName="extract-content" Feb 25 07:42:00 crc kubenswrapper[4978]: E0225 07:42:00.149832 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerName="registry-server" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.149838 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerName="registry-server" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.149987 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cb77d13-cade-4a62-bf2a-d938ed042cc8" containerName="registry-server" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.150458 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533422-t27jr" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.152587 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.152815 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.154027 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.174658 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533422-t27jr"] Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.308331 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmgqc\" (UniqueName: \"kubernetes.io/projected/32cc4832-7c7c-40ae-9aac-30af12e9af4c-kube-api-access-bmgqc\") pod \"auto-csr-approver-29533422-t27jr\" (UID: \"32cc4832-7c7c-40ae-9aac-30af12e9af4c\") " pod="openshift-infra/auto-csr-approver-29533422-t27jr" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.410948 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmgqc\" (UniqueName: \"kubernetes.io/projected/32cc4832-7c7c-40ae-9aac-30af12e9af4c-kube-api-access-bmgqc\") pod \"auto-csr-approver-29533422-t27jr\" (UID: \"32cc4832-7c7c-40ae-9aac-30af12e9af4c\") " pod="openshift-infra/auto-csr-approver-29533422-t27jr" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.437789 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmgqc\" (UniqueName: \"kubernetes.io/projected/32cc4832-7c7c-40ae-9aac-30af12e9af4c-kube-api-access-bmgqc\") pod \"auto-csr-approver-29533422-t27jr\" (UID: \"32cc4832-7c7c-40ae-9aac-30af12e9af4c\") " pod="openshift-infra/auto-csr-approver-29533422-t27jr" Feb 25 07:42:00 crc kubenswrapper[4978]: I0225 07:42:00.471760 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533422-t27jr" Feb 25 07:42:01 crc kubenswrapper[4978]: I0225 07:42:01.008027 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533422-t27jr"] Feb 25 07:42:01 crc kubenswrapper[4978]: W0225 07:42:01.014774 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32cc4832_7c7c_40ae_9aac_30af12e9af4c.slice/crio-257fab13c3abae144481915584ea04a6eb13f99842b63fb43c9c86ec467d8189 WatchSource:0}: Error finding container 257fab13c3abae144481915584ea04a6eb13f99842b63fb43c9c86ec467d8189: Status 404 returned error can't find the container with id 257fab13c3abae144481915584ea04a6eb13f99842b63fb43c9c86ec467d8189 Feb 25 07:42:01 crc kubenswrapper[4978]: I0225 07:42:01.837814 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533422-t27jr" event={"ID":"32cc4832-7c7c-40ae-9aac-30af12e9af4c","Type":"ContainerStarted","Data":"257fab13c3abae144481915584ea04a6eb13f99842b63fb43c9c86ec467d8189"} Feb 25 07:42:02 crc kubenswrapper[4978]: I0225 07:42:02.868020 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533422-t27jr" event={"ID":"32cc4832-7c7c-40ae-9aac-30af12e9af4c","Type":"ContainerStarted","Data":"7080351c1386cb22043a3304b4041d234d1e7da50e9e1f049168bda9fdd48fb1"} Feb 25 07:42:02 crc kubenswrapper[4978]: I0225 07:42:02.889600 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533422-t27jr" podStartSLOduration=1.532921547 podStartE2EDuration="2.889565939s" podCreationTimestamp="2026-02-25 07:42:00 +0000 UTC" firstStartedPulling="2026-02-25 07:42:01.027496525 +0000 UTC m=+3414.466753034" lastFinishedPulling="2026-02-25 07:42:02.384140937 +0000 UTC m=+3415.823397426" observedRunningTime="2026-02-25 07:42:02.882642484 +0000 UTC m=+3416.321898983" watchObservedRunningTime="2026-02-25 07:42:02.889565939 +0000 UTC m=+3416.328822428" Feb 25 07:42:03 crc kubenswrapper[4978]: I0225 07:42:03.879597 4978 generic.go:334] "Generic (PLEG): container finished" podID="32cc4832-7c7c-40ae-9aac-30af12e9af4c" containerID="7080351c1386cb22043a3304b4041d234d1e7da50e9e1f049168bda9fdd48fb1" exitCode=0 Feb 25 07:42:03 crc kubenswrapper[4978]: I0225 07:42:03.879651 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533422-t27jr" event={"ID":"32cc4832-7c7c-40ae-9aac-30af12e9af4c","Type":"ContainerDied","Data":"7080351c1386cb22043a3304b4041d234d1e7da50e9e1f049168bda9fdd48fb1"} Feb 25 07:42:05 crc kubenswrapper[4978]: I0225 07:42:05.233323 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533422-t27jr" Feb 25 07:42:05 crc kubenswrapper[4978]: I0225 07:42:05.394459 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmgqc\" (UniqueName: \"kubernetes.io/projected/32cc4832-7c7c-40ae-9aac-30af12e9af4c-kube-api-access-bmgqc\") pod \"32cc4832-7c7c-40ae-9aac-30af12e9af4c\" (UID: \"32cc4832-7c7c-40ae-9aac-30af12e9af4c\") " Feb 25 07:42:05 crc kubenswrapper[4978]: I0225 07:42:05.402407 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32cc4832-7c7c-40ae-9aac-30af12e9af4c-kube-api-access-bmgqc" (OuterVolumeSpecName: "kube-api-access-bmgqc") pod "32cc4832-7c7c-40ae-9aac-30af12e9af4c" (UID: "32cc4832-7c7c-40ae-9aac-30af12e9af4c"). InnerVolumeSpecName "kube-api-access-bmgqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:42:05 crc kubenswrapper[4978]: I0225 07:42:05.498313 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmgqc\" (UniqueName: \"kubernetes.io/projected/32cc4832-7c7c-40ae-9aac-30af12e9af4c-kube-api-access-bmgqc\") on node \"crc\" DevicePath \"\"" Feb 25 07:42:05 crc kubenswrapper[4978]: I0225 07:42:05.901881 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533422-t27jr" event={"ID":"32cc4832-7c7c-40ae-9aac-30af12e9af4c","Type":"ContainerDied","Data":"257fab13c3abae144481915584ea04a6eb13f99842b63fb43c9c86ec467d8189"} Feb 25 07:42:05 crc kubenswrapper[4978]: I0225 07:42:05.901939 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="257fab13c3abae144481915584ea04a6eb13f99842b63fb43c9c86ec467d8189" Feb 25 07:42:05 crc kubenswrapper[4978]: I0225 07:42:05.901973 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533422-t27jr" Feb 25 07:42:05 crc kubenswrapper[4978]: I0225 07:42:05.975310 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533416-vbpj6"] Feb 25 07:42:05 crc kubenswrapper[4978]: I0225 07:42:05.985880 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533416-vbpj6"] Feb 25 07:42:07 crc kubenswrapper[4978]: I0225 07:42:07.338452 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1806265-d5a0-4e54-bfe8-af444130563a" path="/var/lib/kubelet/pods/f1806265-d5a0-4e54-bfe8-af444130563a/volumes" Feb 25 07:42:16 crc kubenswrapper[4978]: I0225 07:42:16.540919 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:42:16 crc kubenswrapper[4978]: I0225 07:42:16.541005 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:42:22 crc kubenswrapper[4978]: I0225 07:42:22.872957 4978 scope.go:117] "RemoveContainer" containerID="a4366b5ae1c491ae5c1705deb71299a71c119d286e9dff78a6f6b13f7042a1a7" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.032197 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6l4jw"] Feb 25 07:42:29 crc kubenswrapper[4978]: E0225 07:42:29.033584 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32cc4832-7c7c-40ae-9aac-30af12e9af4c" containerName="oc" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.033617 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="32cc4832-7c7c-40ae-9aac-30af12e9af4c" containerName="oc" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.033998 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="32cc4832-7c7c-40ae-9aac-30af12e9af4c" containerName="oc" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.041598 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.047033 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6l4jw"] Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.123762 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpnm7\" (UniqueName: \"kubernetes.io/projected/63791838-0b10-4bbc-9e51-33bd195e3f8c-kube-api-access-mpnm7\") pod \"redhat-operators-6l4jw\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.124023 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-utilities\") pod \"redhat-operators-6l4jw\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.124082 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-catalog-content\") pod \"redhat-operators-6l4jw\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.224933 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpnm7\" (UniqueName: \"kubernetes.io/projected/63791838-0b10-4bbc-9e51-33bd195e3f8c-kube-api-access-mpnm7\") pod \"redhat-operators-6l4jw\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.225251 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-utilities\") pod \"redhat-operators-6l4jw\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.225481 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-catalog-content\") pod \"redhat-operators-6l4jw\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.225847 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-utilities\") pod \"redhat-operators-6l4jw\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.226098 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-catalog-content\") pod \"redhat-operators-6l4jw\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.247660 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpnm7\" (UniqueName: \"kubernetes.io/projected/63791838-0b10-4bbc-9e51-33bd195e3f8c-kube-api-access-mpnm7\") pod \"redhat-operators-6l4jw\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.378550 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:29 crc kubenswrapper[4978]: I0225 07:42:29.614734 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6l4jw"] Feb 25 07:42:30 crc kubenswrapper[4978]: I0225 07:42:30.129226 4978 generic.go:334] "Generic (PLEG): container finished" podID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerID="c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f" exitCode=0 Feb 25 07:42:30 crc kubenswrapper[4978]: I0225 07:42:30.129289 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l4jw" event={"ID":"63791838-0b10-4bbc-9e51-33bd195e3f8c","Type":"ContainerDied","Data":"c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f"} Feb 25 07:42:30 crc kubenswrapper[4978]: I0225 07:42:30.129328 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l4jw" event={"ID":"63791838-0b10-4bbc-9e51-33bd195e3f8c","Type":"ContainerStarted","Data":"cccd9586d48c46975f61c6353a06d82637aa393b2d1ccd537c6d38586cd77a3c"} Feb 25 07:42:31 crc kubenswrapper[4978]: I0225 07:42:31.141646 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l4jw" event={"ID":"63791838-0b10-4bbc-9e51-33bd195e3f8c","Type":"ContainerStarted","Data":"4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9"} Feb 25 07:42:32 crc kubenswrapper[4978]: I0225 07:42:32.155299 4978 generic.go:334] "Generic (PLEG): container finished" podID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerID="4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9" exitCode=0 Feb 25 07:42:32 crc kubenswrapper[4978]: I0225 07:42:32.155393 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l4jw" event={"ID":"63791838-0b10-4bbc-9e51-33bd195e3f8c","Type":"ContainerDied","Data":"4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9"} Feb 25 07:42:33 crc kubenswrapper[4978]: I0225 07:42:33.167030 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l4jw" event={"ID":"63791838-0b10-4bbc-9e51-33bd195e3f8c","Type":"ContainerStarted","Data":"8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878"} Feb 25 07:42:33 crc kubenswrapper[4978]: I0225 07:42:33.188716 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6l4jw" podStartSLOduration=2.664386027 podStartE2EDuration="5.188656276s" podCreationTimestamp="2026-02-25 07:42:28 +0000 UTC" firstStartedPulling="2026-02-25 07:42:30.131727392 +0000 UTC m=+3443.570983901" lastFinishedPulling="2026-02-25 07:42:32.655997661 +0000 UTC m=+3446.095254150" observedRunningTime="2026-02-25 07:42:33.186448908 +0000 UTC m=+3446.625705407" watchObservedRunningTime="2026-02-25 07:42:33.188656276 +0000 UTC m=+3446.627912745" Feb 25 07:42:39 crc kubenswrapper[4978]: I0225 07:42:39.379171 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:39 crc kubenswrapper[4978]: I0225 07:42:39.379525 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:40 crc kubenswrapper[4978]: I0225 07:42:40.450120 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6l4jw" podUID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerName="registry-server" probeResult="failure" output=< Feb 25 07:42:40 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 07:42:40 crc kubenswrapper[4978]: > Feb 25 07:42:46 crc kubenswrapper[4978]: I0225 07:42:46.540427 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:42:46 crc kubenswrapper[4978]: I0225 07:42:46.541525 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:42:46 crc kubenswrapper[4978]: I0225 07:42:46.541609 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:42:46 crc kubenswrapper[4978]: I0225 07:42:46.542558 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e0615218c5cf39b6b1cde753ee4d72d39432f0b0ffecbd6b54f3614eaa3d0ab"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:42:46 crc kubenswrapper[4978]: I0225 07:42:46.542659 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://4e0615218c5cf39b6b1cde753ee4d72d39432f0b0ffecbd6b54f3614eaa3d0ab" gracePeriod=600 Feb 25 07:42:47 crc kubenswrapper[4978]: I0225 07:42:47.281817 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="4e0615218c5cf39b6b1cde753ee4d72d39432f0b0ffecbd6b54f3614eaa3d0ab" exitCode=0 Feb 25 07:42:47 crc kubenswrapper[4978]: I0225 07:42:47.281867 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"4e0615218c5cf39b6b1cde753ee4d72d39432f0b0ffecbd6b54f3614eaa3d0ab"} Feb 25 07:42:47 crc kubenswrapper[4978]: I0225 07:42:47.282400 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5"} Feb 25 07:42:47 crc kubenswrapper[4978]: I0225 07:42:47.282429 4978 scope.go:117] "RemoveContainer" containerID="1e808d616264f48338e3721726d3e4e60273ad0a1c14dc88fea229f515d2f754" Feb 25 07:42:49 crc kubenswrapper[4978]: I0225 07:42:49.460233 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:49 crc kubenswrapper[4978]: I0225 07:42:49.545915 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:49 crc kubenswrapper[4978]: I0225 07:42:49.701923 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6l4jw"] Feb 25 07:42:51 crc kubenswrapper[4978]: I0225 07:42:51.324167 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6l4jw" podUID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerName="registry-server" containerID="cri-o://8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878" gracePeriod=2 Feb 25 07:42:51 crc kubenswrapper[4978]: I0225 07:42:51.826676 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.010137 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpnm7\" (UniqueName: \"kubernetes.io/projected/63791838-0b10-4bbc-9e51-33bd195e3f8c-kube-api-access-mpnm7\") pod \"63791838-0b10-4bbc-9e51-33bd195e3f8c\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.010511 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-catalog-content\") pod \"63791838-0b10-4bbc-9e51-33bd195e3f8c\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.010638 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-utilities\") pod \"63791838-0b10-4bbc-9e51-33bd195e3f8c\" (UID: \"63791838-0b10-4bbc-9e51-33bd195e3f8c\") " Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.012169 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-utilities" (OuterVolumeSpecName: "utilities") pod "63791838-0b10-4bbc-9e51-33bd195e3f8c" (UID: "63791838-0b10-4bbc-9e51-33bd195e3f8c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.018427 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63791838-0b10-4bbc-9e51-33bd195e3f8c-kube-api-access-mpnm7" (OuterVolumeSpecName: "kube-api-access-mpnm7") pod "63791838-0b10-4bbc-9e51-33bd195e3f8c" (UID: "63791838-0b10-4bbc-9e51-33bd195e3f8c"). InnerVolumeSpecName "kube-api-access-mpnm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.112747 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpnm7\" (UniqueName: \"kubernetes.io/projected/63791838-0b10-4bbc-9e51-33bd195e3f8c-kube-api-access-mpnm7\") on node \"crc\" DevicePath \"\"" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.112796 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.232290 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63791838-0b10-4bbc-9e51-33bd195e3f8c" (UID: "63791838-0b10-4bbc-9e51-33bd195e3f8c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.316792 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63791838-0b10-4bbc-9e51-33bd195e3f8c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.338357 4978 generic.go:334] "Generic (PLEG): container finished" podID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerID="8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878" exitCode=0 Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.338453 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l4jw" event={"ID":"63791838-0b10-4bbc-9e51-33bd195e3f8c","Type":"ContainerDied","Data":"8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878"} Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.338493 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6l4jw" event={"ID":"63791838-0b10-4bbc-9e51-33bd195e3f8c","Type":"ContainerDied","Data":"cccd9586d48c46975f61c6353a06d82637aa393b2d1ccd537c6d38586cd77a3c"} Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.338522 4978 scope.go:117] "RemoveContainer" containerID="8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.338775 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6l4jw" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.374618 4978 scope.go:117] "RemoveContainer" containerID="4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.400062 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6l4jw"] Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.411695 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6l4jw"] Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.414273 4978 scope.go:117] "RemoveContainer" containerID="c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.444353 4978 scope.go:117] "RemoveContainer" containerID="8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878" Feb 25 07:42:52 crc kubenswrapper[4978]: E0225 07:42:52.445060 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878\": container with ID starting with 8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878 not found: ID does not exist" containerID="8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.445201 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878"} err="failed to get container status \"8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878\": rpc error: code = NotFound desc = could not find container \"8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878\": container with ID starting with 8b0ed557dbd3d6be5f1fc5a664ff375279564d6d5b35236709bcf7ba74fe2878 not found: ID does not exist" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.445312 4978 scope.go:117] "RemoveContainer" containerID="4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9" Feb 25 07:42:52 crc kubenswrapper[4978]: E0225 07:42:52.445935 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9\": container with ID starting with 4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9 not found: ID does not exist" containerID="4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.445994 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9"} err="failed to get container status \"4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9\": rpc error: code = NotFound desc = could not find container \"4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9\": container with ID starting with 4bceced01bb873abbc476dcefb13b0e6786d03052a7a25764874d163e43ec9f9 not found: ID does not exist" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.446036 4978 scope.go:117] "RemoveContainer" containerID="c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f" Feb 25 07:42:52 crc kubenswrapper[4978]: E0225 07:42:52.446563 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f\": container with ID starting with c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f not found: ID does not exist" containerID="c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f" Feb 25 07:42:52 crc kubenswrapper[4978]: I0225 07:42:52.446675 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f"} err="failed to get container status \"c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f\": rpc error: code = NotFound desc = could not find container \"c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f\": container with ID starting with c1061de623c5eb6062300f7b5d3b3cd6eeee393a8b767862875ea75f5f3ca63f not found: ID does not exist" Feb 25 07:42:53 crc kubenswrapper[4978]: I0225 07:42:53.343793 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63791838-0b10-4bbc-9e51-33bd195e3f8c" path="/var/lib/kubelet/pods/63791838-0b10-4bbc-9e51-33bd195e3f8c/volumes" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.665278 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4fv2z"] Feb 25 07:43:34 crc kubenswrapper[4978]: E0225 07:43:34.668771 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerName="extract-utilities" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.668836 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerName="extract-utilities" Feb 25 07:43:34 crc kubenswrapper[4978]: E0225 07:43:34.668872 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerName="extract-content" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.668893 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerName="extract-content" Feb 25 07:43:34 crc kubenswrapper[4978]: E0225 07:43:34.668917 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerName="registry-server" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.668933 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerName="registry-server" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.669434 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="63791838-0b10-4bbc-9e51-33bd195e3f8c" containerName="registry-server" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.671819 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.689131 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4fv2z"] Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.748963 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnpb7\" (UniqueName: \"kubernetes.io/projected/9f167b01-0645-4db7-88bf-a3c732d9050d-kube-api-access-tnpb7\") pod \"community-operators-4fv2z\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.749016 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-catalog-content\") pod \"community-operators-4fv2z\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.749146 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-utilities\") pod \"community-operators-4fv2z\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.850221 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-utilities\") pod \"community-operators-4fv2z\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.850284 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnpb7\" (UniqueName: \"kubernetes.io/projected/9f167b01-0645-4db7-88bf-a3c732d9050d-kube-api-access-tnpb7\") pod \"community-operators-4fv2z\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.850311 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-catalog-content\") pod \"community-operators-4fv2z\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.850806 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-utilities\") pod \"community-operators-4fv2z\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.850843 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-catalog-content\") pod \"community-operators-4fv2z\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:34 crc kubenswrapper[4978]: I0225 07:43:34.875971 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnpb7\" (UniqueName: \"kubernetes.io/projected/9f167b01-0645-4db7-88bf-a3c732d9050d-kube-api-access-tnpb7\") pod \"community-operators-4fv2z\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:35 crc kubenswrapper[4978]: I0225 07:43:35.006956 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:35 crc kubenswrapper[4978]: I0225 07:43:35.539993 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4fv2z"] Feb 25 07:43:35 crc kubenswrapper[4978]: W0225 07:43:35.544229 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f167b01_0645_4db7_88bf_a3c732d9050d.slice/crio-15c4189a528f34625cbee64c91348869b4a3ecb8fc09a3ea89ad74bd7592d65d WatchSource:0}: Error finding container 15c4189a528f34625cbee64c91348869b4a3ecb8fc09a3ea89ad74bd7592d65d: Status 404 returned error can't find the container with id 15c4189a528f34625cbee64c91348869b4a3ecb8fc09a3ea89ad74bd7592d65d Feb 25 07:43:35 crc kubenswrapper[4978]: I0225 07:43:35.971355 4978 generic.go:334] "Generic (PLEG): container finished" podID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerID="f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351" exitCode=0 Feb 25 07:43:35 crc kubenswrapper[4978]: I0225 07:43:35.971411 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fv2z" event={"ID":"9f167b01-0645-4db7-88bf-a3c732d9050d","Type":"ContainerDied","Data":"f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351"} Feb 25 07:43:35 crc kubenswrapper[4978]: I0225 07:43:35.976412 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fv2z" event={"ID":"9f167b01-0645-4db7-88bf-a3c732d9050d","Type":"ContainerStarted","Data":"15c4189a528f34625cbee64c91348869b4a3ecb8fc09a3ea89ad74bd7592d65d"} Feb 25 07:43:36 crc kubenswrapper[4978]: I0225 07:43:36.990410 4978 generic.go:334] "Generic (PLEG): container finished" podID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerID="5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496" exitCode=0 Feb 25 07:43:36 crc kubenswrapper[4978]: I0225 07:43:36.990506 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fv2z" event={"ID":"9f167b01-0645-4db7-88bf-a3c732d9050d","Type":"ContainerDied","Data":"5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496"} Feb 25 07:43:38 crc kubenswrapper[4978]: I0225 07:43:38.004393 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fv2z" event={"ID":"9f167b01-0645-4db7-88bf-a3c732d9050d","Type":"ContainerStarted","Data":"e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883"} Feb 25 07:43:38 crc kubenswrapper[4978]: I0225 07:43:38.034905 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4fv2z" podStartSLOduration=2.593286619 podStartE2EDuration="4.034885191s" podCreationTimestamp="2026-02-25 07:43:34 +0000 UTC" firstStartedPulling="2026-02-25 07:43:35.972082182 +0000 UTC m=+3509.411338651" lastFinishedPulling="2026-02-25 07:43:37.413680724 +0000 UTC m=+3510.852937223" observedRunningTime="2026-02-25 07:43:38.030343551 +0000 UTC m=+3511.469600010" watchObservedRunningTime="2026-02-25 07:43:38.034885191 +0000 UTC m=+3511.474141650" Feb 25 07:43:45 crc kubenswrapper[4978]: I0225 07:43:45.008431 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:45 crc kubenswrapper[4978]: I0225 07:43:45.009243 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:45 crc kubenswrapper[4978]: I0225 07:43:45.081170 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:45 crc kubenswrapper[4978]: I0225 07:43:45.149056 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:45 crc kubenswrapper[4978]: I0225 07:43:45.328586 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4fv2z"] Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.083168 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4fv2z" podUID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerName="registry-server" containerID="cri-o://e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883" gracePeriod=2 Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.587636 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.655864 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnpb7\" (UniqueName: \"kubernetes.io/projected/9f167b01-0645-4db7-88bf-a3c732d9050d-kube-api-access-tnpb7\") pod \"9f167b01-0645-4db7-88bf-a3c732d9050d\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.655987 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-utilities\") pod \"9f167b01-0645-4db7-88bf-a3c732d9050d\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.656046 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-catalog-content\") pod \"9f167b01-0645-4db7-88bf-a3c732d9050d\" (UID: \"9f167b01-0645-4db7-88bf-a3c732d9050d\") " Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.657739 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-utilities" (OuterVolumeSpecName: "utilities") pod "9f167b01-0645-4db7-88bf-a3c732d9050d" (UID: "9f167b01-0645-4db7-88bf-a3c732d9050d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.664781 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f167b01-0645-4db7-88bf-a3c732d9050d-kube-api-access-tnpb7" (OuterVolumeSpecName: "kube-api-access-tnpb7") pod "9f167b01-0645-4db7-88bf-a3c732d9050d" (UID: "9f167b01-0645-4db7-88bf-a3c732d9050d"). InnerVolumeSpecName "kube-api-access-tnpb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.737831 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f167b01-0645-4db7-88bf-a3c732d9050d" (UID: "9f167b01-0645-4db7-88bf-a3c732d9050d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.757822 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnpb7\" (UniqueName: \"kubernetes.io/projected/9f167b01-0645-4db7-88bf-a3c732d9050d-kube-api-access-tnpb7\") on node \"crc\" DevicePath \"\"" Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.757852 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:43:47 crc kubenswrapper[4978]: I0225 07:43:47.757866 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f167b01-0645-4db7-88bf-a3c732d9050d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.097148 4978 generic.go:334] "Generic (PLEG): container finished" podID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerID="e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883" exitCode=0 Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.097212 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fv2z" event={"ID":"9f167b01-0645-4db7-88bf-a3c732d9050d","Type":"ContainerDied","Data":"e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883"} Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.097251 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4fv2z" event={"ID":"9f167b01-0645-4db7-88bf-a3c732d9050d","Type":"ContainerDied","Data":"15c4189a528f34625cbee64c91348869b4a3ecb8fc09a3ea89ad74bd7592d65d"} Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.097280 4978 scope.go:117] "RemoveContainer" containerID="e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883" Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.097282 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4fv2z" Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.159644 4978 scope.go:117] "RemoveContainer" containerID="5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496" Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.197858 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4fv2z"] Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.204194 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4fv2z"] Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.210693 4978 scope.go:117] "RemoveContainer" containerID="f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351" Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.251411 4978 scope.go:117] "RemoveContainer" containerID="e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883" Feb 25 07:43:48 crc kubenswrapper[4978]: E0225 07:43:48.252086 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883\": container with ID starting with e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883 not found: ID does not exist" containerID="e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883" Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.252140 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883"} err="failed to get container status \"e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883\": rpc error: code = NotFound desc = could not find container \"e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883\": container with ID starting with e9470f9214d451e731c167983c5887d90872a6286497841b0cc0d06d6cdd3883 not found: ID does not exist" Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.252204 4978 scope.go:117] "RemoveContainer" containerID="5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496" Feb 25 07:43:48 crc kubenswrapper[4978]: E0225 07:43:48.252988 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496\": container with ID starting with 5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496 not found: ID does not exist" containerID="5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496" Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.253025 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496"} err="failed to get container status \"5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496\": rpc error: code = NotFound desc = could not find container \"5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496\": container with ID starting with 5a39140cd7fba1232b7edbb60b4a6f2c2fa751dcf252426ae2f378845037f496 not found: ID does not exist" Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.253050 4978 scope.go:117] "RemoveContainer" containerID="f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351" Feb 25 07:43:48 crc kubenswrapper[4978]: E0225 07:43:48.253560 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351\": container with ID starting with f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351 not found: ID does not exist" containerID="f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351" Feb 25 07:43:48 crc kubenswrapper[4978]: I0225 07:43:48.253611 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351"} err="failed to get container status \"f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351\": rpc error: code = NotFound desc = could not find container \"f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351\": container with ID starting with f7adcfbf640a1b375453a1000a08d0fd1dbbe31dc2fd0b713b3a3f1e2712e351 not found: ID does not exist" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.343152 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f167b01-0645-4db7-88bf-a3c732d9050d" path="/var/lib/kubelet/pods/9f167b01-0645-4db7-88bf-a3c732d9050d/volumes" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.554691 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmcx"] Feb 25 07:43:49 crc kubenswrapper[4978]: E0225 07:43:49.555198 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerName="extract-content" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.555226 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerName="extract-content" Feb 25 07:43:49 crc kubenswrapper[4978]: E0225 07:43:49.555255 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerName="extract-utilities" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.555269 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerName="extract-utilities" Feb 25 07:43:49 crc kubenswrapper[4978]: E0225 07:43:49.555283 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerName="registry-server" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.555297 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerName="registry-server" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.555684 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f167b01-0645-4db7-88bf-a3c732d9050d" containerName="registry-server" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.557498 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.568607 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmcx"] Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.687603 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-utilities\") pod \"redhat-marketplace-ckmcx\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.687711 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hhsz\" (UniqueName: \"kubernetes.io/projected/055f1e34-0516-4300-bfd5-ebc17c2b29c2-kube-api-access-8hhsz\") pod \"redhat-marketplace-ckmcx\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.687946 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-catalog-content\") pod \"redhat-marketplace-ckmcx\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.789542 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-utilities\") pod \"redhat-marketplace-ckmcx\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.789736 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hhsz\" (UniqueName: \"kubernetes.io/projected/055f1e34-0516-4300-bfd5-ebc17c2b29c2-kube-api-access-8hhsz\") pod \"redhat-marketplace-ckmcx\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.789869 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-catalog-content\") pod \"redhat-marketplace-ckmcx\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.790512 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-utilities\") pod \"redhat-marketplace-ckmcx\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.790620 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-catalog-content\") pod \"redhat-marketplace-ckmcx\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.825241 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hhsz\" (UniqueName: \"kubernetes.io/projected/055f1e34-0516-4300-bfd5-ebc17c2b29c2-kube-api-access-8hhsz\") pod \"redhat-marketplace-ckmcx\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:49 crc kubenswrapper[4978]: I0225 07:43:49.884217 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:50 crc kubenswrapper[4978]: I0225 07:43:50.162940 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmcx"] Feb 25 07:43:50 crc kubenswrapper[4978]: W0225 07:43:50.176673 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod055f1e34_0516_4300_bfd5_ebc17c2b29c2.slice/crio-71dff3c4c9d391a0e1b7c20bd6f4dd1f8369af63acbd767c7b32ab40a530396b WatchSource:0}: Error finding container 71dff3c4c9d391a0e1b7c20bd6f4dd1f8369af63acbd767c7b32ab40a530396b: Status 404 returned error can't find the container with id 71dff3c4c9d391a0e1b7c20bd6f4dd1f8369af63acbd767c7b32ab40a530396b Feb 25 07:43:51 crc kubenswrapper[4978]: I0225 07:43:51.128596 4978 generic.go:334] "Generic (PLEG): container finished" podID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerID="5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e" exitCode=0 Feb 25 07:43:51 crc kubenswrapper[4978]: I0225 07:43:51.128642 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmcx" event={"ID":"055f1e34-0516-4300-bfd5-ebc17c2b29c2","Type":"ContainerDied","Data":"5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e"} Feb 25 07:43:51 crc kubenswrapper[4978]: I0225 07:43:51.129110 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmcx" event={"ID":"055f1e34-0516-4300-bfd5-ebc17c2b29c2","Type":"ContainerStarted","Data":"71dff3c4c9d391a0e1b7c20bd6f4dd1f8369af63acbd767c7b32ab40a530396b"} Feb 25 07:43:53 crc kubenswrapper[4978]: I0225 07:43:53.157217 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmcx" event={"ID":"055f1e34-0516-4300-bfd5-ebc17c2b29c2","Type":"ContainerStarted","Data":"de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7"} Feb 25 07:43:54 crc kubenswrapper[4978]: I0225 07:43:54.171033 4978 generic.go:334] "Generic (PLEG): container finished" podID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerID="de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7" exitCode=0 Feb 25 07:43:54 crc kubenswrapper[4978]: I0225 07:43:54.171099 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmcx" event={"ID":"055f1e34-0516-4300-bfd5-ebc17c2b29c2","Type":"ContainerDied","Data":"de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7"} Feb 25 07:43:55 crc kubenswrapper[4978]: I0225 07:43:55.183730 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmcx" event={"ID":"055f1e34-0516-4300-bfd5-ebc17c2b29c2","Type":"ContainerStarted","Data":"3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7"} Feb 25 07:43:55 crc kubenswrapper[4978]: I0225 07:43:55.202691 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ckmcx" podStartSLOduration=2.7790038580000003 podStartE2EDuration="6.20266949s" podCreationTimestamp="2026-02-25 07:43:49 +0000 UTC" firstStartedPulling="2026-02-25 07:43:51.130693461 +0000 UTC m=+3524.569949960" lastFinishedPulling="2026-02-25 07:43:54.554359123 +0000 UTC m=+3527.993615592" observedRunningTime="2026-02-25 07:43:55.199512112 +0000 UTC m=+3528.638768591" watchObservedRunningTime="2026-02-25 07:43:55.20266949 +0000 UTC m=+3528.641925959" Feb 25 07:43:59 crc kubenswrapper[4978]: I0225 07:43:59.885294 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:59 crc kubenswrapper[4978]: I0225 07:43:59.885864 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:43:59 crc kubenswrapper[4978]: I0225 07:43:59.931667 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.148816 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533424-5zk4f"] Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.154714 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533424-5zk4f" Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.154708 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533424-5zk4f"] Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.158603 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.159057 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.159978 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.270185 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.305336 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm887\" (UniqueName: \"kubernetes.io/projected/c5e4f2fd-1af2-46f1-adc7-c76913ce6470-kube-api-access-pm887\") pod \"auto-csr-approver-29533424-5zk4f\" (UID: \"c5e4f2fd-1af2-46f1-adc7-c76913ce6470\") " pod="openshift-infra/auto-csr-approver-29533424-5zk4f" Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.330451 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmcx"] Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.407759 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm887\" (UniqueName: \"kubernetes.io/projected/c5e4f2fd-1af2-46f1-adc7-c76913ce6470-kube-api-access-pm887\") pod \"auto-csr-approver-29533424-5zk4f\" (UID: \"c5e4f2fd-1af2-46f1-adc7-c76913ce6470\") " pod="openshift-infra/auto-csr-approver-29533424-5zk4f" Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.448284 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm887\" (UniqueName: \"kubernetes.io/projected/c5e4f2fd-1af2-46f1-adc7-c76913ce6470-kube-api-access-pm887\") pod \"auto-csr-approver-29533424-5zk4f\" (UID: \"c5e4f2fd-1af2-46f1-adc7-c76913ce6470\") " pod="openshift-infra/auto-csr-approver-29533424-5zk4f" Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.471249 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533424-5zk4f" Feb 25 07:44:00 crc kubenswrapper[4978]: I0225 07:44:00.808094 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533424-5zk4f"] Feb 25 07:44:01 crc kubenswrapper[4978]: I0225 07:44:01.234502 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533424-5zk4f" event={"ID":"c5e4f2fd-1af2-46f1-adc7-c76913ce6470","Type":"ContainerStarted","Data":"cb9f0f498afac8637bcc7bc42fd99b28a64dab86d7d669a195d757e4780754e7"} Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.243254 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533424-5zk4f" event={"ID":"c5e4f2fd-1af2-46f1-adc7-c76913ce6470","Type":"ContainerStarted","Data":"7d65562f086b6c4f391e2fbeaaf09e8df3f67cd182d73f4fea37953d452b146d"} Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.243419 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ckmcx" podUID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerName="registry-server" containerID="cri-o://3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7" gracePeriod=2 Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.282253 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533424-5zk4f" podStartSLOduration=1.402682667 podStartE2EDuration="2.282217884s" podCreationTimestamp="2026-02-25 07:44:00 +0000 UTC" firstStartedPulling="2026-02-25 07:44:00.812389367 +0000 UTC m=+3534.251645826" lastFinishedPulling="2026-02-25 07:44:01.691924574 +0000 UTC m=+3535.131181043" observedRunningTime="2026-02-25 07:44:02.273752672 +0000 UTC m=+3535.713009161" watchObservedRunningTime="2026-02-25 07:44:02.282217884 +0000 UTC m=+3535.721474383" Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.698680 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.844489 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-utilities\") pod \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.844641 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hhsz\" (UniqueName: \"kubernetes.io/projected/055f1e34-0516-4300-bfd5-ebc17c2b29c2-kube-api-access-8hhsz\") pod \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.844674 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-catalog-content\") pod \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\" (UID: \"055f1e34-0516-4300-bfd5-ebc17c2b29c2\") " Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.845339 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-utilities" (OuterVolumeSpecName: "utilities") pod "055f1e34-0516-4300-bfd5-ebc17c2b29c2" (UID: "055f1e34-0516-4300-bfd5-ebc17c2b29c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.851435 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/055f1e34-0516-4300-bfd5-ebc17c2b29c2-kube-api-access-8hhsz" (OuterVolumeSpecName: "kube-api-access-8hhsz") pod "055f1e34-0516-4300-bfd5-ebc17c2b29c2" (UID: "055f1e34-0516-4300-bfd5-ebc17c2b29c2"). InnerVolumeSpecName "kube-api-access-8hhsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.880032 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "055f1e34-0516-4300-bfd5-ebc17c2b29c2" (UID: "055f1e34-0516-4300-bfd5-ebc17c2b29c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.946317 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hhsz\" (UniqueName: \"kubernetes.io/projected/055f1e34-0516-4300-bfd5-ebc17c2b29c2-kube-api-access-8hhsz\") on node \"crc\" DevicePath \"\"" Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.946356 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:44:02 crc kubenswrapper[4978]: I0225 07:44:02.946400 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/055f1e34-0516-4300-bfd5-ebc17c2b29c2-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.252055 4978 generic.go:334] "Generic (PLEG): container finished" podID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerID="3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7" exitCode=0 Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.252103 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmcx" event={"ID":"055f1e34-0516-4300-bfd5-ebc17c2b29c2","Type":"ContainerDied","Data":"3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7"} Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.252434 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckmcx" event={"ID":"055f1e34-0516-4300-bfd5-ebc17c2b29c2","Type":"ContainerDied","Data":"71dff3c4c9d391a0e1b7c20bd6f4dd1f8369af63acbd767c7b32ab40a530396b"} Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.252456 4978 scope.go:117] "RemoveContainer" containerID="3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.252214 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckmcx" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.255133 4978 generic.go:334] "Generic (PLEG): container finished" podID="c5e4f2fd-1af2-46f1-adc7-c76913ce6470" containerID="7d65562f086b6c4f391e2fbeaaf09e8df3f67cd182d73f4fea37953d452b146d" exitCode=0 Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.255170 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533424-5zk4f" event={"ID":"c5e4f2fd-1af2-46f1-adc7-c76913ce6470","Type":"ContainerDied","Data":"7d65562f086b6c4f391e2fbeaaf09e8df3f67cd182d73f4fea37953d452b146d"} Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.303906 4978 scope.go:117] "RemoveContainer" containerID="de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.318099 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmcx"] Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.325757 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckmcx"] Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.330784 4978 scope.go:117] "RemoveContainer" containerID="5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.336573 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" path="/var/lib/kubelet/pods/055f1e34-0516-4300-bfd5-ebc17c2b29c2/volumes" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.360913 4978 scope.go:117] "RemoveContainer" containerID="3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7" Feb 25 07:44:03 crc kubenswrapper[4978]: E0225 07:44:03.361361 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7\": container with ID starting with 3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7 not found: ID does not exist" containerID="3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.361443 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7"} err="failed to get container status \"3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7\": rpc error: code = NotFound desc = could not find container \"3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7\": container with ID starting with 3aaeeef56365e8ca02aeac99dcd8dcee1e7f4a2aa63604eb070cccc48f5b21b7 not found: ID does not exist" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.361501 4978 scope.go:117] "RemoveContainer" containerID="de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7" Feb 25 07:44:03 crc kubenswrapper[4978]: E0225 07:44:03.361832 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7\": container with ID starting with de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7 not found: ID does not exist" containerID="de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.361922 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7"} err="failed to get container status \"de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7\": rpc error: code = NotFound desc = could not find container \"de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7\": container with ID starting with de7ed70a79de80fa664efd3b869276bd4113b617d72401a79880fae765c54eb7 not found: ID does not exist" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.361993 4978 scope.go:117] "RemoveContainer" containerID="5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e" Feb 25 07:44:03 crc kubenswrapper[4978]: E0225 07:44:03.362390 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e\": container with ID starting with 5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e not found: ID does not exist" containerID="5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e" Feb 25 07:44:03 crc kubenswrapper[4978]: I0225 07:44:03.362431 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e"} err="failed to get container status \"5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e\": rpc error: code = NotFound desc = could not find container \"5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e\": container with ID starting with 5834bbadd0c9bb7573e9e1bcf7bb2917af4295574852eb833c22f93f6af0fb3e not found: ID does not exist" Feb 25 07:44:04 crc kubenswrapper[4978]: I0225 07:44:04.637623 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533424-5zk4f" Feb 25 07:44:04 crc kubenswrapper[4978]: I0225 07:44:04.771244 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm887\" (UniqueName: \"kubernetes.io/projected/c5e4f2fd-1af2-46f1-adc7-c76913ce6470-kube-api-access-pm887\") pod \"c5e4f2fd-1af2-46f1-adc7-c76913ce6470\" (UID: \"c5e4f2fd-1af2-46f1-adc7-c76913ce6470\") " Feb 25 07:44:04 crc kubenswrapper[4978]: I0225 07:44:04.778892 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5e4f2fd-1af2-46f1-adc7-c76913ce6470-kube-api-access-pm887" (OuterVolumeSpecName: "kube-api-access-pm887") pod "c5e4f2fd-1af2-46f1-adc7-c76913ce6470" (UID: "c5e4f2fd-1af2-46f1-adc7-c76913ce6470"). InnerVolumeSpecName "kube-api-access-pm887". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:44:04 crc kubenswrapper[4978]: I0225 07:44:04.873285 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm887\" (UniqueName: \"kubernetes.io/projected/c5e4f2fd-1af2-46f1-adc7-c76913ce6470-kube-api-access-pm887\") on node \"crc\" DevicePath \"\"" Feb 25 07:44:05 crc kubenswrapper[4978]: I0225 07:44:05.275812 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533424-5zk4f" event={"ID":"c5e4f2fd-1af2-46f1-adc7-c76913ce6470","Type":"ContainerDied","Data":"cb9f0f498afac8637bcc7bc42fd99b28a64dab86d7d669a195d757e4780754e7"} Feb 25 07:44:05 crc kubenswrapper[4978]: I0225 07:44:05.275853 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb9f0f498afac8637bcc7bc42fd99b28a64dab86d7d669a195d757e4780754e7" Feb 25 07:44:05 crc kubenswrapper[4978]: I0225 07:44:05.275909 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533424-5zk4f" Feb 25 07:44:05 crc kubenswrapper[4978]: I0225 07:44:05.336049 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533418-42npc"] Feb 25 07:44:05 crc kubenswrapper[4978]: I0225 07:44:05.336219 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533418-42npc"] Feb 25 07:44:07 crc kubenswrapper[4978]: I0225 07:44:07.347464 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="325ca167-ab03-4703-b1f3-ce57e98dfdbb" path="/var/lib/kubelet/pods/325ca167-ab03-4703-b1f3-ce57e98dfdbb/volumes" Feb 25 07:44:23 crc kubenswrapper[4978]: I0225 07:44:23.038592 4978 scope.go:117] "RemoveContainer" containerID="0c412716ec71b35a9ef60e3ca82a45fef3671679b2456fbc4c16354d5858a5d1" Feb 25 07:44:46 crc kubenswrapper[4978]: I0225 07:44:46.540572 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:44:46 crc kubenswrapper[4978]: I0225 07:44:46.541309 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.165419 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp"] Feb 25 07:45:00 crc kubenswrapper[4978]: E0225 07:45:00.166302 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerName="registry-server" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.166320 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerName="registry-server" Feb 25 07:45:00 crc kubenswrapper[4978]: E0225 07:45:00.166338 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerName="extract-utilities" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.166346 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerName="extract-utilities" Feb 25 07:45:00 crc kubenswrapper[4978]: E0225 07:45:00.166394 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5e4f2fd-1af2-46f1-adc7-c76913ce6470" containerName="oc" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.166403 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5e4f2fd-1af2-46f1-adc7-c76913ce6470" containerName="oc" Feb 25 07:45:00 crc kubenswrapper[4978]: E0225 07:45:00.166426 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerName="extract-content" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.166435 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerName="extract-content" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.166596 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="055f1e34-0516-4300-bfd5-ebc17c2b29c2" containerName="registry-server" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.166629 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5e4f2fd-1af2-46f1-adc7-c76913ce6470" containerName="oc" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.167190 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.172457 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.173061 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.174551 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp"] Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.308958 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-config-volume\") pod \"collect-profiles-29533425-4d6zp\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.309179 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkklh\" (UniqueName: \"kubernetes.io/projected/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-kube-api-access-bkklh\") pod \"collect-profiles-29533425-4d6zp\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.309244 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-secret-volume\") pod \"collect-profiles-29533425-4d6zp\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.410999 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-config-volume\") pod \"collect-profiles-29533425-4d6zp\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.411131 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkklh\" (UniqueName: \"kubernetes.io/projected/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-kube-api-access-bkklh\") pod \"collect-profiles-29533425-4d6zp\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.411189 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-secret-volume\") pod \"collect-profiles-29533425-4d6zp\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.412815 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-config-volume\") pod \"collect-profiles-29533425-4d6zp\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.418776 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-secret-volume\") pod \"collect-profiles-29533425-4d6zp\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.429782 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkklh\" (UniqueName: \"kubernetes.io/projected/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-kube-api-access-bkklh\") pod \"collect-profiles-29533425-4d6zp\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.484731 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:00 crc kubenswrapper[4978]: I0225 07:45:00.945240 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp"] Feb 25 07:45:00 crc kubenswrapper[4978]: W0225 07:45:00.959055 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0295eb9_4a00_4bb7_9eb0_b8c35e9a506a.slice/crio-e3adce74bcef749d3b84d96d0c9ced3d7aedc885c886aa81f9ef15b46e7bbffc WatchSource:0}: Error finding container e3adce74bcef749d3b84d96d0c9ced3d7aedc885c886aa81f9ef15b46e7bbffc: Status 404 returned error can't find the container with id e3adce74bcef749d3b84d96d0c9ced3d7aedc885c886aa81f9ef15b46e7bbffc Feb 25 07:45:01 crc kubenswrapper[4978]: I0225 07:45:01.799976 4978 generic.go:334] "Generic (PLEG): container finished" podID="c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a" containerID="6588bfe6ec137758f126ff6eaaa355f2fc9cf342f3cf6887768a3e31fe853fa1" exitCode=0 Feb 25 07:45:01 crc kubenswrapper[4978]: I0225 07:45:01.800085 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" event={"ID":"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a","Type":"ContainerDied","Data":"6588bfe6ec137758f126ff6eaaa355f2fc9cf342f3cf6887768a3e31fe853fa1"} Feb 25 07:45:01 crc kubenswrapper[4978]: I0225 07:45:01.800511 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" event={"ID":"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a","Type":"ContainerStarted","Data":"e3adce74bcef749d3b84d96d0c9ced3d7aedc885c886aa81f9ef15b46e7bbffc"} Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.196193 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.359285 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-config-volume\") pod \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.359437 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkklh\" (UniqueName: \"kubernetes.io/projected/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-kube-api-access-bkklh\") pod \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.359505 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-secret-volume\") pod \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\" (UID: \"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a\") " Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.360694 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-config-volume" (OuterVolumeSpecName: "config-volume") pod "c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a" (UID: "c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.365009 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a" (UID: "c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.365505 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-kube-api-access-bkklh" (OuterVolumeSpecName: "kube-api-access-bkklh") pod "c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a" (UID: "c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a"). InnerVolumeSpecName "kube-api-access-bkklh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.461573 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.461600 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkklh\" (UniqueName: \"kubernetes.io/projected/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-kube-api-access-bkklh\") on node \"crc\" DevicePath \"\"" Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.461612 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.820867 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" event={"ID":"c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a","Type":"ContainerDied","Data":"e3adce74bcef749d3b84d96d0c9ced3d7aedc885c886aa81f9ef15b46e7bbffc"} Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.820926 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3adce74bcef749d3b84d96d0c9ced3d7aedc885c886aa81f9ef15b46e7bbffc" Feb 25 07:45:03 crc kubenswrapper[4978]: I0225 07:45:03.820951 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp" Feb 25 07:45:04 crc kubenswrapper[4978]: I0225 07:45:04.288444 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8"] Feb 25 07:45:04 crc kubenswrapper[4978]: I0225 07:45:04.295019 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533380-7vwq8"] Feb 25 07:45:05 crc kubenswrapper[4978]: I0225 07:45:05.343168 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2bef0e-1e45-4eb3-9525-3328e351fa3d" path="/var/lib/kubelet/pods/1b2bef0e-1e45-4eb3-9525-3328e351fa3d/volumes" Feb 25 07:45:16 crc kubenswrapper[4978]: I0225 07:45:16.540873 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:45:16 crc kubenswrapper[4978]: I0225 07:45:16.541701 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:45:23 crc kubenswrapper[4978]: I0225 07:45:23.130927 4978 scope.go:117] "RemoveContainer" containerID="5027f2c80274584aa70cb25ea3445ad7171efdbc698dfc62136aa6c03ee2eec7" Feb 25 07:45:46 crc kubenswrapper[4978]: I0225 07:45:46.540802 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:45:46 crc kubenswrapper[4978]: I0225 07:45:46.541665 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:45:46 crc kubenswrapper[4978]: I0225 07:45:46.541739 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:45:46 crc kubenswrapper[4978]: I0225 07:45:46.542722 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:45:46 crc kubenswrapper[4978]: I0225 07:45:46.542811 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" gracePeriod=600 Feb 25 07:45:46 crc kubenswrapper[4978]: E0225 07:45:46.679299 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:45:47 crc kubenswrapper[4978]: I0225 07:45:47.355410 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" exitCode=0 Feb 25 07:45:47 crc kubenswrapper[4978]: I0225 07:45:47.355470 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5"} Feb 25 07:45:47 crc kubenswrapper[4978]: I0225 07:45:47.355519 4978 scope.go:117] "RemoveContainer" containerID="4e0615218c5cf39b6b1cde753ee4d72d39432f0b0ffecbd6b54f3614eaa3d0ab" Feb 25 07:45:47 crc kubenswrapper[4978]: I0225 07:45:47.356539 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:45:47 crc kubenswrapper[4978]: E0225 07:45:47.357038 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.142617 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533426-cgd2c"] Feb 25 07:46:00 crc kubenswrapper[4978]: E0225 07:46:00.143513 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a" containerName="collect-profiles" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.143529 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a" containerName="collect-profiles" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.143707 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a" containerName="collect-profiles" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.144217 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533426-cgd2c" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.146669 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.147750 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.148344 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.162236 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533426-cgd2c"] Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.250199 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr794\" (UniqueName: \"kubernetes.io/projected/d561355a-1c08-4a5e-8584-af557e06909d-kube-api-access-pr794\") pod \"auto-csr-approver-29533426-cgd2c\" (UID: \"d561355a-1c08-4a5e-8584-af557e06909d\") " pod="openshift-infra/auto-csr-approver-29533426-cgd2c" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.353245 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr794\" (UniqueName: \"kubernetes.io/projected/d561355a-1c08-4a5e-8584-af557e06909d-kube-api-access-pr794\") pod \"auto-csr-approver-29533426-cgd2c\" (UID: \"d561355a-1c08-4a5e-8584-af557e06909d\") " pod="openshift-infra/auto-csr-approver-29533426-cgd2c" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.381876 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr794\" (UniqueName: \"kubernetes.io/projected/d561355a-1c08-4a5e-8584-af557e06909d-kube-api-access-pr794\") pod \"auto-csr-approver-29533426-cgd2c\" (UID: \"d561355a-1c08-4a5e-8584-af557e06909d\") " pod="openshift-infra/auto-csr-approver-29533426-cgd2c" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.467068 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533426-cgd2c" Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.723358 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533426-cgd2c"] Feb 25 07:46:00 crc kubenswrapper[4978]: W0225 07:46:00.738290 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd561355a_1c08_4a5e_8584_af557e06909d.slice/crio-be8e9e23753c15f1906a5d43234818822d3d67bac3771e30fed9a81c4c5e8294 WatchSource:0}: Error finding container be8e9e23753c15f1906a5d43234818822d3d67bac3771e30fed9a81c4c5e8294: Status 404 returned error can't find the container with id be8e9e23753c15f1906a5d43234818822d3d67bac3771e30fed9a81c4c5e8294 Feb 25 07:46:00 crc kubenswrapper[4978]: I0225 07:46:00.741767 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 07:46:01 crc kubenswrapper[4978]: I0225 07:46:01.475735 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533426-cgd2c" event={"ID":"d561355a-1c08-4a5e-8584-af557e06909d","Type":"ContainerStarted","Data":"be8e9e23753c15f1906a5d43234818822d3d67bac3771e30fed9a81c4c5e8294"} Feb 25 07:46:02 crc kubenswrapper[4978]: I0225 07:46:02.327467 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:46:02 crc kubenswrapper[4978]: E0225 07:46:02.327959 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:46:02 crc kubenswrapper[4978]: I0225 07:46:02.482200 4978 generic.go:334] "Generic (PLEG): container finished" podID="d561355a-1c08-4a5e-8584-af557e06909d" containerID="2c3b0ed54ece8630241797593b43e50ab4d5cdbd0497e404712182c36ae5ce56" exitCode=0 Feb 25 07:46:02 crc kubenswrapper[4978]: I0225 07:46:02.482236 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533426-cgd2c" event={"ID":"d561355a-1c08-4a5e-8584-af557e06909d","Type":"ContainerDied","Data":"2c3b0ed54ece8630241797593b43e50ab4d5cdbd0497e404712182c36ae5ce56"} Feb 25 07:46:03 crc kubenswrapper[4978]: I0225 07:46:03.813594 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533426-cgd2c" Feb 25 07:46:03 crc kubenswrapper[4978]: I0225 07:46:03.910479 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr794\" (UniqueName: \"kubernetes.io/projected/d561355a-1c08-4a5e-8584-af557e06909d-kube-api-access-pr794\") pod \"d561355a-1c08-4a5e-8584-af557e06909d\" (UID: \"d561355a-1c08-4a5e-8584-af557e06909d\") " Feb 25 07:46:03 crc kubenswrapper[4978]: I0225 07:46:03.915663 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d561355a-1c08-4a5e-8584-af557e06909d-kube-api-access-pr794" (OuterVolumeSpecName: "kube-api-access-pr794") pod "d561355a-1c08-4a5e-8584-af557e06909d" (UID: "d561355a-1c08-4a5e-8584-af557e06909d"). InnerVolumeSpecName "kube-api-access-pr794". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:46:04 crc kubenswrapper[4978]: I0225 07:46:04.011626 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr794\" (UniqueName: \"kubernetes.io/projected/d561355a-1c08-4a5e-8584-af557e06909d-kube-api-access-pr794\") on node \"crc\" DevicePath \"\"" Feb 25 07:46:04 crc kubenswrapper[4978]: I0225 07:46:04.501809 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533426-cgd2c" event={"ID":"d561355a-1c08-4a5e-8584-af557e06909d","Type":"ContainerDied","Data":"be8e9e23753c15f1906a5d43234818822d3d67bac3771e30fed9a81c4c5e8294"} Feb 25 07:46:04 crc kubenswrapper[4978]: I0225 07:46:04.501856 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be8e9e23753c15f1906a5d43234818822d3d67bac3771e30fed9a81c4c5e8294" Feb 25 07:46:04 crc kubenswrapper[4978]: I0225 07:46:04.502337 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533426-cgd2c" Feb 25 07:46:04 crc kubenswrapper[4978]: I0225 07:46:04.932258 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533420-zwpxs"] Feb 25 07:46:04 crc kubenswrapper[4978]: I0225 07:46:04.944168 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533420-zwpxs"] Feb 25 07:46:05 crc kubenswrapper[4978]: I0225 07:46:05.334699 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b9eb274-2592-4593-bb61-4387f8416f36" path="/var/lib/kubelet/pods/4b9eb274-2592-4593-bb61-4387f8416f36/volumes" Feb 25 07:46:13 crc kubenswrapper[4978]: I0225 07:46:13.365842 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:46:13 crc kubenswrapper[4978]: E0225 07:46:13.368248 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:46:23 crc kubenswrapper[4978]: I0225 07:46:23.215987 4978 scope.go:117] "RemoveContainer" containerID="b523bc6d2b5aaa6355d9541669197aae9f484b45984acc20e5a8b7800b13359c" Feb 25 07:46:25 crc kubenswrapper[4978]: I0225 07:46:25.328678 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:46:25 crc kubenswrapper[4978]: E0225 07:46:25.329582 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:46:40 crc kubenswrapper[4978]: I0225 07:46:40.327642 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:46:40 crc kubenswrapper[4978]: E0225 07:46:40.328291 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:46:55 crc kubenswrapper[4978]: I0225 07:46:55.329176 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:46:55 crc kubenswrapper[4978]: E0225 07:46:55.329982 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:47:06 crc kubenswrapper[4978]: I0225 07:47:06.327133 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:47:06 crc kubenswrapper[4978]: E0225 07:47:06.327948 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:47:19 crc kubenswrapper[4978]: I0225 07:47:19.327594 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:47:19 crc kubenswrapper[4978]: E0225 07:47:19.328907 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:47:32 crc kubenswrapper[4978]: I0225 07:47:32.327595 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:47:32 crc kubenswrapper[4978]: E0225 07:47:32.328919 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:47:46 crc kubenswrapper[4978]: I0225 07:47:46.327622 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:47:46 crc kubenswrapper[4978]: E0225 07:47:46.328484 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:47:57 crc kubenswrapper[4978]: I0225 07:47:57.334889 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:47:57 crc kubenswrapper[4978]: E0225 07:47:57.336265 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.169858 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533428-4vnzj"] Feb 25 07:48:00 crc kubenswrapper[4978]: E0225 07:48:00.170531 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d561355a-1c08-4a5e-8584-af557e06909d" containerName="oc" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.170563 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d561355a-1c08-4a5e-8584-af557e06909d" containerName="oc" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.170900 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d561355a-1c08-4a5e-8584-af557e06909d" containerName="oc" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.171742 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533428-4vnzj" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.175557 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.175878 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.177583 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.188621 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533428-4vnzj"] Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.266222 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv47g\" (UniqueName: \"kubernetes.io/projected/396e5a14-f984-49fd-84d6-27498a036777-kube-api-access-fv47g\") pod \"auto-csr-approver-29533428-4vnzj\" (UID: \"396e5a14-f984-49fd-84d6-27498a036777\") " pod="openshift-infra/auto-csr-approver-29533428-4vnzj" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.368023 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv47g\" (UniqueName: \"kubernetes.io/projected/396e5a14-f984-49fd-84d6-27498a036777-kube-api-access-fv47g\") pod \"auto-csr-approver-29533428-4vnzj\" (UID: \"396e5a14-f984-49fd-84d6-27498a036777\") " pod="openshift-infra/auto-csr-approver-29533428-4vnzj" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.398196 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv47g\" (UniqueName: \"kubernetes.io/projected/396e5a14-f984-49fd-84d6-27498a036777-kube-api-access-fv47g\") pod \"auto-csr-approver-29533428-4vnzj\" (UID: \"396e5a14-f984-49fd-84d6-27498a036777\") " pod="openshift-infra/auto-csr-approver-29533428-4vnzj" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.493321 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533428-4vnzj" Feb 25 07:48:00 crc kubenswrapper[4978]: I0225 07:48:00.826749 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533428-4vnzj"] Feb 25 07:48:01 crc kubenswrapper[4978]: I0225 07:48:01.592795 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533428-4vnzj" event={"ID":"396e5a14-f984-49fd-84d6-27498a036777","Type":"ContainerStarted","Data":"f49f1e647a5afedc6f7c2229c0ddafb24cbf12a178189241583fdf6fcd0b46aa"} Feb 25 07:48:03 crc kubenswrapper[4978]: I0225 07:48:03.617188 4978 generic.go:334] "Generic (PLEG): container finished" podID="396e5a14-f984-49fd-84d6-27498a036777" containerID="0dfb3bd24dcdbb57854e3a71cc583d4b4251b43b3e0da0d3c955bfbb7491e454" exitCode=0 Feb 25 07:48:03 crc kubenswrapper[4978]: I0225 07:48:03.617268 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533428-4vnzj" event={"ID":"396e5a14-f984-49fd-84d6-27498a036777","Type":"ContainerDied","Data":"0dfb3bd24dcdbb57854e3a71cc583d4b4251b43b3e0da0d3c955bfbb7491e454"} Feb 25 07:48:04 crc kubenswrapper[4978]: I0225 07:48:04.980987 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533428-4vnzj" Feb 25 07:48:05 crc kubenswrapper[4978]: I0225 07:48:05.045349 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv47g\" (UniqueName: \"kubernetes.io/projected/396e5a14-f984-49fd-84d6-27498a036777-kube-api-access-fv47g\") pod \"396e5a14-f984-49fd-84d6-27498a036777\" (UID: \"396e5a14-f984-49fd-84d6-27498a036777\") " Feb 25 07:48:05 crc kubenswrapper[4978]: I0225 07:48:05.050865 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/396e5a14-f984-49fd-84d6-27498a036777-kube-api-access-fv47g" (OuterVolumeSpecName: "kube-api-access-fv47g") pod "396e5a14-f984-49fd-84d6-27498a036777" (UID: "396e5a14-f984-49fd-84d6-27498a036777"). InnerVolumeSpecName "kube-api-access-fv47g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:48:05 crc kubenswrapper[4978]: I0225 07:48:05.147415 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv47g\" (UniqueName: \"kubernetes.io/projected/396e5a14-f984-49fd-84d6-27498a036777-kube-api-access-fv47g\") on node \"crc\" DevicePath \"\"" Feb 25 07:48:05 crc kubenswrapper[4978]: I0225 07:48:05.640132 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533428-4vnzj" event={"ID":"396e5a14-f984-49fd-84d6-27498a036777","Type":"ContainerDied","Data":"f49f1e647a5afedc6f7c2229c0ddafb24cbf12a178189241583fdf6fcd0b46aa"} Feb 25 07:48:05 crc kubenswrapper[4978]: I0225 07:48:05.640196 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f49f1e647a5afedc6f7c2229c0ddafb24cbf12a178189241583fdf6fcd0b46aa" Feb 25 07:48:05 crc kubenswrapper[4978]: I0225 07:48:05.640245 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533428-4vnzj" Feb 25 07:48:06 crc kubenswrapper[4978]: I0225 07:48:06.092670 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533422-t27jr"] Feb 25 07:48:06 crc kubenswrapper[4978]: I0225 07:48:06.099758 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533422-t27jr"] Feb 25 07:48:07 crc kubenswrapper[4978]: I0225 07:48:07.353738 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32cc4832-7c7c-40ae-9aac-30af12e9af4c" path="/var/lib/kubelet/pods/32cc4832-7c7c-40ae-9aac-30af12e9af4c/volumes" Feb 25 07:48:12 crc kubenswrapper[4978]: I0225 07:48:12.328864 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:48:12 crc kubenswrapper[4978]: E0225 07:48:12.330319 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:48:23 crc kubenswrapper[4978]: I0225 07:48:23.349015 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:48:23 crc kubenswrapper[4978]: E0225 07:48:23.349599 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:48:23 crc kubenswrapper[4978]: I0225 07:48:23.351000 4978 scope.go:117] "RemoveContainer" containerID="7080351c1386cb22043a3304b4041d234d1e7da50e9e1f049168bda9fdd48fb1" Feb 25 07:48:38 crc kubenswrapper[4978]: I0225 07:48:38.328800 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:48:38 crc kubenswrapper[4978]: E0225 07:48:38.329806 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:48:49 crc kubenswrapper[4978]: I0225 07:48:49.328299 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:48:49 crc kubenswrapper[4978]: E0225 07:48:49.329398 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:49:04 crc kubenswrapper[4978]: I0225 07:49:04.328123 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:49:04 crc kubenswrapper[4978]: E0225 07:49:04.328796 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:49:16 crc kubenswrapper[4978]: I0225 07:49:16.328309 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:49:16 crc kubenswrapper[4978]: E0225 07:49:16.329063 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:49:30 crc kubenswrapper[4978]: I0225 07:49:30.327836 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:49:30 crc kubenswrapper[4978]: E0225 07:49:30.328767 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:49:44 crc kubenswrapper[4978]: I0225 07:49:44.327489 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:49:44 crc kubenswrapper[4978]: E0225 07:49:44.328121 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:49:59 crc kubenswrapper[4978]: I0225 07:49:59.327646 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:49:59 crc kubenswrapper[4978]: E0225 07:49:59.328858 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.178965 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533430-w2682"] Feb 25 07:50:00 crc kubenswrapper[4978]: E0225 07:50:00.179625 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="396e5a14-f984-49fd-84d6-27498a036777" containerName="oc" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.179662 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="396e5a14-f984-49fd-84d6-27498a036777" containerName="oc" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.179990 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="396e5a14-f984-49fd-84d6-27498a036777" containerName="oc" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.181492 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533430-w2682" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.183775 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.183931 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.185125 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.196735 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533430-w2682"] Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.241695 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt6mz\" (UniqueName: \"kubernetes.io/projected/f8752058-4f5d-40b2-a8b7-42105c8d85a2-kube-api-access-bt6mz\") pod \"auto-csr-approver-29533430-w2682\" (UID: \"f8752058-4f5d-40b2-a8b7-42105c8d85a2\") " pod="openshift-infra/auto-csr-approver-29533430-w2682" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.343619 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt6mz\" (UniqueName: \"kubernetes.io/projected/f8752058-4f5d-40b2-a8b7-42105c8d85a2-kube-api-access-bt6mz\") pod \"auto-csr-approver-29533430-w2682\" (UID: \"f8752058-4f5d-40b2-a8b7-42105c8d85a2\") " pod="openshift-infra/auto-csr-approver-29533430-w2682" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.368429 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt6mz\" (UniqueName: \"kubernetes.io/projected/f8752058-4f5d-40b2-a8b7-42105c8d85a2-kube-api-access-bt6mz\") pod \"auto-csr-approver-29533430-w2682\" (UID: \"f8752058-4f5d-40b2-a8b7-42105c8d85a2\") " pod="openshift-infra/auto-csr-approver-29533430-w2682" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.507671 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533430-w2682" Feb 25 07:50:00 crc kubenswrapper[4978]: I0225 07:50:00.827449 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533430-w2682"] Feb 25 07:50:01 crc kubenswrapper[4978]: I0225 07:50:01.744816 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533430-w2682" event={"ID":"f8752058-4f5d-40b2-a8b7-42105c8d85a2","Type":"ContainerStarted","Data":"43edeb7bc90b42a9f11bdaec8a0d49cabbfce80cd695d372e8e240d6a8a5396b"} Feb 25 07:50:02 crc kubenswrapper[4978]: I0225 07:50:02.766406 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533430-w2682" event={"ID":"f8752058-4f5d-40b2-a8b7-42105c8d85a2","Type":"ContainerStarted","Data":"6f96aa23a5b930fa2a3420549881cba4ee8d2839fc425259495da07c928fdf8c"} Feb 25 07:50:02 crc kubenswrapper[4978]: I0225 07:50:02.786961 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533430-w2682" podStartSLOduration=1.391669737 podStartE2EDuration="2.786936124s" podCreationTimestamp="2026-02-25 07:50:00 +0000 UTC" firstStartedPulling="2026-02-25 07:50:00.825977627 +0000 UTC m=+3894.265234106" lastFinishedPulling="2026-02-25 07:50:02.221244024 +0000 UTC m=+3895.660500493" observedRunningTime="2026-02-25 07:50:02.782935919 +0000 UTC m=+3896.222192388" watchObservedRunningTime="2026-02-25 07:50:02.786936124 +0000 UTC m=+3896.226192633" Feb 25 07:50:03 crc kubenswrapper[4978]: I0225 07:50:03.778736 4978 generic.go:334] "Generic (PLEG): container finished" podID="f8752058-4f5d-40b2-a8b7-42105c8d85a2" containerID="6f96aa23a5b930fa2a3420549881cba4ee8d2839fc425259495da07c928fdf8c" exitCode=0 Feb 25 07:50:03 crc kubenswrapper[4978]: I0225 07:50:03.778794 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533430-w2682" event={"ID":"f8752058-4f5d-40b2-a8b7-42105c8d85a2","Type":"ContainerDied","Data":"6f96aa23a5b930fa2a3420549881cba4ee8d2839fc425259495da07c928fdf8c"} Feb 25 07:50:05 crc kubenswrapper[4978]: I0225 07:50:05.122786 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533430-w2682" Feb 25 07:50:05 crc kubenswrapper[4978]: I0225 07:50:05.218089 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bt6mz\" (UniqueName: \"kubernetes.io/projected/f8752058-4f5d-40b2-a8b7-42105c8d85a2-kube-api-access-bt6mz\") pod \"f8752058-4f5d-40b2-a8b7-42105c8d85a2\" (UID: \"f8752058-4f5d-40b2-a8b7-42105c8d85a2\") " Feb 25 07:50:05 crc kubenswrapper[4978]: I0225 07:50:05.224690 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8752058-4f5d-40b2-a8b7-42105c8d85a2-kube-api-access-bt6mz" (OuterVolumeSpecName: "kube-api-access-bt6mz") pod "f8752058-4f5d-40b2-a8b7-42105c8d85a2" (UID: "f8752058-4f5d-40b2-a8b7-42105c8d85a2"). InnerVolumeSpecName "kube-api-access-bt6mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:50:05 crc kubenswrapper[4978]: I0225 07:50:05.320686 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bt6mz\" (UniqueName: \"kubernetes.io/projected/f8752058-4f5d-40b2-a8b7-42105c8d85a2-kube-api-access-bt6mz\") on node \"crc\" DevicePath \"\"" Feb 25 07:50:05 crc kubenswrapper[4978]: I0225 07:50:05.798365 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533430-w2682" event={"ID":"f8752058-4f5d-40b2-a8b7-42105c8d85a2","Type":"ContainerDied","Data":"43edeb7bc90b42a9f11bdaec8a0d49cabbfce80cd695d372e8e240d6a8a5396b"} Feb 25 07:50:05 crc kubenswrapper[4978]: I0225 07:50:05.798421 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43edeb7bc90b42a9f11bdaec8a0d49cabbfce80cd695d372e8e240d6a8a5396b" Feb 25 07:50:05 crc kubenswrapper[4978]: I0225 07:50:05.798495 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533430-w2682" Feb 25 07:50:05 crc kubenswrapper[4978]: I0225 07:50:05.876141 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533424-5zk4f"] Feb 25 07:50:05 crc kubenswrapper[4978]: I0225 07:50:05.883637 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533424-5zk4f"] Feb 25 07:50:07 crc kubenswrapper[4978]: I0225 07:50:07.351723 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5e4f2fd-1af2-46f1-adc7-c76913ce6470" path="/var/lib/kubelet/pods/c5e4f2fd-1af2-46f1-adc7-c76913ce6470/volumes" Feb 25 07:50:12 crc kubenswrapper[4978]: I0225 07:50:12.327472 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:50:12 crc kubenswrapper[4978]: E0225 07:50:12.328481 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:50:23 crc kubenswrapper[4978]: I0225 07:50:23.442301 4978 scope.go:117] "RemoveContainer" containerID="7d65562f086b6c4f391e2fbeaaf09e8df3f67cd182d73f4fea37953d452b146d" Feb 25 07:50:24 crc kubenswrapper[4978]: I0225 07:50:24.328670 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:50:24 crc kubenswrapper[4978]: E0225 07:50:24.329307 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:50:35 crc kubenswrapper[4978]: I0225 07:50:35.327869 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:50:35 crc kubenswrapper[4978]: E0225 07:50:35.328827 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:50:46 crc kubenswrapper[4978]: I0225 07:50:46.328056 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:50:46 crc kubenswrapper[4978]: E0225 07:50:46.329081 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:51:00 crc kubenswrapper[4978]: I0225 07:51:00.328286 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:51:01 crc kubenswrapper[4978]: I0225 07:51:01.272852 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"f9e105f61bc1f34c83b82dd9ca3ab861d524f5a79b497ff70dc0ba62730799a2"} Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.165581 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533432-z6k75"] Feb 25 07:52:00 crc kubenswrapper[4978]: E0225 07:52:00.166853 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8752058-4f5d-40b2-a8b7-42105c8d85a2" containerName="oc" Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.166879 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8752058-4f5d-40b2-a8b7-42105c8d85a2" containerName="oc" Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.167152 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8752058-4f5d-40b2-a8b7-42105c8d85a2" containerName="oc" Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.167944 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533432-z6k75" Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.171017 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.171323 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.171346 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.180788 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533432-z6k75"] Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.262354 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpmwr\" (UniqueName: \"kubernetes.io/projected/12cfaf66-9c14-4b0e-9d84-d0aec0608250-kube-api-access-mpmwr\") pod \"auto-csr-approver-29533432-z6k75\" (UID: \"12cfaf66-9c14-4b0e-9d84-d0aec0608250\") " pod="openshift-infra/auto-csr-approver-29533432-z6k75" Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.363302 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpmwr\" (UniqueName: \"kubernetes.io/projected/12cfaf66-9c14-4b0e-9d84-d0aec0608250-kube-api-access-mpmwr\") pod \"auto-csr-approver-29533432-z6k75\" (UID: \"12cfaf66-9c14-4b0e-9d84-d0aec0608250\") " pod="openshift-infra/auto-csr-approver-29533432-z6k75" Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.386024 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpmwr\" (UniqueName: \"kubernetes.io/projected/12cfaf66-9c14-4b0e-9d84-d0aec0608250-kube-api-access-mpmwr\") pod \"auto-csr-approver-29533432-z6k75\" (UID: \"12cfaf66-9c14-4b0e-9d84-d0aec0608250\") " pod="openshift-infra/auto-csr-approver-29533432-z6k75" Feb 25 07:52:00 crc kubenswrapper[4978]: I0225 07:52:00.524749 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533432-z6k75" Feb 25 07:52:01 crc kubenswrapper[4978]: I0225 07:52:01.032731 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533432-z6k75"] Feb 25 07:52:01 crc kubenswrapper[4978]: I0225 07:52:01.048618 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 07:52:01 crc kubenswrapper[4978]: I0225 07:52:01.825739 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533432-z6k75" event={"ID":"12cfaf66-9c14-4b0e-9d84-d0aec0608250","Type":"ContainerStarted","Data":"eaba2134fd46993725cfabad4add197e3b7e8e61f92d3963fb23fc139d8874a3"} Feb 25 07:52:02 crc kubenswrapper[4978]: I0225 07:52:02.838323 4978 generic.go:334] "Generic (PLEG): container finished" podID="12cfaf66-9c14-4b0e-9d84-d0aec0608250" containerID="e5430d47390d8a19cc1fef2ac2993b6fc4f258e02f4a2c8e3e7cd49a68390151" exitCode=0 Feb 25 07:52:02 crc kubenswrapper[4978]: I0225 07:52:02.838419 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533432-z6k75" event={"ID":"12cfaf66-9c14-4b0e-9d84-d0aec0608250","Type":"ContainerDied","Data":"e5430d47390d8a19cc1fef2ac2993b6fc4f258e02f4a2c8e3e7cd49a68390151"} Feb 25 07:52:04 crc kubenswrapper[4978]: I0225 07:52:04.860976 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533432-z6k75" event={"ID":"12cfaf66-9c14-4b0e-9d84-d0aec0608250","Type":"ContainerDied","Data":"eaba2134fd46993725cfabad4add197e3b7e8e61f92d3963fb23fc139d8874a3"} Feb 25 07:52:04 crc kubenswrapper[4978]: I0225 07:52:04.861362 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaba2134fd46993725cfabad4add197e3b7e8e61f92d3963fb23fc139d8874a3" Feb 25 07:52:04 crc kubenswrapper[4978]: I0225 07:52:04.875976 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533432-z6k75" Feb 25 07:52:05 crc kubenswrapper[4978]: I0225 07:52:05.057915 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpmwr\" (UniqueName: \"kubernetes.io/projected/12cfaf66-9c14-4b0e-9d84-d0aec0608250-kube-api-access-mpmwr\") pod \"12cfaf66-9c14-4b0e-9d84-d0aec0608250\" (UID: \"12cfaf66-9c14-4b0e-9d84-d0aec0608250\") " Feb 25 07:52:05 crc kubenswrapper[4978]: I0225 07:52:05.064557 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12cfaf66-9c14-4b0e-9d84-d0aec0608250-kube-api-access-mpmwr" (OuterVolumeSpecName: "kube-api-access-mpmwr") pod "12cfaf66-9c14-4b0e-9d84-d0aec0608250" (UID: "12cfaf66-9c14-4b0e-9d84-d0aec0608250"). InnerVolumeSpecName "kube-api-access-mpmwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:52:05 crc kubenswrapper[4978]: I0225 07:52:05.159336 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpmwr\" (UniqueName: \"kubernetes.io/projected/12cfaf66-9c14-4b0e-9d84-d0aec0608250-kube-api-access-mpmwr\") on node \"crc\" DevicePath \"\"" Feb 25 07:52:05 crc kubenswrapper[4978]: I0225 07:52:05.868681 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533432-z6k75" Feb 25 07:52:05 crc kubenswrapper[4978]: I0225 07:52:05.965355 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533426-cgd2c"] Feb 25 07:52:05 crc kubenswrapper[4978]: I0225 07:52:05.972746 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533426-cgd2c"] Feb 25 07:52:07 crc kubenswrapper[4978]: I0225 07:52:07.342099 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d561355a-1c08-4a5e-8584-af557e06909d" path="/var/lib/kubelet/pods/d561355a-1c08-4a5e-8584-af557e06909d/volumes" Feb 25 07:52:23 crc kubenswrapper[4978]: I0225 07:52:23.547173 4978 scope.go:117] "RemoveContainer" containerID="2c3b0ed54ece8630241797593b43e50ab4d5cdbd0497e404712182c36ae5ce56" Feb 25 07:53:16 crc kubenswrapper[4978]: I0225 07:53:16.540213 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:53:16 crc kubenswrapper[4978]: I0225 07:53:16.540816 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:53:46 crc kubenswrapper[4978]: I0225 07:53:46.540637 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:53:46 crc kubenswrapper[4978]: I0225 07:53:46.541331 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.164666 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533434-v4sxc"] Feb 25 07:54:00 crc kubenswrapper[4978]: E0225 07:54:00.165772 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12cfaf66-9c14-4b0e-9d84-d0aec0608250" containerName="oc" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.165792 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="12cfaf66-9c14-4b0e-9d84-d0aec0608250" containerName="oc" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.166021 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="12cfaf66-9c14-4b0e-9d84-d0aec0608250" containerName="oc" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.166763 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533434-v4sxc" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.176219 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.176787 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.176861 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.178495 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533434-v4sxc"] Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.216316 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctj84\" (UniqueName: \"kubernetes.io/projected/96c8796f-7879-434f-8e7d-3e3944ae3f12-kube-api-access-ctj84\") pod \"auto-csr-approver-29533434-v4sxc\" (UID: \"96c8796f-7879-434f-8e7d-3e3944ae3f12\") " pod="openshift-infra/auto-csr-approver-29533434-v4sxc" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.318709 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctj84\" (UniqueName: \"kubernetes.io/projected/96c8796f-7879-434f-8e7d-3e3944ae3f12-kube-api-access-ctj84\") pod \"auto-csr-approver-29533434-v4sxc\" (UID: \"96c8796f-7879-434f-8e7d-3e3944ae3f12\") " pod="openshift-infra/auto-csr-approver-29533434-v4sxc" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.587541 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctj84\" (UniqueName: \"kubernetes.io/projected/96c8796f-7879-434f-8e7d-3e3944ae3f12-kube-api-access-ctj84\") pod \"auto-csr-approver-29533434-v4sxc\" (UID: \"96c8796f-7879-434f-8e7d-3e3944ae3f12\") " pod="openshift-infra/auto-csr-approver-29533434-v4sxc" Feb 25 07:54:00 crc kubenswrapper[4978]: I0225 07:54:00.804770 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533434-v4sxc" Feb 25 07:54:01 crc kubenswrapper[4978]: I0225 07:54:01.304742 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533434-v4sxc"] Feb 25 07:54:01 crc kubenswrapper[4978]: W0225 07:54:01.311246 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96c8796f_7879_434f_8e7d_3e3944ae3f12.slice/crio-e270bde3a7e53e06cb1c477034ae785bd8320b339ad1db260619f45a246115b7 WatchSource:0}: Error finding container e270bde3a7e53e06cb1c477034ae785bd8320b339ad1db260619f45a246115b7: Status 404 returned error can't find the container with id e270bde3a7e53e06cb1c477034ae785bd8320b339ad1db260619f45a246115b7 Feb 25 07:54:01 crc kubenswrapper[4978]: I0225 07:54:01.985311 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533434-v4sxc" event={"ID":"96c8796f-7879-434f-8e7d-3e3944ae3f12","Type":"ContainerStarted","Data":"e270bde3a7e53e06cb1c477034ae785bd8320b339ad1db260619f45a246115b7"} Feb 25 07:54:02 crc kubenswrapper[4978]: I0225 07:54:02.998335 4978 generic.go:334] "Generic (PLEG): container finished" podID="96c8796f-7879-434f-8e7d-3e3944ae3f12" containerID="0846736cde16b94670c9f47d51d41aa92e9d6cf636f03dfb05d9a798436b00ec" exitCode=0 Feb 25 07:54:02 crc kubenswrapper[4978]: I0225 07:54:02.998430 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533434-v4sxc" event={"ID":"96c8796f-7879-434f-8e7d-3e3944ae3f12","Type":"ContainerDied","Data":"0846736cde16b94670c9f47d51d41aa92e9d6cf636f03dfb05d9a798436b00ec"} Feb 25 07:54:04 crc kubenswrapper[4978]: I0225 07:54:04.369561 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533434-v4sxc" Feb 25 07:54:04 crc kubenswrapper[4978]: I0225 07:54:04.379896 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctj84\" (UniqueName: \"kubernetes.io/projected/96c8796f-7879-434f-8e7d-3e3944ae3f12-kube-api-access-ctj84\") pod \"96c8796f-7879-434f-8e7d-3e3944ae3f12\" (UID: \"96c8796f-7879-434f-8e7d-3e3944ae3f12\") " Feb 25 07:54:04 crc kubenswrapper[4978]: I0225 07:54:04.387338 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96c8796f-7879-434f-8e7d-3e3944ae3f12-kube-api-access-ctj84" (OuterVolumeSpecName: "kube-api-access-ctj84") pod "96c8796f-7879-434f-8e7d-3e3944ae3f12" (UID: "96c8796f-7879-434f-8e7d-3e3944ae3f12"). InnerVolumeSpecName "kube-api-access-ctj84". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:54:04 crc kubenswrapper[4978]: I0225 07:54:04.481991 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctj84\" (UniqueName: \"kubernetes.io/projected/96c8796f-7879-434f-8e7d-3e3944ae3f12-kube-api-access-ctj84\") on node \"crc\" DevicePath \"\"" Feb 25 07:54:05 crc kubenswrapper[4978]: I0225 07:54:05.018618 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533434-v4sxc" event={"ID":"96c8796f-7879-434f-8e7d-3e3944ae3f12","Type":"ContainerDied","Data":"e270bde3a7e53e06cb1c477034ae785bd8320b339ad1db260619f45a246115b7"} Feb 25 07:54:05 crc kubenswrapper[4978]: I0225 07:54:05.018664 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e270bde3a7e53e06cb1c477034ae785bd8320b339ad1db260619f45a246115b7" Feb 25 07:54:05 crc kubenswrapper[4978]: I0225 07:54:05.018728 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533434-v4sxc" Feb 25 07:54:05 crc kubenswrapper[4978]: I0225 07:54:05.458789 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533428-4vnzj"] Feb 25 07:54:05 crc kubenswrapper[4978]: I0225 07:54:05.466664 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533428-4vnzj"] Feb 25 07:54:07 crc kubenswrapper[4978]: I0225 07:54:07.342231 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="396e5a14-f984-49fd-84d6-27498a036777" path="/var/lib/kubelet/pods/396e5a14-f984-49fd-84d6-27498a036777/volumes" Feb 25 07:54:16 crc kubenswrapper[4978]: I0225 07:54:16.540278 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:54:16 crc kubenswrapper[4978]: I0225 07:54:16.541066 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:54:16 crc kubenswrapper[4978]: I0225 07:54:16.541154 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:54:16 crc kubenswrapper[4978]: I0225 07:54:16.542253 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9e105f61bc1f34c83b82dd9ca3ab861d524f5a79b497ff70dc0ba62730799a2"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:54:16 crc kubenswrapper[4978]: I0225 07:54:16.542490 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://f9e105f61bc1f34c83b82dd9ca3ab861d524f5a79b497ff70dc0ba62730799a2" gracePeriod=600 Feb 25 07:54:17 crc kubenswrapper[4978]: I0225 07:54:17.113973 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="f9e105f61bc1f34c83b82dd9ca3ab861d524f5a79b497ff70dc0ba62730799a2" exitCode=0 Feb 25 07:54:17 crc kubenswrapper[4978]: I0225 07:54:17.114091 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"f9e105f61bc1f34c83b82dd9ca3ab861d524f5a79b497ff70dc0ba62730799a2"} Feb 25 07:54:17 crc kubenswrapper[4978]: I0225 07:54:17.114682 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab"} Feb 25 07:54:17 crc kubenswrapper[4978]: I0225 07:54:17.114754 4978 scope.go:117] "RemoveContainer" containerID="8db0b5a0f63a3278a31c030f7f54df82ffd3612f4760a589450aac1c066e45d5" Feb 25 07:54:23 crc kubenswrapper[4978]: I0225 07:54:23.642873 4978 scope.go:117] "RemoveContainer" containerID="0dfb3bd24dcdbb57854e3a71cc583d4b4251b43b3e0da0d3c955bfbb7491e454" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.309764 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7tps9"] Feb 25 07:54:35 crc kubenswrapper[4978]: E0225 07:54:35.310881 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96c8796f-7879-434f-8e7d-3e3944ae3f12" containerName="oc" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.310903 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="96c8796f-7879-434f-8e7d-3e3944ae3f12" containerName="oc" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.311144 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="96c8796f-7879-434f-8e7d-3e3944ae3f12" containerName="oc" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.312964 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.320628 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tps9"] Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.502741 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-catalog-content\") pod \"redhat-marketplace-7tps9\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.502809 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-utilities\") pod \"redhat-marketplace-7tps9\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.502839 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62w8k\" (UniqueName: \"kubernetes.io/projected/6765fd49-209a-47aa-ba75-218ff26110d8-kube-api-access-62w8k\") pod \"redhat-marketplace-7tps9\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.604552 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-catalog-content\") pod \"redhat-marketplace-7tps9\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.604607 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-utilities\") pod \"redhat-marketplace-7tps9\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.604630 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62w8k\" (UniqueName: \"kubernetes.io/projected/6765fd49-209a-47aa-ba75-218ff26110d8-kube-api-access-62w8k\") pod \"redhat-marketplace-7tps9\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.605205 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-catalog-content\") pod \"redhat-marketplace-7tps9\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.605429 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-utilities\") pod \"redhat-marketplace-7tps9\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.623172 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62w8k\" (UniqueName: \"kubernetes.io/projected/6765fd49-209a-47aa-ba75-218ff26110d8-kube-api-access-62w8k\") pod \"redhat-marketplace-7tps9\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:35 crc kubenswrapper[4978]: I0225 07:54:35.631818 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:36 crc kubenswrapper[4978]: I0225 07:54:36.081048 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tps9"] Feb 25 07:54:36 crc kubenswrapper[4978]: I0225 07:54:36.302203 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tps9" event={"ID":"6765fd49-209a-47aa-ba75-218ff26110d8","Type":"ContainerStarted","Data":"c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc"} Feb 25 07:54:36 crc kubenswrapper[4978]: I0225 07:54:36.302259 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tps9" event={"ID":"6765fd49-209a-47aa-ba75-218ff26110d8","Type":"ContainerStarted","Data":"a7b6b374259894c53093a45a228fef7d2e49e96e45fe6ae19972f4ad84a8a948"} Feb 25 07:54:37 crc kubenswrapper[4978]: I0225 07:54:37.310954 4978 generic.go:334] "Generic (PLEG): container finished" podID="6765fd49-209a-47aa-ba75-218ff26110d8" containerID="c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc" exitCode=0 Feb 25 07:54:37 crc kubenswrapper[4978]: I0225 07:54:37.311683 4978 generic.go:334] "Generic (PLEG): container finished" podID="6765fd49-209a-47aa-ba75-218ff26110d8" containerID="1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4" exitCode=0 Feb 25 07:54:37 crc kubenswrapper[4978]: I0225 07:54:37.311074 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tps9" event={"ID":"6765fd49-209a-47aa-ba75-218ff26110d8","Type":"ContainerDied","Data":"c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc"} Feb 25 07:54:37 crc kubenswrapper[4978]: I0225 07:54:37.311746 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tps9" event={"ID":"6765fd49-209a-47aa-ba75-218ff26110d8","Type":"ContainerDied","Data":"1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4"} Feb 25 07:54:38 crc kubenswrapper[4978]: I0225 07:54:38.325449 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tps9" event={"ID":"6765fd49-209a-47aa-ba75-218ff26110d8","Type":"ContainerStarted","Data":"1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc"} Feb 25 07:54:38 crc kubenswrapper[4978]: I0225 07:54:38.370865 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7tps9" podStartSLOduration=1.953066678 podStartE2EDuration="3.370835203s" podCreationTimestamp="2026-02-25 07:54:35 +0000 UTC" firstStartedPulling="2026-02-25 07:54:36.304036499 +0000 UTC m=+4169.743292958" lastFinishedPulling="2026-02-25 07:54:37.721804994 +0000 UTC m=+4171.161061483" observedRunningTime="2026-02-25 07:54:38.356014033 +0000 UTC m=+4171.795270512" watchObservedRunningTime="2026-02-25 07:54:38.370835203 +0000 UTC m=+4171.810091692" Feb 25 07:54:45 crc kubenswrapper[4978]: I0225 07:54:45.632342 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:45 crc kubenswrapper[4978]: I0225 07:54:45.632884 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:45 crc kubenswrapper[4978]: I0225 07:54:45.690470 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:46 crc kubenswrapper[4978]: I0225 07:54:46.476151 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:46 crc kubenswrapper[4978]: I0225 07:54:46.543977 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tps9"] Feb 25 07:54:48 crc kubenswrapper[4978]: I0225 07:54:48.413410 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7tps9" podUID="6765fd49-209a-47aa-ba75-218ff26110d8" containerName="registry-server" containerID="cri-o://1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc" gracePeriod=2 Feb 25 07:54:48 crc kubenswrapper[4978]: I0225 07:54:48.868885 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.012249 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62w8k\" (UniqueName: \"kubernetes.io/projected/6765fd49-209a-47aa-ba75-218ff26110d8-kube-api-access-62w8k\") pod \"6765fd49-209a-47aa-ba75-218ff26110d8\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.012733 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-utilities\") pod \"6765fd49-209a-47aa-ba75-218ff26110d8\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.012808 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-catalog-content\") pod \"6765fd49-209a-47aa-ba75-218ff26110d8\" (UID: \"6765fd49-209a-47aa-ba75-218ff26110d8\") " Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.014550 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-utilities" (OuterVolumeSpecName: "utilities") pod "6765fd49-209a-47aa-ba75-218ff26110d8" (UID: "6765fd49-209a-47aa-ba75-218ff26110d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.021408 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6765fd49-209a-47aa-ba75-218ff26110d8-kube-api-access-62w8k" (OuterVolumeSpecName: "kube-api-access-62w8k") pod "6765fd49-209a-47aa-ba75-218ff26110d8" (UID: "6765fd49-209a-47aa-ba75-218ff26110d8"). InnerVolumeSpecName "kube-api-access-62w8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.058180 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6765fd49-209a-47aa-ba75-218ff26110d8" (UID: "6765fd49-209a-47aa-ba75-218ff26110d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.115418 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.115464 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6765fd49-209a-47aa-ba75-218ff26110d8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.115486 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62w8k\" (UniqueName: \"kubernetes.io/projected/6765fd49-209a-47aa-ba75-218ff26110d8-kube-api-access-62w8k\") on node \"crc\" DevicePath \"\"" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.428890 4978 generic.go:334] "Generic (PLEG): container finished" podID="6765fd49-209a-47aa-ba75-218ff26110d8" containerID="1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc" exitCode=0 Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.429237 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tps9" event={"ID":"6765fd49-209a-47aa-ba75-218ff26110d8","Type":"ContainerDied","Data":"1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc"} Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.429338 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7tps9" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.429423 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7tps9" event={"ID":"6765fd49-209a-47aa-ba75-218ff26110d8","Type":"ContainerDied","Data":"a7b6b374259894c53093a45a228fef7d2e49e96e45fe6ae19972f4ad84a8a948"} Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.429468 4978 scope.go:117] "RemoveContainer" containerID="1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.474123 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tps9"] Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.475545 4978 scope.go:117] "RemoveContainer" containerID="1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.487013 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7tps9"] Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.503214 4978 scope.go:117] "RemoveContainer" containerID="c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.552022 4978 scope.go:117] "RemoveContainer" containerID="1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc" Feb 25 07:54:49 crc kubenswrapper[4978]: E0225 07:54:49.552431 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc\": container with ID starting with 1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc not found: ID does not exist" containerID="1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.552491 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc"} err="failed to get container status \"1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc\": rpc error: code = NotFound desc = could not find container \"1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc\": container with ID starting with 1dc940ada86530665c32c3bc16c6569d29bf5399e31ce0f5429ef9849cbed3cc not found: ID does not exist" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.552532 4978 scope.go:117] "RemoveContainer" containerID="1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4" Feb 25 07:54:49 crc kubenswrapper[4978]: E0225 07:54:49.552874 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4\": container with ID starting with 1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4 not found: ID does not exist" containerID="1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.552912 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4"} err="failed to get container status \"1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4\": rpc error: code = NotFound desc = could not find container \"1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4\": container with ID starting with 1e905e83387c0347e6452ac56dd03452a6a786deaa126f52c33c99b4212ed2c4 not found: ID does not exist" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.552946 4978 scope.go:117] "RemoveContainer" containerID="c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc" Feb 25 07:54:49 crc kubenswrapper[4978]: E0225 07:54:49.553404 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc\": container with ID starting with c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc not found: ID does not exist" containerID="c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc" Feb 25 07:54:49 crc kubenswrapper[4978]: I0225 07:54:49.553451 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc"} err="failed to get container status \"c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc\": rpc error: code = NotFound desc = could not find container \"c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc\": container with ID starting with c849e51e3e5aff83fc91f378b9a31e67583ba4b8f81aaa005f9ad952c444c0dc not found: ID does not exist" Feb 25 07:54:51 crc kubenswrapper[4978]: I0225 07:54:51.343633 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6765fd49-209a-47aa-ba75-218ff26110d8" path="/var/lib/kubelet/pods/6765fd49-209a-47aa-ba75-218ff26110d8/volumes" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.166547 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533436-vssbg"] Feb 25 07:56:00 crc kubenswrapper[4978]: E0225 07:56:00.167416 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6765fd49-209a-47aa-ba75-218ff26110d8" containerName="extract-utilities" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.167437 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6765fd49-209a-47aa-ba75-218ff26110d8" containerName="extract-utilities" Feb 25 07:56:00 crc kubenswrapper[4978]: E0225 07:56:00.167453 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6765fd49-209a-47aa-ba75-218ff26110d8" containerName="extract-content" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.167463 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6765fd49-209a-47aa-ba75-218ff26110d8" containerName="extract-content" Feb 25 07:56:00 crc kubenswrapper[4978]: E0225 07:56:00.167498 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6765fd49-209a-47aa-ba75-218ff26110d8" containerName="registry-server" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.167510 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6765fd49-209a-47aa-ba75-218ff26110d8" containerName="registry-server" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.167722 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6765fd49-209a-47aa-ba75-218ff26110d8" containerName="registry-server" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.168415 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533436-vssbg" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.171085 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.171313 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.173099 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.191126 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533436-vssbg"] Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.264450 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf75j\" (UniqueName: \"kubernetes.io/projected/c7785c3b-db54-4780-ae6c-5a94dc82325a-kube-api-access-xf75j\") pod \"auto-csr-approver-29533436-vssbg\" (UID: \"c7785c3b-db54-4780-ae6c-5a94dc82325a\") " pod="openshift-infra/auto-csr-approver-29533436-vssbg" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.366617 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf75j\" (UniqueName: \"kubernetes.io/projected/c7785c3b-db54-4780-ae6c-5a94dc82325a-kube-api-access-xf75j\") pod \"auto-csr-approver-29533436-vssbg\" (UID: \"c7785c3b-db54-4780-ae6c-5a94dc82325a\") " pod="openshift-infra/auto-csr-approver-29533436-vssbg" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.687652 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf75j\" (UniqueName: \"kubernetes.io/projected/c7785c3b-db54-4780-ae6c-5a94dc82325a-kube-api-access-xf75j\") pod \"auto-csr-approver-29533436-vssbg\" (UID: \"c7785c3b-db54-4780-ae6c-5a94dc82325a\") " pod="openshift-infra/auto-csr-approver-29533436-vssbg" Feb 25 07:56:00 crc kubenswrapper[4978]: I0225 07:56:00.790165 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533436-vssbg" Feb 25 07:56:01 crc kubenswrapper[4978]: I0225 07:56:01.104010 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533436-vssbg"] Feb 25 07:56:02 crc kubenswrapper[4978]: I0225 07:56:02.089726 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533436-vssbg" event={"ID":"c7785c3b-db54-4780-ae6c-5a94dc82325a","Type":"ContainerStarted","Data":"5521e57e99d5730d04c8efc70ff726f4f52edf854fc91c53d2559e561d4f7d46"} Feb 25 07:56:03 crc kubenswrapper[4978]: I0225 07:56:03.100628 4978 generic.go:334] "Generic (PLEG): container finished" podID="c7785c3b-db54-4780-ae6c-5a94dc82325a" containerID="6dcf9c486f81a90df79fd878b640fc004dcdaccb24fdcf991ff7de9b21ebd477" exitCode=0 Feb 25 07:56:03 crc kubenswrapper[4978]: I0225 07:56:03.100694 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533436-vssbg" event={"ID":"c7785c3b-db54-4780-ae6c-5a94dc82325a","Type":"ContainerDied","Data":"6dcf9c486f81a90df79fd878b640fc004dcdaccb24fdcf991ff7de9b21ebd477"} Feb 25 07:56:04 crc kubenswrapper[4978]: I0225 07:56:04.445425 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533436-vssbg" Feb 25 07:56:04 crc kubenswrapper[4978]: I0225 07:56:04.540169 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf75j\" (UniqueName: \"kubernetes.io/projected/c7785c3b-db54-4780-ae6c-5a94dc82325a-kube-api-access-xf75j\") pod \"c7785c3b-db54-4780-ae6c-5a94dc82325a\" (UID: \"c7785c3b-db54-4780-ae6c-5a94dc82325a\") " Feb 25 07:56:04 crc kubenswrapper[4978]: I0225 07:56:04.554186 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7785c3b-db54-4780-ae6c-5a94dc82325a-kube-api-access-xf75j" (OuterVolumeSpecName: "kube-api-access-xf75j") pod "c7785c3b-db54-4780-ae6c-5a94dc82325a" (UID: "c7785c3b-db54-4780-ae6c-5a94dc82325a"). InnerVolumeSpecName "kube-api-access-xf75j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:56:04 crc kubenswrapper[4978]: I0225 07:56:04.642622 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf75j\" (UniqueName: \"kubernetes.io/projected/c7785c3b-db54-4780-ae6c-5a94dc82325a-kube-api-access-xf75j\") on node \"crc\" DevicePath \"\"" Feb 25 07:56:05 crc kubenswrapper[4978]: I0225 07:56:05.117233 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533436-vssbg" event={"ID":"c7785c3b-db54-4780-ae6c-5a94dc82325a","Type":"ContainerDied","Data":"5521e57e99d5730d04c8efc70ff726f4f52edf854fc91c53d2559e561d4f7d46"} Feb 25 07:56:05 crc kubenswrapper[4978]: I0225 07:56:05.117583 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5521e57e99d5730d04c8efc70ff726f4f52edf854fc91c53d2559e561d4f7d46" Feb 25 07:56:05 crc kubenswrapper[4978]: I0225 07:56:05.117504 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533436-vssbg" Feb 25 07:56:05 crc kubenswrapper[4978]: I0225 07:56:05.531749 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533430-w2682"] Feb 25 07:56:05 crc kubenswrapper[4978]: I0225 07:56:05.537903 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533430-w2682"] Feb 25 07:56:07 crc kubenswrapper[4978]: I0225 07:56:07.335500 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8752058-4f5d-40b2-a8b7-42105c8d85a2" path="/var/lib/kubelet/pods/f8752058-4f5d-40b2-a8b7-42105c8d85a2/volumes" Feb 25 07:56:16 crc kubenswrapper[4978]: I0225 07:56:16.540817 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:56:16 crc kubenswrapper[4978]: I0225 07:56:16.542543 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.511697 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mtwcf"] Feb 25 07:56:17 crc kubenswrapper[4978]: E0225 07:56:17.512721 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7785c3b-db54-4780-ae6c-5a94dc82325a" containerName="oc" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.512757 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7785c3b-db54-4780-ae6c-5a94dc82325a" containerName="oc" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.514240 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7785c3b-db54-4780-ae6c-5a94dc82325a" containerName="oc" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.520808 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.524663 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mtwcf"] Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.690581 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-catalog-content\") pod \"community-operators-mtwcf\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.690662 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2kb8\" (UniqueName: \"kubernetes.io/projected/d1d097cc-7cac-41b8-9c33-0be1d52e9918-kube-api-access-r2kb8\") pod \"community-operators-mtwcf\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.690732 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-utilities\") pod \"community-operators-mtwcf\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.791577 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-catalog-content\") pod \"community-operators-mtwcf\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.791653 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2kb8\" (UniqueName: \"kubernetes.io/projected/d1d097cc-7cac-41b8-9c33-0be1d52e9918-kube-api-access-r2kb8\") pod \"community-operators-mtwcf\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.791747 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-utilities\") pod \"community-operators-mtwcf\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.792204 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-catalog-content\") pod \"community-operators-mtwcf\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.792459 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-utilities\") pod \"community-operators-mtwcf\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.810410 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2kb8\" (UniqueName: \"kubernetes.io/projected/d1d097cc-7cac-41b8-9c33-0be1d52e9918-kube-api-access-r2kb8\") pod \"community-operators-mtwcf\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:17 crc kubenswrapper[4978]: I0225 07:56:17.855660 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:18 crc kubenswrapper[4978]: I0225 07:56:18.355849 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mtwcf"] Feb 25 07:56:19 crc kubenswrapper[4978]: I0225 07:56:19.238318 4978 generic.go:334] "Generic (PLEG): container finished" podID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerID="ee21fe05906659321c182ff687d4eea1413885e64e6af52f40a6ca3514e76d8f" exitCode=0 Feb 25 07:56:19 crc kubenswrapper[4978]: I0225 07:56:19.238399 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtwcf" event={"ID":"d1d097cc-7cac-41b8-9c33-0be1d52e9918","Type":"ContainerDied","Data":"ee21fe05906659321c182ff687d4eea1413885e64e6af52f40a6ca3514e76d8f"} Feb 25 07:56:19 crc kubenswrapper[4978]: I0225 07:56:19.238706 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtwcf" event={"ID":"d1d097cc-7cac-41b8-9c33-0be1d52e9918","Type":"ContainerStarted","Data":"98c41544e660bde518d03c1e9f91556c8f955c6a79d9aa811df17d0426ec55d1"} Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.283398 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5ncm4"] Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.284802 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.304278 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5ncm4"] Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.327727 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-catalog-content\") pod \"redhat-operators-5ncm4\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.327827 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-utilities\") pod \"redhat-operators-5ncm4\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.327891 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s49mr\" (UniqueName: \"kubernetes.io/projected/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-kube-api-access-s49mr\") pod \"redhat-operators-5ncm4\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.428595 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-utilities\") pod \"redhat-operators-5ncm4\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.428654 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s49mr\" (UniqueName: \"kubernetes.io/projected/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-kube-api-access-s49mr\") pod \"redhat-operators-5ncm4\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.428707 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-catalog-content\") pod \"redhat-operators-5ncm4\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.429217 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-catalog-content\") pod \"redhat-operators-5ncm4\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.430473 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-utilities\") pod \"redhat-operators-5ncm4\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.460272 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s49mr\" (UniqueName: \"kubernetes.io/projected/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-kube-api-access-s49mr\") pod \"redhat-operators-5ncm4\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:20 crc kubenswrapper[4978]: I0225 07:56:20.631920 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:21 crc kubenswrapper[4978]: I0225 07:56:21.055491 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5ncm4"] Feb 25 07:56:21 crc kubenswrapper[4978]: W0225 07:56:21.067675 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c7ee72e_0b83_4fab_b0bc_3ac7db465a4c.slice/crio-81d2c983104f54a051ec48bf2638cc36b4a3f3a9d2164d31283f83643dbf11bf WatchSource:0}: Error finding container 81d2c983104f54a051ec48bf2638cc36b4a3f3a9d2164d31283f83643dbf11bf: Status 404 returned error can't find the container with id 81d2c983104f54a051ec48bf2638cc36b4a3f3a9d2164d31283f83643dbf11bf Feb 25 07:56:21 crc kubenswrapper[4978]: I0225 07:56:21.252131 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ncm4" event={"ID":"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c","Type":"ContainerStarted","Data":"006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53"} Feb 25 07:56:21 crc kubenswrapper[4978]: I0225 07:56:21.252174 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ncm4" event={"ID":"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c","Type":"ContainerStarted","Data":"81d2c983104f54a051ec48bf2638cc36b4a3f3a9d2164d31283f83643dbf11bf"} Feb 25 07:56:22 crc kubenswrapper[4978]: I0225 07:56:22.261120 4978 generic.go:334] "Generic (PLEG): container finished" podID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerID="006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53" exitCode=0 Feb 25 07:56:22 crc kubenswrapper[4978]: I0225 07:56:22.261168 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ncm4" event={"ID":"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c","Type":"ContainerDied","Data":"006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53"} Feb 25 07:56:22 crc kubenswrapper[4978]: I0225 07:56:22.892788 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p5j75"] Feb 25 07:56:22 crc kubenswrapper[4978]: I0225 07:56:22.894936 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:22 crc kubenswrapper[4978]: I0225 07:56:22.915450 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p5j75"] Feb 25 07:56:22 crc kubenswrapper[4978]: I0225 07:56:22.968793 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-catalog-content\") pod \"certified-operators-p5j75\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:22 crc kubenswrapper[4978]: I0225 07:56:22.968923 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-utilities\") pod \"certified-operators-p5j75\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:22 crc kubenswrapper[4978]: I0225 07:56:22.968976 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wwhq\" (UniqueName: \"kubernetes.io/projected/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-kube-api-access-4wwhq\") pod \"certified-operators-p5j75\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:23 crc kubenswrapper[4978]: I0225 07:56:23.069937 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-utilities\") pod \"certified-operators-p5j75\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:23 crc kubenswrapper[4978]: I0225 07:56:23.070012 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wwhq\" (UniqueName: \"kubernetes.io/projected/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-kube-api-access-4wwhq\") pod \"certified-operators-p5j75\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:23 crc kubenswrapper[4978]: I0225 07:56:23.070062 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-catalog-content\") pod \"certified-operators-p5j75\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:23 crc kubenswrapper[4978]: I0225 07:56:23.070593 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-utilities\") pod \"certified-operators-p5j75\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:23 crc kubenswrapper[4978]: I0225 07:56:23.070712 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-catalog-content\") pod \"certified-operators-p5j75\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:23 crc kubenswrapper[4978]: I0225 07:56:23.093951 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wwhq\" (UniqueName: \"kubernetes.io/projected/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-kube-api-access-4wwhq\") pod \"certified-operators-p5j75\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:23 crc kubenswrapper[4978]: I0225 07:56:23.231502 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:23 crc kubenswrapper[4978]: I0225 07:56:23.780228 4978 scope.go:117] "RemoveContainer" containerID="6f96aa23a5b930fa2a3420549881cba4ee8d2839fc425259495da07c928fdf8c" Feb 25 07:56:24 crc kubenswrapper[4978]: I0225 07:56:24.455935 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p5j75"] Feb 25 07:56:24 crc kubenswrapper[4978]: W0225 07:56:24.459960 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06e10d6b_ca0e_42ef_8062_33ac8e820ae0.slice/crio-ee361dcac350e5f85b6d3af94ca770cd59ec15a7e5389dfffa5fcb0a009f55fb WatchSource:0}: Error finding container ee361dcac350e5f85b6d3af94ca770cd59ec15a7e5389dfffa5fcb0a009f55fb: Status 404 returned error can't find the container with id ee361dcac350e5f85b6d3af94ca770cd59ec15a7e5389dfffa5fcb0a009f55fb Feb 25 07:56:25 crc kubenswrapper[4978]: I0225 07:56:25.283876 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ncm4" event={"ID":"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c","Type":"ContainerStarted","Data":"423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce"} Feb 25 07:56:25 crc kubenswrapper[4978]: I0225 07:56:25.294624 4978 generic.go:334] "Generic (PLEG): container finished" podID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerID="138d1d06c48ea092c213ef55aaf340f0d4a712d611fae9de13cf47ce838f14c4" exitCode=0 Feb 25 07:56:25 crc kubenswrapper[4978]: I0225 07:56:25.294697 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtwcf" event={"ID":"d1d097cc-7cac-41b8-9c33-0be1d52e9918","Type":"ContainerDied","Data":"138d1d06c48ea092c213ef55aaf340f0d4a712d611fae9de13cf47ce838f14c4"} Feb 25 07:56:25 crc kubenswrapper[4978]: I0225 07:56:25.297968 4978 generic.go:334] "Generic (PLEG): container finished" podID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerID="2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4" exitCode=0 Feb 25 07:56:25 crc kubenswrapper[4978]: I0225 07:56:25.298027 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5j75" event={"ID":"06e10d6b-ca0e-42ef-8062-33ac8e820ae0","Type":"ContainerDied","Data":"2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4"} Feb 25 07:56:25 crc kubenswrapper[4978]: I0225 07:56:25.298062 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5j75" event={"ID":"06e10d6b-ca0e-42ef-8062-33ac8e820ae0","Type":"ContainerStarted","Data":"ee361dcac350e5f85b6d3af94ca770cd59ec15a7e5389dfffa5fcb0a009f55fb"} Feb 25 07:56:26 crc kubenswrapper[4978]: I0225 07:56:26.310654 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5j75" event={"ID":"06e10d6b-ca0e-42ef-8062-33ac8e820ae0","Type":"ContainerStarted","Data":"726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef"} Feb 25 07:56:26 crc kubenswrapper[4978]: I0225 07:56:26.315033 4978 generic.go:334] "Generic (PLEG): container finished" podID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerID="423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce" exitCode=0 Feb 25 07:56:26 crc kubenswrapper[4978]: I0225 07:56:26.315105 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ncm4" event={"ID":"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c","Type":"ContainerDied","Data":"423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce"} Feb 25 07:56:26 crc kubenswrapper[4978]: I0225 07:56:26.324435 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtwcf" event={"ID":"d1d097cc-7cac-41b8-9c33-0be1d52e9918","Type":"ContainerStarted","Data":"6f10dbb2be6cba00db21a4ae7cc9bd70d3550e26bb8e5fce9eed23507f8de59c"} Feb 25 07:56:26 crc kubenswrapper[4978]: I0225 07:56:26.488444 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mtwcf" podStartSLOduration=3.024260004 podStartE2EDuration="9.488423599s" podCreationTimestamp="2026-02-25 07:56:17 +0000 UTC" firstStartedPulling="2026-02-25 07:56:19.240785639 +0000 UTC m=+4272.680042108" lastFinishedPulling="2026-02-25 07:56:25.704949234 +0000 UTC m=+4279.144205703" observedRunningTime="2026-02-25 07:56:26.483659812 +0000 UTC m=+4279.922916281" watchObservedRunningTime="2026-02-25 07:56:26.488423599 +0000 UTC m=+4279.927680058" Feb 25 07:56:27 crc kubenswrapper[4978]: I0225 07:56:27.333222 4978 generic.go:334] "Generic (PLEG): container finished" podID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerID="726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef" exitCode=0 Feb 25 07:56:27 crc kubenswrapper[4978]: I0225 07:56:27.335756 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5j75" event={"ID":"06e10d6b-ca0e-42ef-8062-33ac8e820ae0","Type":"ContainerDied","Data":"726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef"} Feb 25 07:56:27 crc kubenswrapper[4978]: I0225 07:56:27.336163 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ncm4" event={"ID":"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c","Type":"ContainerStarted","Data":"8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee"} Feb 25 07:56:27 crc kubenswrapper[4978]: I0225 07:56:27.384029 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5ncm4" podStartSLOduration=4.782589056 podStartE2EDuration="7.384007359s" podCreationTimestamp="2026-02-25 07:56:20 +0000 UTC" firstStartedPulling="2026-02-25 07:56:24.1216991 +0000 UTC m=+4277.560955559" lastFinishedPulling="2026-02-25 07:56:26.723117403 +0000 UTC m=+4280.162373862" observedRunningTime="2026-02-25 07:56:27.379423657 +0000 UTC m=+4280.818680146" watchObservedRunningTime="2026-02-25 07:56:27.384007359 +0000 UTC m=+4280.823263818" Feb 25 07:56:27 crc kubenswrapper[4978]: I0225 07:56:27.856641 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:27 crc kubenswrapper[4978]: I0225 07:56:27.856699 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:28 crc kubenswrapper[4978]: I0225 07:56:28.344083 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5j75" event={"ID":"06e10d6b-ca0e-42ef-8062-33ac8e820ae0","Type":"ContainerStarted","Data":"51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710"} Feb 25 07:56:28 crc kubenswrapper[4978]: I0225 07:56:28.375607 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p5j75" podStartSLOduration=3.906295475 podStartE2EDuration="6.375591924s" podCreationTimestamp="2026-02-25 07:56:22 +0000 UTC" firstStartedPulling="2026-02-25 07:56:25.300200799 +0000 UTC m=+4278.739457278" lastFinishedPulling="2026-02-25 07:56:27.769497268 +0000 UTC m=+4281.208753727" observedRunningTime="2026-02-25 07:56:28.372011943 +0000 UTC m=+4281.811268402" watchObservedRunningTime="2026-02-25 07:56:28.375591924 +0000 UTC m=+4281.814848383" Feb 25 07:56:28 crc kubenswrapper[4978]: I0225 07:56:28.920639 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mtwcf" podUID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerName="registry-server" probeResult="failure" output=< Feb 25 07:56:28 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 07:56:28 crc kubenswrapper[4978]: > Feb 25 07:56:30 crc kubenswrapper[4978]: I0225 07:56:30.632670 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:30 crc kubenswrapper[4978]: I0225 07:56:30.632728 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:31 crc kubenswrapper[4978]: I0225 07:56:31.671032 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5ncm4" podUID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerName="registry-server" probeResult="failure" output=< Feb 25 07:56:31 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 07:56:31 crc kubenswrapper[4978]: > Feb 25 07:56:33 crc kubenswrapper[4978]: I0225 07:56:33.231760 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:33 crc kubenswrapper[4978]: I0225 07:56:33.235005 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:33 crc kubenswrapper[4978]: I0225 07:56:33.311546 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:33 crc kubenswrapper[4978]: I0225 07:56:33.458725 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:34 crc kubenswrapper[4978]: I0225 07:56:34.674313 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p5j75"] Feb 25 07:56:36 crc kubenswrapper[4978]: I0225 07:56:36.425464 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p5j75" podUID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerName="registry-server" containerID="cri-o://51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710" gracePeriod=2 Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.396151 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.441065 4978 generic.go:334] "Generic (PLEG): container finished" podID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerID="51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710" exitCode=0 Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.441117 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5j75" event={"ID":"06e10d6b-ca0e-42ef-8062-33ac8e820ae0","Type":"ContainerDied","Data":"51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710"} Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.441146 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5j75" event={"ID":"06e10d6b-ca0e-42ef-8062-33ac8e820ae0","Type":"ContainerDied","Data":"ee361dcac350e5f85b6d3af94ca770cd59ec15a7e5389dfffa5fcb0a009f55fb"} Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.441169 4978 scope.go:117] "RemoveContainer" containerID="51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.441296 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5j75" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.459691 4978 scope.go:117] "RemoveContainer" containerID="726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.477521 4978 scope.go:117] "RemoveContainer" containerID="2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.505480 4978 scope.go:117] "RemoveContainer" containerID="51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710" Feb 25 07:56:37 crc kubenswrapper[4978]: E0225 07:56:37.505944 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710\": container with ID starting with 51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710 not found: ID does not exist" containerID="51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.505998 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710"} err="failed to get container status \"51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710\": rpc error: code = NotFound desc = could not find container \"51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710\": container with ID starting with 51f8adaf392ab41a5acfbd2889d6b212c31ba4f178f75d22b714d28bcf73c710 not found: ID does not exist" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.506036 4978 scope.go:117] "RemoveContainer" containerID="726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef" Feb 25 07:56:37 crc kubenswrapper[4978]: E0225 07:56:37.506310 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef\": container with ID starting with 726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef not found: ID does not exist" containerID="726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.506335 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef"} err="failed to get container status \"726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef\": rpc error: code = NotFound desc = could not find container \"726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef\": container with ID starting with 726e509d64c880c51561b9c4924344306457215ffeba5dfd6e2235b69d22e9ef not found: ID does not exist" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.506357 4978 scope.go:117] "RemoveContainer" containerID="2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4" Feb 25 07:56:37 crc kubenswrapper[4978]: E0225 07:56:37.506779 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4\": container with ID starting with 2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4 not found: ID does not exist" containerID="2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.506843 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4"} err="failed to get container status \"2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4\": rpc error: code = NotFound desc = could not find container \"2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4\": container with ID starting with 2b3e47e46f9834dead6ef84996359abd41ea3d571b8eb979b062d4f3a477c8f4 not found: ID does not exist" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.586619 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-utilities\") pod \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.586774 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-catalog-content\") pod \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.586846 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wwhq\" (UniqueName: \"kubernetes.io/projected/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-kube-api-access-4wwhq\") pod \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\" (UID: \"06e10d6b-ca0e-42ef-8062-33ac8e820ae0\") " Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.587752 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-utilities" (OuterVolumeSpecName: "utilities") pod "06e10d6b-ca0e-42ef-8062-33ac8e820ae0" (UID: "06e10d6b-ca0e-42ef-8062-33ac8e820ae0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.595460 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-kube-api-access-4wwhq" (OuterVolumeSpecName: "kube-api-access-4wwhq") pod "06e10d6b-ca0e-42ef-8062-33ac8e820ae0" (UID: "06e10d6b-ca0e-42ef-8062-33ac8e820ae0"). InnerVolumeSpecName "kube-api-access-4wwhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.646099 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06e10d6b-ca0e-42ef-8062-33ac8e820ae0" (UID: "06e10d6b-ca0e-42ef-8062-33ac8e820ae0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.688421 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.688650 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.689141 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wwhq\" (UniqueName: \"kubernetes.io/projected/06e10d6b-ca0e-42ef-8062-33ac8e820ae0-kube-api-access-4wwhq\") on node \"crc\" DevicePath \"\"" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.780262 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p5j75"] Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.791106 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p5j75"] Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.917046 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:37 crc kubenswrapper[4978]: I0225 07:56:37.983892 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mtwcf" Feb 25 07:56:39 crc kubenswrapper[4978]: I0225 07:56:39.344992 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" path="/var/lib/kubelet/pods/06e10d6b-ca0e-42ef-8062-33ac8e820ae0/volumes" Feb 25 07:56:39 crc kubenswrapper[4978]: I0225 07:56:39.915594 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mtwcf"] Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.074051 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9qw7l"] Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.074432 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9qw7l" podUID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerName="registry-server" containerID="cri-o://f7d98c0a14d86a3e620f5bc3b576d8cecca71fb39836d9b77252c4c09e6ca202" gracePeriod=2 Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.476551 4978 generic.go:334] "Generic (PLEG): container finished" podID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerID="f7d98c0a14d86a3e620f5bc3b576d8cecca71fb39836d9b77252c4c09e6ca202" exitCode=0 Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.476906 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qw7l" event={"ID":"75fbd0de-dd88-4079-a5c9-72aff9976339","Type":"ContainerDied","Data":"f7d98c0a14d86a3e620f5bc3b576d8cecca71fb39836d9b77252c4c09e6ca202"} Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.476938 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qw7l" event={"ID":"75fbd0de-dd88-4079-a5c9-72aff9976339","Type":"ContainerDied","Data":"7248abf77ec7dfb8cbdfad57f49f813fe18f666ce7003c67ab00beb7cebb16b2"} Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.476951 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7248abf77ec7dfb8cbdfad57f49f813fe18f666ce7003c67ab00beb7cebb16b2" Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.491320 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qw7l" Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.643513 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-utilities\") pod \"75fbd0de-dd88-4079-a5c9-72aff9976339\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.643803 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-catalog-content\") pod \"75fbd0de-dd88-4079-a5c9-72aff9976339\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.643896 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4ff8\" (UniqueName: \"kubernetes.io/projected/75fbd0de-dd88-4079-a5c9-72aff9976339-kube-api-access-g4ff8\") pod \"75fbd0de-dd88-4079-a5c9-72aff9976339\" (UID: \"75fbd0de-dd88-4079-a5c9-72aff9976339\") " Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.644014 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-utilities" (OuterVolumeSpecName: "utilities") pod "75fbd0de-dd88-4079-a5c9-72aff9976339" (UID: "75fbd0de-dd88-4079-a5c9-72aff9976339"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.644544 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.649549 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75fbd0de-dd88-4079-a5c9-72aff9976339-kube-api-access-g4ff8" (OuterVolumeSpecName: "kube-api-access-g4ff8") pod "75fbd0de-dd88-4079-a5c9-72aff9976339" (UID: "75fbd0de-dd88-4079-a5c9-72aff9976339"). InnerVolumeSpecName "kube-api-access-g4ff8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.692706 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75fbd0de-dd88-4079-a5c9-72aff9976339" (UID: "75fbd0de-dd88-4079-a5c9-72aff9976339"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.699120 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.743476 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.746101 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75fbd0de-dd88-4079-a5c9-72aff9976339-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:56:40 crc kubenswrapper[4978]: I0225 07:56:40.746130 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4ff8\" (UniqueName: \"kubernetes.io/projected/75fbd0de-dd88-4079-a5c9-72aff9976339-kube-api-access-g4ff8\") on node \"crc\" DevicePath \"\"" Feb 25 07:56:41 crc kubenswrapper[4978]: I0225 07:56:41.484234 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qw7l" Feb 25 07:56:41 crc kubenswrapper[4978]: I0225 07:56:41.504916 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9qw7l"] Feb 25 07:56:41 crc kubenswrapper[4978]: I0225 07:56:41.510420 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9qw7l"] Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.079322 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5ncm4"] Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.079597 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5ncm4" podUID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerName="registry-server" containerID="cri-o://8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee" gracePeriod=2 Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.339643 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75fbd0de-dd88-4079-a5c9-72aff9976339" path="/var/lib/kubelet/pods/75fbd0de-dd88-4079-a5c9-72aff9976339/volumes" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.441322 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.501847 4978 generic.go:334] "Generic (PLEG): container finished" podID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerID="8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee" exitCode=0 Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.501891 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ncm4" event={"ID":"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c","Type":"ContainerDied","Data":"8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee"} Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.501919 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5ncm4" event={"ID":"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c","Type":"ContainerDied","Data":"81d2c983104f54a051ec48bf2638cc36b4a3f3a9d2164d31283f83643dbf11bf"} Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.501936 4978 scope.go:117] "RemoveContainer" containerID="8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.502053 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5ncm4" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.519945 4978 scope.go:117] "RemoveContainer" containerID="423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.544069 4978 scope.go:117] "RemoveContainer" containerID="006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.578014 4978 scope.go:117] "RemoveContainer" containerID="8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee" Feb 25 07:56:43 crc kubenswrapper[4978]: E0225 07:56:43.578363 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee\": container with ID starting with 8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee not found: ID does not exist" containerID="8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.578406 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee"} err="failed to get container status \"8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee\": rpc error: code = NotFound desc = could not find container \"8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee\": container with ID starting with 8f0121ef3244bef04635ad0c978358f3511551aee98d1616aae15a232ae786ee not found: ID does not exist" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.578425 4978 scope.go:117] "RemoveContainer" containerID="423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce" Feb 25 07:56:43 crc kubenswrapper[4978]: E0225 07:56:43.578785 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce\": container with ID starting with 423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce not found: ID does not exist" containerID="423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.578806 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce"} err="failed to get container status \"423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce\": rpc error: code = NotFound desc = could not find container \"423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce\": container with ID starting with 423f9a2ddb2b82cdd33cdf59b69a27686b64649b203876b4a522ad8e772b94ce not found: ID does not exist" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.578821 4978 scope.go:117] "RemoveContainer" containerID="006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53" Feb 25 07:56:43 crc kubenswrapper[4978]: E0225 07:56:43.579030 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53\": container with ID starting with 006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53 not found: ID does not exist" containerID="006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.579050 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53"} err="failed to get container status \"006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53\": rpc error: code = NotFound desc = could not find container \"006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53\": container with ID starting with 006ef8e3f74d0aef458e7a339adce18f059e86cadc798ff1e33b9b3c6a6b2e53 not found: ID does not exist" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.582537 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s49mr\" (UniqueName: \"kubernetes.io/projected/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-kube-api-access-s49mr\") pod \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.582588 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-catalog-content\") pod \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.582668 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-utilities\") pod \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\" (UID: \"0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c\") " Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.583552 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-utilities" (OuterVolumeSpecName: "utilities") pod "0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" (UID: "0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.588398 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-kube-api-access-s49mr" (OuterVolumeSpecName: "kube-api-access-s49mr") pod "0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" (UID: "0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c"). InnerVolumeSpecName "kube-api-access-s49mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.683761 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s49mr\" (UniqueName: \"kubernetes.io/projected/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-kube-api-access-s49mr\") on node \"crc\" DevicePath \"\"" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.683802 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.708545 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" (UID: "0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.784993 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.832799 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5ncm4"] Feb 25 07:56:43 crc kubenswrapper[4978]: I0225 07:56:43.840807 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5ncm4"] Feb 25 07:56:45 crc kubenswrapper[4978]: I0225 07:56:45.342425 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" path="/var/lib/kubelet/pods/0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c/volumes" Feb 25 07:56:46 crc kubenswrapper[4978]: I0225 07:56:46.540836 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:56:46 crc kubenswrapper[4978]: I0225 07:56:46.540928 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:57:16 crc kubenswrapper[4978]: I0225 07:57:16.541040 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 07:57:16 crc kubenswrapper[4978]: I0225 07:57:16.541734 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 07:57:16 crc kubenswrapper[4978]: I0225 07:57:16.541816 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 07:57:16 crc kubenswrapper[4978]: I0225 07:57:16.542626 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 07:57:16 crc kubenswrapper[4978]: I0225 07:57:16.542714 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" gracePeriod=600 Feb 25 07:57:16 crc kubenswrapper[4978]: E0225 07:57:16.692992 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:57:16 crc kubenswrapper[4978]: I0225 07:57:16.852272 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" exitCode=0 Feb 25 07:57:16 crc kubenswrapper[4978]: I0225 07:57:16.852353 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab"} Feb 25 07:57:16 crc kubenswrapper[4978]: I0225 07:57:16.852450 4978 scope.go:117] "RemoveContainer" containerID="f9e105f61bc1f34c83b82dd9ca3ab861d524f5a79b497ff70dc0ba62730799a2" Feb 25 07:57:16 crc kubenswrapper[4978]: I0225 07:57:16.853276 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:57:16 crc kubenswrapper[4978]: E0225 07:57:16.853872 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:57:24 crc kubenswrapper[4978]: I0225 07:57:24.250048 4978 scope.go:117] "RemoveContainer" containerID="f7d98c0a14d86a3e620f5bc3b576d8cecca71fb39836d9b77252c4c09e6ca202" Feb 25 07:57:24 crc kubenswrapper[4978]: I0225 07:57:24.283656 4978 scope.go:117] "RemoveContainer" containerID="c4fe110e6b4c66227e91f464c9a2f21257b98defd9e1b377a4183e36582da88d" Feb 25 07:57:24 crc kubenswrapper[4978]: I0225 07:57:24.320033 4978 scope.go:117] "RemoveContainer" containerID="99ea167ab72e2e78b8cfd016f128caf8117a21c93283c67947ca6f4c74868140" Feb 25 07:57:30 crc kubenswrapper[4978]: I0225 07:57:30.327904 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:57:30 crc kubenswrapper[4978]: E0225 07:57:30.328702 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:57:42 crc kubenswrapper[4978]: I0225 07:57:42.328527 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:57:42 crc kubenswrapper[4978]: E0225 07:57:42.329570 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:57:55 crc kubenswrapper[4978]: I0225 07:57:55.327618 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:57:55 crc kubenswrapper[4978]: E0225 07:57:55.328459 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.171323 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533438-f5lxj"] Feb 25 07:58:00 crc kubenswrapper[4978]: E0225 07:58:00.172116 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerName="extract-utilities" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172136 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerName="extract-utilities" Feb 25 07:58:00 crc kubenswrapper[4978]: E0225 07:58:00.172166 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerName="extract-content" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172180 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerName="extract-content" Feb 25 07:58:00 crc kubenswrapper[4978]: E0225 07:58:00.172204 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerName="registry-server" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172216 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerName="registry-server" Feb 25 07:58:00 crc kubenswrapper[4978]: E0225 07:58:00.172254 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerName="extract-utilities" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172266 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerName="extract-utilities" Feb 25 07:58:00 crc kubenswrapper[4978]: E0225 07:58:00.172289 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerName="extract-content" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172304 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerName="extract-content" Feb 25 07:58:00 crc kubenswrapper[4978]: E0225 07:58:00.172320 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerName="extract-content" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172331 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerName="extract-content" Feb 25 07:58:00 crc kubenswrapper[4978]: E0225 07:58:00.172357 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerName="registry-server" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172391 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerName="registry-server" Feb 25 07:58:00 crc kubenswrapper[4978]: E0225 07:58:00.172407 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerName="extract-utilities" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172418 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerName="extract-utilities" Feb 25 07:58:00 crc kubenswrapper[4978]: E0225 07:58:00.172445 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerName="registry-server" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172459 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerName="registry-server" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172694 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="75fbd0de-dd88-4079-a5c9-72aff9976339" containerName="registry-server" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172718 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c7ee72e-0b83-4fab-b0bc-3ac7db465a4c" containerName="registry-server" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.172743 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e10d6b-ca0e-42ef-8062-33ac8e820ae0" containerName="registry-server" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.174005 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533438-f5lxj" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.177624 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.177660 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.182718 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.188096 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533438-f5lxj"] Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.305050 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg4z4\" (UniqueName: \"kubernetes.io/projected/f87353da-73da-45b5-af2b-bca6106dc7dc-kube-api-access-bg4z4\") pod \"auto-csr-approver-29533438-f5lxj\" (UID: \"f87353da-73da-45b5-af2b-bca6106dc7dc\") " pod="openshift-infra/auto-csr-approver-29533438-f5lxj" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.407107 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg4z4\" (UniqueName: \"kubernetes.io/projected/f87353da-73da-45b5-af2b-bca6106dc7dc-kube-api-access-bg4z4\") pod \"auto-csr-approver-29533438-f5lxj\" (UID: \"f87353da-73da-45b5-af2b-bca6106dc7dc\") " pod="openshift-infra/auto-csr-approver-29533438-f5lxj" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.441351 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg4z4\" (UniqueName: \"kubernetes.io/projected/f87353da-73da-45b5-af2b-bca6106dc7dc-kube-api-access-bg4z4\") pod \"auto-csr-approver-29533438-f5lxj\" (UID: \"f87353da-73da-45b5-af2b-bca6106dc7dc\") " pod="openshift-infra/auto-csr-approver-29533438-f5lxj" Feb 25 07:58:00 crc kubenswrapper[4978]: I0225 07:58:00.506185 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533438-f5lxj" Feb 25 07:58:01 crc kubenswrapper[4978]: I0225 07:58:01.076492 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533438-f5lxj"] Feb 25 07:58:01 crc kubenswrapper[4978]: I0225 07:58:01.085519 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 07:58:01 crc kubenswrapper[4978]: I0225 07:58:01.342057 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533438-f5lxj" event={"ID":"f87353da-73da-45b5-af2b-bca6106dc7dc","Type":"ContainerStarted","Data":"54068631bd3ee1692fb6b02feaf898cf739821bf69e76b5ce2e4912ffa1fcdc6"} Feb 25 07:58:02 crc kubenswrapper[4978]: I0225 07:58:02.347249 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533438-f5lxj" event={"ID":"f87353da-73da-45b5-af2b-bca6106dc7dc","Type":"ContainerStarted","Data":"7217898c803478bba38491f99209628b67f2a99ff1cb505680206bbdbac88287"} Feb 25 07:58:02 crc kubenswrapper[4978]: I0225 07:58:02.375157 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533438-f5lxj" podStartSLOduration=1.502329078 podStartE2EDuration="2.37509735s" podCreationTimestamp="2026-02-25 07:58:00 +0000 UTC" firstStartedPulling="2026-02-25 07:58:01.085078805 +0000 UTC m=+4374.524335294" lastFinishedPulling="2026-02-25 07:58:01.957847067 +0000 UTC m=+4375.397103566" observedRunningTime="2026-02-25 07:58:02.371897851 +0000 UTC m=+4375.811154380" watchObservedRunningTime="2026-02-25 07:58:02.37509735 +0000 UTC m=+4375.814353849" Feb 25 07:58:03 crc kubenswrapper[4978]: I0225 07:58:03.354783 4978 generic.go:334] "Generic (PLEG): container finished" podID="f87353da-73da-45b5-af2b-bca6106dc7dc" containerID="7217898c803478bba38491f99209628b67f2a99ff1cb505680206bbdbac88287" exitCode=0 Feb 25 07:58:03 crc kubenswrapper[4978]: I0225 07:58:03.354856 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533438-f5lxj" event={"ID":"f87353da-73da-45b5-af2b-bca6106dc7dc","Type":"ContainerDied","Data":"7217898c803478bba38491f99209628b67f2a99ff1cb505680206bbdbac88287"} Feb 25 07:58:04 crc kubenswrapper[4978]: I0225 07:58:04.735950 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533438-f5lxj" Feb 25 07:58:04 crc kubenswrapper[4978]: I0225 07:58:04.876538 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg4z4\" (UniqueName: \"kubernetes.io/projected/f87353da-73da-45b5-af2b-bca6106dc7dc-kube-api-access-bg4z4\") pod \"f87353da-73da-45b5-af2b-bca6106dc7dc\" (UID: \"f87353da-73da-45b5-af2b-bca6106dc7dc\") " Feb 25 07:58:04 crc kubenswrapper[4978]: I0225 07:58:04.886260 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f87353da-73da-45b5-af2b-bca6106dc7dc-kube-api-access-bg4z4" (OuterVolumeSpecName: "kube-api-access-bg4z4") pod "f87353da-73da-45b5-af2b-bca6106dc7dc" (UID: "f87353da-73da-45b5-af2b-bca6106dc7dc"). InnerVolumeSpecName "kube-api-access-bg4z4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 07:58:04 crc kubenswrapper[4978]: I0225 07:58:04.978068 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg4z4\" (UniqueName: \"kubernetes.io/projected/f87353da-73da-45b5-af2b-bca6106dc7dc-kube-api-access-bg4z4\") on node \"crc\" DevicePath \"\"" Feb 25 07:58:05 crc kubenswrapper[4978]: I0225 07:58:05.376127 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533438-f5lxj" event={"ID":"f87353da-73da-45b5-af2b-bca6106dc7dc","Type":"ContainerDied","Data":"54068631bd3ee1692fb6b02feaf898cf739821bf69e76b5ce2e4912ffa1fcdc6"} Feb 25 07:58:05 crc kubenswrapper[4978]: I0225 07:58:05.376433 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54068631bd3ee1692fb6b02feaf898cf739821bf69e76b5ce2e4912ffa1fcdc6" Feb 25 07:58:05 crc kubenswrapper[4978]: I0225 07:58:05.376212 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533438-f5lxj" Feb 25 07:58:05 crc kubenswrapper[4978]: I0225 07:58:05.463516 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533432-z6k75"] Feb 25 07:58:05 crc kubenswrapper[4978]: I0225 07:58:05.470173 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533432-z6k75"] Feb 25 07:58:07 crc kubenswrapper[4978]: I0225 07:58:07.341549 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12cfaf66-9c14-4b0e-9d84-d0aec0608250" path="/var/lib/kubelet/pods/12cfaf66-9c14-4b0e-9d84-d0aec0608250/volumes" Feb 25 07:58:08 crc kubenswrapper[4978]: I0225 07:58:08.327328 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:58:08 crc kubenswrapper[4978]: E0225 07:58:08.327603 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:58:23 crc kubenswrapper[4978]: I0225 07:58:23.328310 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:58:23 crc kubenswrapper[4978]: E0225 07:58:23.329041 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:58:24 crc kubenswrapper[4978]: I0225 07:58:24.399082 4978 scope.go:117] "RemoveContainer" containerID="e5430d47390d8a19cc1fef2ac2993b6fc4f258e02f4a2c8e3e7cd49a68390151" Feb 25 07:58:37 crc kubenswrapper[4978]: I0225 07:58:37.338953 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:58:37 crc kubenswrapper[4978]: E0225 07:58:37.339988 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:58:48 crc kubenswrapper[4978]: I0225 07:58:48.328651 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:58:48 crc kubenswrapper[4978]: E0225 07:58:48.329585 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:59:00 crc kubenswrapper[4978]: I0225 07:59:00.328088 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:59:00 crc kubenswrapper[4978]: E0225 07:59:00.329033 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:59:11 crc kubenswrapper[4978]: I0225 07:59:11.328135 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:59:11 crc kubenswrapper[4978]: E0225 07:59:11.329315 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:59:26 crc kubenswrapper[4978]: I0225 07:59:26.328862 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:59:26 crc kubenswrapper[4978]: E0225 07:59:26.329854 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:59:40 crc kubenswrapper[4978]: I0225 07:59:40.328454 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:59:40 crc kubenswrapper[4978]: E0225 07:59:40.329568 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 07:59:52 crc kubenswrapper[4978]: I0225 07:59:52.327523 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 07:59:52 crc kubenswrapper[4978]: E0225 07:59:52.328492 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.166576 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533440-rx4fz"] Feb 25 08:00:00 crc kubenswrapper[4978]: E0225 08:00:00.167607 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f87353da-73da-45b5-af2b-bca6106dc7dc" containerName="oc" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.167628 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f87353da-73da-45b5-af2b-bca6106dc7dc" containerName="oc" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.167867 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f87353da-73da-45b5-af2b-bca6106dc7dc" containerName="oc" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.168983 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533440-rx4fz" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.171697 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.172033 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.172178 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.177673 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn"] Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.178498 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.182245 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.182709 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.207218 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn"] Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.226715 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533440-rx4fz"] Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.335796 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad8c7810-e508-400f-86de-6aae3c2b99c2-secret-volume\") pod \"collect-profiles-29533440-hvsjn\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.335943 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxttn\" (UniqueName: \"kubernetes.io/projected/ad8c7810-e508-400f-86de-6aae3c2b99c2-kube-api-access-kxttn\") pod \"collect-profiles-29533440-hvsjn\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.336246 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st9h6\" (UniqueName: \"kubernetes.io/projected/d39ce697-b043-4407-8f0a-e481842e606a-kube-api-access-st9h6\") pod \"auto-csr-approver-29533440-rx4fz\" (UID: \"d39ce697-b043-4407-8f0a-e481842e606a\") " pod="openshift-infra/auto-csr-approver-29533440-rx4fz" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.336443 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad8c7810-e508-400f-86de-6aae3c2b99c2-config-volume\") pod \"collect-profiles-29533440-hvsjn\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.437314 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st9h6\" (UniqueName: \"kubernetes.io/projected/d39ce697-b043-4407-8f0a-e481842e606a-kube-api-access-st9h6\") pod \"auto-csr-approver-29533440-rx4fz\" (UID: \"d39ce697-b043-4407-8f0a-e481842e606a\") " pod="openshift-infra/auto-csr-approver-29533440-rx4fz" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.437796 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad8c7810-e508-400f-86de-6aae3c2b99c2-config-volume\") pod \"collect-profiles-29533440-hvsjn\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.437934 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad8c7810-e508-400f-86de-6aae3c2b99c2-secret-volume\") pod \"collect-profiles-29533440-hvsjn\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.438097 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxttn\" (UniqueName: \"kubernetes.io/projected/ad8c7810-e508-400f-86de-6aae3c2b99c2-kube-api-access-kxttn\") pod \"collect-profiles-29533440-hvsjn\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.440285 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad8c7810-e508-400f-86de-6aae3c2b99c2-config-volume\") pod \"collect-profiles-29533440-hvsjn\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.445653 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad8c7810-e508-400f-86de-6aae3c2b99c2-secret-volume\") pod \"collect-profiles-29533440-hvsjn\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.469076 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxttn\" (UniqueName: \"kubernetes.io/projected/ad8c7810-e508-400f-86de-6aae3c2b99c2-kube-api-access-kxttn\") pod \"collect-profiles-29533440-hvsjn\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.469747 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st9h6\" (UniqueName: \"kubernetes.io/projected/d39ce697-b043-4407-8f0a-e481842e606a-kube-api-access-st9h6\") pod \"auto-csr-approver-29533440-rx4fz\" (UID: \"d39ce697-b043-4407-8f0a-e481842e606a\") " pod="openshift-infra/auto-csr-approver-29533440-rx4fz" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.510409 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533440-rx4fz" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.524090 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.851003 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn"] Feb 25 08:00:00 crc kubenswrapper[4978]: I0225 08:00:00.906625 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533440-rx4fz"] Feb 25 08:00:01 crc kubenswrapper[4978]: E0225 08:00:01.365987 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad8c7810_e508_400f_86de_6aae3c2b99c2.slice/crio-8aa1ba4fff420c5b41239f8e1cdcc3abf5bd6e845b1a71ad19b7fae8a1514784.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad8c7810_e508_400f_86de_6aae3c2b99c2.slice/crio-conmon-8aa1ba4fff420c5b41239f8e1cdcc3abf5bd6e845b1a71ad19b7fae8a1514784.scope\": RecentStats: unable to find data in memory cache]" Feb 25 08:00:01 crc kubenswrapper[4978]: I0225 08:00:01.461494 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533440-rx4fz" event={"ID":"d39ce697-b043-4407-8f0a-e481842e606a","Type":"ContainerStarted","Data":"77c6ff26f8c8b4c54d8aa2c2078383e304914d905cc6d81cfa282d23c98e0017"} Feb 25 08:00:01 crc kubenswrapper[4978]: I0225 08:00:01.462997 4978 generic.go:334] "Generic (PLEG): container finished" podID="ad8c7810-e508-400f-86de-6aae3c2b99c2" containerID="8aa1ba4fff420c5b41239f8e1cdcc3abf5bd6e845b1a71ad19b7fae8a1514784" exitCode=0 Feb 25 08:00:01 crc kubenswrapper[4978]: I0225 08:00:01.463038 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" event={"ID":"ad8c7810-e508-400f-86de-6aae3c2b99c2","Type":"ContainerDied","Data":"8aa1ba4fff420c5b41239f8e1cdcc3abf5bd6e845b1a71ad19b7fae8a1514784"} Feb 25 08:00:01 crc kubenswrapper[4978]: I0225 08:00:01.463066 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" event={"ID":"ad8c7810-e508-400f-86de-6aae3c2b99c2","Type":"ContainerStarted","Data":"299942d353abdaab61e033a37a307cf933350975b7f9ab301d658b968967809b"} Feb 25 08:00:02 crc kubenswrapper[4978]: I0225 08:00:02.763632 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:02 crc kubenswrapper[4978]: I0225 08:00:02.873057 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad8c7810-e508-400f-86de-6aae3c2b99c2-config-volume\") pod \"ad8c7810-e508-400f-86de-6aae3c2b99c2\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " Feb 25 08:00:02 crc kubenswrapper[4978]: I0225 08:00:02.873114 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad8c7810-e508-400f-86de-6aae3c2b99c2-secret-volume\") pod \"ad8c7810-e508-400f-86de-6aae3c2b99c2\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " Feb 25 08:00:02 crc kubenswrapper[4978]: I0225 08:00:02.873142 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxttn\" (UniqueName: \"kubernetes.io/projected/ad8c7810-e508-400f-86de-6aae3c2b99c2-kube-api-access-kxttn\") pod \"ad8c7810-e508-400f-86de-6aae3c2b99c2\" (UID: \"ad8c7810-e508-400f-86de-6aae3c2b99c2\") " Feb 25 08:00:02 crc kubenswrapper[4978]: I0225 08:00:02.873825 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad8c7810-e508-400f-86de-6aae3c2b99c2-config-volume" (OuterVolumeSpecName: "config-volume") pod "ad8c7810-e508-400f-86de-6aae3c2b99c2" (UID: "ad8c7810-e508-400f-86de-6aae3c2b99c2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:00:02 crc kubenswrapper[4978]: I0225 08:00:02.880858 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad8c7810-e508-400f-86de-6aae3c2b99c2-kube-api-access-kxttn" (OuterVolumeSpecName: "kube-api-access-kxttn") pod "ad8c7810-e508-400f-86de-6aae3c2b99c2" (UID: "ad8c7810-e508-400f-86de-6aae3c2b99c2"). InnerVolumeSpecName "kube-api-access-kxttn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:00:02 crc kubenswrapper[4978]: I0225 08:00:02.881588 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad8c7810-e508-400f-86de-6aae3c2b99c2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ad8c7810-e508-400f-86de-6aae3c2b99c2" (UID: "ad8c7810-e508-400f-86de-6aae3c2b99c2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:00:02 crc kubenswrapper[4978]: I0225 08:00:02.974735 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad8c7810-e508-400f-86de-6aae3c2b99c2-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 08:00:02 crc kubenswrapper[4978]: I0225 08:00:02.975021 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxttn\" (UniqueName: \"kubernetes.io/projected/ad8c7810-e508-400f-86de-6aae3c2b99c2-kube-api-access-kxttn\") on node \"crc\" DevicePath \"\"" Feb 25 08:00:02 crc kubenswrapper[4978]: I0225 08:00:02.975141 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad8c7810-e508-400f-86de-6aae3c2b99c2-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 08:00:03 crc kubenswrapper[4978]: I0225 08:00:03.327963 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:00:03 crc kubenswrapper[4978]: E0225 08:00:03.328290 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:00:03 crc kubenswrapper[4978]: I0225 08:00:03.482258 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" event={"ID":"ad8c7810-e508-400f-86de-6aae3c2b99c2","Type":"ContainerDied","Data":"299942d353abdaab61e033a37a307cf933350975b7f9ab301d658b968967809b"} Feb 25 08:00:03 crc kubenswrapper[4978]: I0225 08:00:03.482294 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="299942d353abdaab61e033a37a307cf933350975b7f9ab301d658b968967809b" Feb 25 08:00:03 crc kubenswrapper[4978]: I0225 08:00:03.482305 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn" Feb 25 08:00:03 crc kubenswrapper[4978]: I0225 08:00:03.859534 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6"] Feb 25 08:00:03 crc kubenswrapper[4978]: I0225 08:00:03.864864 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533395-qmcs6"] Feb 25 08:00:04 crc kubenswrapper[4978]: I0225 08:00:04.495818 4978 generic.go:334] "Generic (PLEG): container finished" podID="d39ce697-b043-4407-8f0a-e481842e606a" containerID="606c397a14c6e7e3ca5ff2984d2a4feb70ede057f97489d17ca3129fd105261a" exitCode=0 Feb 25 08:00:04 crc kubenswrapper[4978]: I0225 08:00:04.495906 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533440-rx4fz" event={"ID":"d39ce697-b043-4407-8f0a-e481842e606a","Type":"ContainerDied","Data":"606c397a14c6e7e3ca5ff2984d2a4feb70ede057f97489d17ca3129fd105261a"} Feb 25 08:00:05 crc kubenswrapper[4978]: I0225 08:00:05.345062 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fada399d-f50f-4e12-bceb-dc03a02b25c9" path="/var/lib/kubelet/pods/fada399d-f50f-4e12-bceb-dc03a02b25c9/volumes" Feb 25 08:00:05 crc kubenswrapper[4978]: I0225 08:00:05.891982 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533440-rx4fz" Feb 25 08:00:06 crc kubenswrapper[4978]: I0225 08:00:06.033354 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-st9h6\" (UniqueName: \"kubernetes.io/projected/d39ce697-b043-4407-8f0a-e481842e606a-kube-api-access-st9h6\") pod \"d39ce697-b043-4407-8f0a-e481842e606a\" (UID: \"d39ce697-b043-4407-8f0a-e481842e606a\") " Feb 25 08:00:06 crc kubenswrapper[4978]: I0225 08:00:06.043063 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d39ce697-b043-4407-8f0a-e481842e606a-kube-api-access-st9h6" (OuterVolumeSpecName: "kube-api-access-st9h6") pod "d39ce697-b043-4407-8f0a-e481842e606a" (UID: "d39ce697-b043-4407-8f0a-e481842e606a"). InnerVolumeSpecName "kube-api-access-st9h6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:00:06 crc kubenswrapper[4978]: I0225 08:00:06.135468 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-st9h6\" (UniqueName: \"kubernetes.io/projected/d39ce697-b043-4407-8f0a-e481842e606a-kube-api-access-st9h6\") on node \"crc\" DevicePath \"\"" Feb 25 08:00:06 crc kubenswrapper[4978]: I0225 08:00:06.517138 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533440-rx4fz" event={"ID":"d39ce697-b043-4407-8f0a-e481842e606a","Type":"ContainerDied","Data":"77c6ff26f8c8b4c54d8aa2c2078383e304914d905cc6d81cfa282d23c98e0017"} Feb 25 08:00:06 crc kubenswrapper[4978]: I0225 08:00:06.517201 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77c6ff26f8c8b4c54d8aa2c2078383e304914d905cc6d81cfa282d23c98e0017" Feb 25 08:00:06 crc kubenswrapper[4978]: I0225 08:00:06.517316 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533440-rx4fz" Feb 25 08:00:06 crc kubenswrapper[4978]: I0225 08:00:06.972445 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533434-v4sxc"] Feb 25 08:00:06 crc kubenswrapper[4978]: I0225 08:00:06.983013 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533434-v4sxc"] Feb 25 08:00:07 crc kubenswrapper[4978]: I0225 08:00:07.344209 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96c8796f-7879-434f-8e7d-3e3944ae3f12" path="/var/lib/kubelet/pods/96c8796f-7879-434f-8e7d-3e3944ae3f12/volumes" Feb 25 08:00:18 crc kubenswrapper[4978]: I0225 08:00:18.328166 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:00:18 crc kubenswrapper[4978]: E0225 08:00:18.329084 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:00:24 crc kubenswrapper[4978]: I0225 08:00:24.525765 4978 scope.go:117] "RemoveContainer" containerID="0846736cde16b94670c9f47d51d41aa92e9d6cf636f03dfb05d9a798436b00ec" Feb 25 08:00:24 crc kubenswrapper[4978]: I0225 08:00:24.597705 4978 scope.go:117] "RemoveContainer" containerID="a7950e92dd757d74be2aa8134bb0bfe21452f72b9cf5dfadecd2140af87a1cf7" Feb 25 08:00:33 crc kubenswrapper[4978]: I0225 08:00:33.328078 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:00:33 crc kubenswrapper[4978]: E0225 08:00:33.329022 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:00:44 crc kubenswrapper[4978]: I0225 08:00:44.327831 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:00:44 crc kubenswrapper[4978]: E0225 08:00:44.328856 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:00:59 crc kubenswrapper[4978]: I0225 08:00:59.328899 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:00:59 crc kubenswrapper[4978]: E0225 08:00:59.330208 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:01:13 crc kubenswrapper[4978]: I0225 08:01:13.328768 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:01:13 crc kubenswrapper[4978]: E0225 08:01:13.329809 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:01:27 crc kubenswrapper[4978]: I0225 08:01:27.335524 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:01:27 crc kubenswrapper[4978]: E0225 08:01:27.336924 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:01:42 crc kubenswrapper[4978]: I0225 08:01:42.328155 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:01:42 crc kubenswrapper[4978]: E0225 08:01:42.328934 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:01:56 crc kubenswrapper[4978]: I0225 08:01:56.327416 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:01:56 crc kubenswrapper[4978]: E0225 08:01:56.329621 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.159205 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533442-wx27s"] Feb 25 08:02:00 crc kubenswrapper[4978]: E0225 08:02:00.159574 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad8c7810-e508-400f-86de-6aae3c2b99c2" containerName="collect-profiles" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.159589 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad8c7810-e508-400f-86de-6aae3c2b99c2" containerName="collect-profiles" Feb 25 08:02:00 crc kubenswrapper[4978]: E0225 08:02:00.159603 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d39ce697-b043-4407-8f0a-e481842e606a" containerName="oc" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.159612 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d39ce697-b043-4407-8f0a-e481842e606a" containerName="oc" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.159773 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad8c7810-e508-400f-86de-6aae3c2b99c2" containerName="collect-profiles" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.159805 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d39ce697-b043-4407-8f0a-e481842e606a" containerName="oc" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.160284 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533442-wx27s" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.164431 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.165392 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.165637 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.179004 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533442-wx27s"] Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.279455 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfrnf\" (UniqueName: \"kubernetes.io/projected/18188410-1214-4177-83a5-210c22331997-kube-api-access-hfrnf\") pod \"auto-csr-approver-29533442-wx27s\" (UID: \"18188410-1214-4177-83a5-210c22331997\") " pod="openshift-infra/auto-csr-approver-29533442-wx27s" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.381228 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfrnf\" (UniqueName: \"kubernetes.io/projected/18188410-1214-4177-83a5-210c22331997-kube-api-access-hfrnf\") pod \"auto-csr-approver-29533442-wx27s\" (UID: \"18188410-1214-4177-83a5-210c22331997\") " pod="openshift-infra/auto-csr-approver-29533442-wx27s" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.409064 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfrnf\" (UniqueName: \"kubernetes.io/projected/18188410-1214-4177-83a5-210c22331997-kube-api-access-hfrnf\") pod \"auto-csr-approver-29533442-wx27s\" (UID: \"18188410-1214-4177-83a5-210c22331997\") " pod="openshift-infra/auto-csr-approver-29533442-wx27s" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.489791 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533442-wx27s" Feb 25 08:02:00 crc kubenswrapper[4978]: I0225 08:02:00.777041 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533442-wx27s"] Feb 25 08:02:01 crc kubenswrapper[4978]: I0225 08:02:01.599939 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533442-wx27s" event={"ID":"18188410-1214-4177-83a5-210c22331997","Type":"ContainerStarted","Data":"36738b19a5f0c3e4ee8ca44b44db3187e2398f31068e5badadb759176f6591d4"} Feb 25 08:02:02 crc kubenswrapper[4978]: I0225 08:02:02.608869 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533442-wx27s" event={"ID":"18188410-1214-4177-83a5-210c22331997","Type":"ContainerStarted","Data":"69d320b9375b16f69e47c396d05ef08887258369cd56bdb619b07612d3e620bc"} Feb 25 08:02:02 crc kubenswrapper[4978]: I0225 08:02:02.626070 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533442-wx27s" podStartSLOduration=1.206652358 podStartE2EDuration="2.626053905s" podCreationTimestamp="2026-02-25 08:02:00 +0000 UTC" firstStartedPulling="2026-02-25 08:02:00.793209258 +0000 UTC m=+4614.232465747" lastFinishedPulling="2026-02-25 08:02:02.212610785 +0000 UTC m=+4615.651867294" observedRunningTime="2026-02-25 08:02:02.622522006 +0000 UTC m=+4616.061778485" watchObservedRunningTime="2026-02-25 08:02:02.626053905 +0000 UTC m=+4616.065310364" Feb 25 08:02:03 crc kubenswrapper[4978]: I0225 08:02:03.637589 4978 generic.go:334] "Generic (PLEG): container finished" podID="18188410-1214-4177-83a5-210c22331997" containerID="69d320b9375b16f69e47c396d05ef08887258369cd56bdb619b07612d3e620bc" exitCode=0 Feb 25 08:02:03 crc kubenswrapper[4978]: I0225 08:02:03.637680 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533442-wx27s" event={"ID":"18188410-1214-4177-83a5-210c22331997","Type":"ContainerDied","Data":"69d320b9375b16f69e47c396d05ef08887258369cd56bdb619b07612d3e620bc"} Feb 25 08:02:04 crc kubenswrapper[4978]: I0225 08:02:04.966383 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533442-wx27s" Feb 25 08:02:05 crc kubenswrapper[4978]: I0225 08:02:05.158606 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfrnf\" (UniqueName: \"kubernetes.io/projected/18188410-1214-4177-83a5-210c22331997-kube-api-access-hfrnf\") pod \"18188410-1214-4177-83a5-210c22331997\" (UID: \"18188410-1214-4177-83a5-210c22331997\") " Feb 25 08:02:05 crc kubenswrapper[4978]: I0225 08:02:05.164915 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18188410-1214-4177-83a5-210c22331997-kube-api-access-hfrnf" (OuterVolumeSpecName: "kube-api-access-hfrnf") pod "18188410-1214-4177-83a5-210c22331997" (UID: "18188410-1214-4177-83a5-210c22331997"). InnerVolumeSpecName "kube-api-access-hfrnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:02:05 crc kubenswrapper[4978]: I0225 08:02:05.260531 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfrnf\" (UniqueName: \"kubernetes.io/projected/18188410-1214-4177-83a5-210c22331997-kube-api-access-hfrnf\") on node \"crc\" DevicePath \"\"" Feb 25 08:02:05 crc kubenswrapper[4978]: I0225 08:02:05.662743 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533442-wx27s" event={"ID":"18188410-1214-4177-83a5-210c22331997","Type":"ContainerDied","Data":"36738b19a5f0c3e4ee8ca44b44db3187e2398f31068e5badadb759176f6591d4"} Feb 25 08:02:05 crc kubenswrapper[4978]: I0225 08:02:05.662798 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36738b19a5f0c3e4ee8ca44b44db3187e2398f31068e5badadb759176f6591d4" Feb 25 08:02:05 crc kubenswrapper[4978]: I0225 08:02:05.662857 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533442-wx27s" Feb 25 08:02:05 crc kubenswrapper[4978]: I0225 08:02:05.698441 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533436-vssbg"] Feb 25 08:02:05 crc kubenswrapper[4978]: I0225 08:02:05.707587 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533436-vssbg"] Feb 25 08:02:07 crc kubenswrapper[4978]: I0225 08:02:07.336535 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7785c3b-db54-4780-ae6c-5a94dc82325a" path="/var/lib/kubelet/pods/c7785c3b-db54-4780-ae6c-5a94dc82325a/volumes" Feb 25 08:02:08 crc kubenswrapper[4978]: I0225 08:02:08.328207 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:02:08 crc kubenswrapper[4978]: E0225 08:02:08.328505 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:02:19 crc kubenswrapper[4978]: I0225 08:02:19.327525 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:02:19 crc kubenswrapper[4978]: I0225 08:02:19.784187 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"f9dec921ad75d0b8998d2608dca1c1f1323cda0691df3e981653b48923c773f6"} Feb 25 08:02:24 crc kubenswrapper[4978]: I0225 08:02:24.697311 4978 scope.go:117] "RemoveContainer" containerID="6dcf9c486f81a90df79fd878b640fc004dcdaccb24fdcf991ff7de9b21ebd477" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.161448 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533444-m5mhx"] Feb 25 08:04:00 crc kubenswrapper[4978]: E0225 08:04:00.162543 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18188410-1214-4177-83a5-210c22331997" containerName="oc" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.162564 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="18188410-1214-4177-83a5-210c22331997" containerName="oc" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.162790 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="18188410-1214-4177-83a5-210c22331997" containerName="oc" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.163464 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533444-m5mhx" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.167326 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.167670 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.167814 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.169895 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533444-m5mhx"] Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.225517 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tddgg\" (UniqueName: \"kubernetes.io/projected/3ff2384d-779b-4524-a068-60bc02c36c5c-kube-api-access-tddgg\") pod \"auto-csr-approver-29533444-m5mhx\" (UID: \"3ff2384d-779b-4524-a068-60bc02c36c5c\") " pod="openshift-infra/auto-csr-approver-29533444-m5mhx" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.326647 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tddgg\" (UniqueName: \"kubernetes.io/projected/3ff2384d-779b-4524-a068-60bc02c36c5c-kube-api-access-tddgg\") pod \"auto-csr-approver-29533444-m5mhx\" (UID: \"3ff2384d-779b-4524-a068-60bc02c36c5c\") " pod="openshift-infra/auto-csr-approver-29533444-m5mhx" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.361110 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tddgg\" (UniqueName: \"kubernetes.io/projected/3ff2384d-779b-4524-a068-60bc02c36c5c-kube-api-access-tddgg\") pod \"auto-csr-approver-29533444-m5mhx\" (UID: \"3ff2384d-779b-4524-a068-60bc02c36c5c\") " pod="openshift-infra/auto-csr-approver-29533444-m5mhx" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.493642 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533444-m5mhx" Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.820199 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533444-m5mhx"] Feb 25 08:04:00 crc kubenswrapper[4978]: I0225 08:04:00.832471 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 08:04:01 crc kubenswrapper[4978]: I0225 08:04:01.766756 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533444-m5mhx" event={"ID":"3ff2384d-779b-4524-a068-60bc02c36c5c","Type":"ContainerStarted","Data":"d661a6b20b63640f3ebae224b09c3e2f2aa51bea44a7b8004c22f0650e498b3a"} Feb 25 08:04:04 crc kubenswrapper[4978]: I0225 08:04:04.801344 4978 generic.go:334] "Generic (PLEG): container finished" podID="3ff2384d-779b-4524-a068-60bc02c36c5c" containerID="152f4ea577273890a1e5add29f61fd5aaa807674afaed0100a6d7e282f474d0f" exitCode=0 Feb 25 08:04:04 crc kubenswrapper[4978]: I0225 08:04:04.801461 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533444-m5mhx" event={"ID":"3ff2384d-779b-4524-a068-60bc02c36c5c","Type":"ContainerDied","Data":"152f4ea577273890a1e5add29f61fd5aaa807674afaed0100a6d7e282f474d0f"} Feb 25 08:04:06 crc kubenswrapper[4978]: I0225 08:04:06.194350 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533444-m5mhx" Feb 25 08:04:06 crc kubenswrapper[4978]: I0225 08:04:06.334841 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tddgg\" (UniqueName: \"kubernetes.io/projected/3ff2384d-779b-4524-a068-60bc02c36c5c-kube-api-access-tddgg\") pod \"3ff2384d-779b-4524-a068-60bc02c36c5c\" (UID: \"3ff2384d-779b-4524-a068-60bc02c36c5c\") " Feb 25 08:04:06 crc kubenswrapper[4978]: I0225 08:04:06.343714 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ff2384d-779b-4524-a068-60bc02c36c5c-kube-api-access-tddgg" (OuterVolumeSpecName: "kube-api-access-tddgg") pod "3ff2384d-779b-4524-a068-60bc02c36c5c" (UID: "3ff2384d-779b-4524-a068-60bc02c36c5c"). InnerVolumeSpecName "kube-api-access-tddgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:04:06 crc kubenswrapper[4978]: I0225 08:04:06.437676 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tddgg\" (UniqueName: \"kubernetes.io/projected/3ff2384d-779b-4524-a068-60bc02c36c5c-kube-api-access-tddgg\") on node \"crc\" DevicePath \"\"" Feb 25 08:04:06 crc kubenswrapper[4978]: I0225 08:04:06.821341 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533444-m5mhx" event={"ID":"3ff2384d-779b-4524-a068-60bc02c36c5c","Type":"ContainerDied","Data":"d661a6b20b63640f3ebae224b09c3e2f2aa51bea44a7b8004c22f0650e498b3a"} Feb 25 08:04:06 crc kubenswrapper[4978]: I0225 08:04:06.821770 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d661a6b20b63640f3ebae224b09c3e2f2aa51bea44a7b8004c22f0650e498b3a" Feb 25 08:04:06 crc kubenswrapper[4978]: I0225 08:04:06.821419 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533444-m5mhx" Feb 25 08:04:07 crc kubenswrapper[4978]: I0225 08:04:07.287965 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533438-f5lxj"] Feb 25 08:04:07 crc kubenswrapper[4978]: I0225 08:04:07.299806 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533438-f5lxj"] Feb 25 08:04:07 crc kubenswrapper[4978]: I0225 08:04:07.342822 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f87353da-73da-45b5-af2b-bca6106dc7dc" path="/var/lib/kubelet/pods/f87353da-73da-45b5-af2b-bca6106dc7dc/volumes" Feb 25 08:04:24 crc kubenswrapper[4978]: I0225 08:04:24.809275 4978 scope.go:117] "RemoveContainer" containerID="7217898c803478bba38491f99209628b67f2a99ff1cb505680206bbdbac88287" Feb 25 08:04:46 crc kubenswrapper[4978]: I0225 08:04:46.540465 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:04:46 crc kubenswrapper[4978]: I0225 08:04:46.541506 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:05:16 crc kubenswrapper[4978]: I0225 08:05:16.541191 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:05:16 crc kubenswrapper[4978]: I0225 08:05:16.542106 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:05:46 crc kubenswrapper[4978]: I0225 08:05:46.541005 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:05:46 crc kubenswrapper[4978]: I0225 08:05:46.541736 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:05:46 crc kubenswrapper[4978]: I0225 08:05:46.541818 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 08:05:46 crc kubenswrapper[4978]: I0225 08:05:46.542745 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9dec921ad75d0b8998d2608dca1c1f1323cda0691df3e981653b48923c773f6"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 08:05:46 crc kubenswrapper[4978]: I0225 08:05:46.542847 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://f9dec921ad75d0b8998d2608dca1c1f1323cda0691df3e981653b48923c773f6" gracePeriod=600 Feb 25 08:05:46 crc kubenswrapper[4978]: I0225 08:05:46.737925 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="f9dec921ad75d0b8998d2608dca1c1f1323cda0691df3e981653b48923c773f6" exitCode=0 Feb 25 08:05:46 crc kubenswrapper[4978]: I0225 08:05:46.738019 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"f9dec921ad75d0b8998d2608dca1c1f1323cda0691df3e981653b48923c773f6"} Feb 25 08:05:46 crc kubenswrapper[4978]: I0225 08:05:46.738084 4978 scope.go:117] "RemoveContainer" containerID="3e7b182fea093600fbb5d8b6eb7def3571f8d6fca7a9eaaa2d481a389c04c0ab" Feb 25 08:05:47 crc kubenswrapper[4978]: I0225 08:05:47.751488 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34"} Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.158795 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533446-48jcj"] Feb 25 08:06:00 crc kubenswrapper[4978]: E0225 08:06:00.159560 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ff2384d-779b-4524-a068-60bc02c36c5c" containerName="oc" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.159575 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ff2384d-779b-4524-a068-60bc02c36c5c" containerName="oc" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.159784 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ff2384d-779b-4524-a068-60bc02c36c5c" containerName="oc" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.160287 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533446-48jcj" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.162351 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.162466 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.162736 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.175272 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533446-48jcj"] Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.326493 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wds6m"] Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.338765 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8v88\" (UniqueName: \"kubernetes.io/projected/67638e0a-8134-4a91-bb00-a356a33843b0-kube-api-access-c8v88\") pod \"auto-csr-approver-29533446-48jcj\" (UID: \"67638e0a-8134-4a91-bb00-a356a33843b0\") " pod="openshift-infra/auto-csr-approver-29533446-48jcj" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.339204 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.382659 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wds6m"] Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.440745 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8v88\" (UniqueName: \"kubernetes.io/projected/67638e0a-8134-4a91-bb00-a356a33843b0-kube-api-access-c8v88\") pod \"auto-csr-approver-29533446-48jcj\" (UID: \"67638e0a-8134-4a91-bb00-a356a33843b0\") " pod="openshift-infra/auto-csr-approver-29533446-48jcj" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.440888 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8hww\" (UniqueName: \"kubernetes.io/projected/6e25f757-2727-4824-bb91-e4d02aff8738-kube-api-access-p8hww\") pod \"redhat-marketplace-wds6m\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.440978 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-catalog-content\") pod \"redhat-marketplace-wds6m\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.441026 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-utilities\") pod \"redhat-marketplace-wds6m\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.475988 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8v88\" (UniqueName: \"kubernetes.io/projected/67638e0a-8134-4a91-bb00-a356a33843b0-kube-api-access-c8v88\") pod \"auto-csr-approver-29533446-48jcj\" (UID: \"67638e0a-8134-4a91-bb00-a356a33843b0\") " pod="openshift-infra/auto-csr-approver-29533446-48jcj" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.478246 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533446-48jcj" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.542500 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8hww\" (UniqueName: \"kubernetes.io/projected/6e25f757-2727-4824-bb91-e4d02aff8738-kube-api-access-p8hww\") pod \"redhat-marketplace-wds6m\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.542610 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-catalog-content\") pod \"redhat-marketplace-wds6m\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.542662 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-utilities\") pod \"redhat-marketplace-wds6m\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.543397 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-utilities\") pod \"redhat-marketplace-wds6m\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.543615 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-catalog-content\") pod \"redhat-marketplace-wds6m\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.565914 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8hww\" (UniqueName: \"kubernetes.io/projected/6e25f757-2727-4824-bb91-e4d02aff8738-kube-api-access-p8hww\") pod \"redhat-marketplace-wds6m\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.667123 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.759696 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533446-48jcj"] Feb 25 08:06:00 crc kubenswrapper[4978]: W0225 08:06:00.767896 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67638e0a_8134_4a91_bb00_a356a33843b0.slice/crio-8462d4d9e94cd95e24b3d0e50cec4acfe305f26ec96e734262c7afd6e72634e1 WatchSource:0}: Error finding container 8462d4d9e94cd95e24b3d0e50cec4acfe305f26ec96e734262c7afd6e72634e1: Status 404 returned error can't find the container with id 8462d4d9e94cd95e24b3d0e50cec4acfe305f26ec96e734262c7afd6e72634e1 Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.867986 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533446-48jcj" event={"ID":"67638e0a-8134-4a91-bb00-a356a33843b0","Type":"ContainerStarted","Data":"8462d4d9e94cd95e24b3d0e50cec4acfe305f26ec96e734262c7afd6e72634e1"} Feb 25 08:06:00 crc kubenswrapper[4978]: I0225 08:06:00.899129 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wds6m"] Feb 25 08:06:00 crc kubenswrapper[4978]: W0225 08:06:00.903074 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e25f757_2727_4824_bb91_e4d02aff8738.slice/crio-256adfbc1561b9411761dee59ef16ea1eec4174c585937dd3015fcfb68ec5c4b WatchSource:0}: Error finding container 256adfbc1561b9411761dee59ef16ea1eec4174c585937dd3015fcfb68ec5c4b: Status 404 returned error can't find the container with id 256adfbc1561b9411761dee59ef16ea1eec4174c585937dd3015fcfb68ec5c4b Feb 25 08:06:01 crc kubenswrapper[4978]: I0225 08:06:01.880222 4978 generic.go:334] "Generic (PLEG): container finished" podID="6e25f757-2727-4824-bb91-e4d02aff8738" containerID="85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46" exitCode=0 Feb 25 08:06:01 crc kubenswrapper[4978]: I0225 08:06:01.880299 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wds6m" event={"ID":"6e25f757-2727-4824-bb91-e4d02aff8738","Type":"ContainerDied","Data":"85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46"} Feb 25 08:06:01 crc kubenswrapper[4978]: I0225 08:06:01.880575 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wds6m" event={"ID":"6e25f757-2727-4824-bb91-e4d02aff8738","Type":"ContainerStarted","Data":"256adfbc1561b9411761dee59ef16ea1eec4174c585937dd3015fcfb68ec5c4b"} Feb 25 08:06:02 crc kubenswrapper[4978]: I0225 08:06:02.892897 4978 generic.go:334] "Generic (PLEG): container finished" podID="67638e0a-8134-4a91-bb00-a356a33843b0" containerID="b7b9d5122c29649e67a9a7d62c49880b078f8d6250b9dd5e072a4b9e51327290" exitCode=0 Feb 25 08:06:02 crc kubenswrapper[4978]: I0225 08:06:02.893020 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533446-48jcj" event={"ID":"67638e0a-8134-4a91-bb00-a356a33843b0","Type":"ContainerDied","Data":"b7b9d5122c29649e67a9a7d62c49880b078f8d6250b9dd5e072a4b9e51327290"} Feb 25 08:06:03 crc kubenswrapper[4978]: I0225 08:06:03.910293 4978 generic.go:334] "Generic (PLEG): container finished" podID="6e25f757-2727-4824-bb91-e4d02aff8738" containerID="f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0" exitCode=0 Feb 25 08:06:03 crc kubenswrapper[4978]: I0225 08:06:03.910406 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wds6m" event={"ID":"6e25f757-2727-4824-bb91-e4d02aff8738","Type":"ContainerDied","Data":"f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0"} Feb 25 08:06:04 crc kubenswrapper[4978]: I0225 08:06:04.482063 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533446-48jcj" Feb 25 08:06:04 crc kubenswrapper[4978]: I0225 08:06:04.605857 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8v88\" (UniqueName: \"kubernetes.io/projected/67638e0a-8134-4a91-bb00-a356a33843b0-kube-api-access-c8v88\") pod \"67638e0a-8134-4a91-bb00-a356a33843b0\" (UID: \"67638e0a-8134-4a91-bb00-a356a33843b0\") " Feb 25 08:06:04 crc kubenswrapper[4978]: I0225 08:06:04.612576 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67638e0a-8134-4a91-bb00-a356a33843b0-kube-api-access-c8v88" (OuterVolumeSpecName: "kube-api-access-c8v88") pod "67638e0a-8134-4a91-bb00-a356a33843b0" (UID: "67638e0a-8134-4a91-bb00-a356a33843b0"). InnerVolumeSpecName "kube-api-access-c8v88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:06:04 crc kubenswrapper[4978]: I0225 08:06:04.708054 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8v88\" (UniqueName: \"kubernetes.io/projected/67638e0a-8134-4a91-bb00-a356a33843b0-kube-api-access-c8v88\") on node \"crc\" DevicePath \"\"" Feb 25 08:06:04 crc kubenswrapper[4978]: I0225 08:06:04.949557 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wds6m" event={"ID":"6e25f757-2727-4824-bb91-e4d02aff8738","Type":"ContainerStarted","Data":"6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7"} Feb 25 08:06:04 crc kubenswrapper[4978]: I0225 08:06:04.952488 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533446-48jcj" event={"ID":"67638e0a-8134-4a91-bb00-a356a33843b0","Type":"ContainerDied","Data":"8462d4d9e94cd95e24b3d0e50cec4acfe305f26ec96e734262c7afd6e72634e1"} Feb 25 08:06:04 crc kubenswrapper[4978]: I0225 08:06:04.952548 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8462d4d9e94cd95e24b3d0e50cec4acfe305f26ec96e734262c7afd6e72634e1" Feb 25 08:06:04 crc kubenswrapper[4978]: I0225 08:06:04.952645 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533446-48jcj" Feb 25 08:06:05 crc kubenswrapper[4978]: I0225 08:06:05.008079 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wds6m" podStartSLOduration=2.468506101 podStartE2EDuration="5.008047349s" podCreationTimestamp="2026-02-25 08:06:00 +0000 UTC" firstStartedPulling="2026-02-25 08:06:01.884521107 +0000 UTC m=+4855.323777576" lastFinishedPulling="2026-02-25 08:06:04.424062325 +0000 UTC m=+4857.863318824" observedRunningTime="2026-02-25 08:06:04.977085232 +0000 UTC m=+4858.416341711" watchObservedRunningTime="2026-02-25 08:06:05.008047349 +0000 UTC m=+4858.447303848" Feb 25 08:06:05 crc kubenswrapper[4978]: I0225 08:06:05.577750 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533440-rx4fz"] Feb 25 08:06:05 crc kubenswrapper[4978]: I0225 08:06:05.588235 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533440-rx4fz"] Feb 25 08:06:07 crc kubenswrapper[4978]: I0225 08:06:07.341142 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d39ce697-b043-4407-8f0a-e481842e606a" path="/var/lib/kubelet/pods/d39ce697-b043-4407-8f0a-e481842e606a/volumes" Feb 25 08:06:10 crc kubenswrapper[4978]: I0225 08:06:10.667763 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:10 crc kubenswrapper[4978]: I0225 08:06:10.668284 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:10 crc kubenswrapper[4978]: I0225 08:06:10.753795 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:11 crc kubenswrapper[4978]: I0225 08:06:11.076795 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:11 crc kubenswrapper[4978]: I0225 08:06:11.152749 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wds6m"] Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.026424 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wds6m" podUID="6e25f757-2727-4824-bb91-e4d02aff8738" containerName="registry-server" containerID="cri-o://6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7" gracePeriod=2 Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.527245 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.684848 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8hww\" (UniqueName: \"kubernetes.io/projected/6e25f757-2727-4824-bb91-e4d02aff8738-kube-api-access-p8hww\") pod \"6e25f757-2727-4824-bb91-e4d02aff8738\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.684939 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-catalog-content\") pod \"6e25f757-2727-4824-bb91-e4d02aff8738\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.685118 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-utilities\") pod \"6e25f757-2727-4824-bb91-e4d02aff8738\" (UID: \"6e25f757-2727-4824-bb91-e4d02aff8738\") " Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.686208 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-utilities" (OuterVolumeSpecName: "utilities") pod "6e25f757-2727-4824-bb91-e4d02aff8738" (UID: "6e25f757-2727-4824-bb91-e4d02aff8738"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.697449 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e25f757-2727-4824-bb91-e4d02aff8738-kube-api-access-p8hww" (OuterVolumeSpecName: "kube-api-access-p8hww") pod "6e25f757-2727-4824-bb91-e4d02aff8738" (UID: "6e25f757-2727-4824-bb91-e4d02aff8738"). InnerVolumeSpecName "kube-api-access-p8hww". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.717345 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e25f757-2727-4824-bb91-e4d02aff8738" (UID: "6e25f757-2727-4824-bb91-e4d02aff8738"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.786893 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.786931 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8hww\" (UniqueName: \"kubernetes.io/projected/6e25f757-2727-4824-bb91-e4d02aff8738-kube-api-access-p8hww\") on node \"crc\" DevicePath \"\"" Feb 25 08:06:13 crc kubenswrapper[4978]: I0225 08:06:13.786944 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e25f757-2727-4824-bb91-e4d02aff8738-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.041532 4978 generic.go:334] "Generic (PLEG): container finished" podID="6e25f757-2727-4824-bb91-e4d02aff8738" containerID="6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7" exitCode=0 Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.041598 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wds6m" event={"ID":"6e25f757-2727-4824-bb91-e4d02aff8738","Type":"ContainerDied","Data":"6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7"} Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.041650 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wds6m" event={"ID":"6e25f757-2727-4824-bb91-e4d02aff8738","Type":"ContainerDied","Data":"256adfbc1561b9411761dee59ef16ea1eec4174c585937dd3015fcfb68ec5c4b"} Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.041644 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wds6m" Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.041671 4978 scope.go:117] "RemoveContainer" containerID="6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7" Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.082034 4978 scope.go:117] "RemoveContainer" containerID="f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0" Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.105271 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wds6m"] Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.120517 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wds6m"] Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.145439 4978 scope.go:117] "RemoveContainer" containerID="85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46" Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.174501 4978 scope.go:117] "RemoveContainer" containerID="6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7" Feb 25 08:06:14 crc kubenswrapper[4978]: E0225 08:06:14.175244 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7\": container with ID starting with 6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7 not found: ID does not exist" containerID="6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7" Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.175310 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7"} err="failed to get container status \"6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7\": rpc error: code = NotFound desc = could not find container \"6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7\": container with ID starting with 6bce1e2265cfd81313b7b75155654a3901ae31733cb50c61bc635139f9a173e7 not found: ID does not exist" Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.175354 4978 scope.go:117] "RemoveContainer" containerID="f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0" Feb 25 08:06:14 crc kubenswrapper[4978]: E0225 08:06:14.175913 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0\": container with ID starting with f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0 not found: ID does not exist" containerID="f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0" Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.175964 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0"} err="failed to get container status \"f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0\": rpc error: code = NotFound desc = could not find container \"f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0\": container with ID starting with f3ad0a37d5792b1cc88a60b9ed415247d901f3c301b186f799dca7df6d0e7bc0 not found: ID does not exist" Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.175996 4978 scope.go:117] "RemoveContainer" containerID="85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46" Feb 25 08:06:14 crc kubenswrapper[4978]: E0225 08:06:14.176356 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46\": container with ID starting with 85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46 not found: ID does not exist" containerID="85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46" Feb 25 08:06:14 crc kubenswrapper[4978]: I0225 08:06:14.176422 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46"} err="failed to get container status \"85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46\": rpc error: code = NotFound desc = could not find container \"85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46\": container with ID starting with 85aa5bf4e63c0ee01e054250e10a82da8c4d15fc22cac2d59485fe0f12aa5f46 not found: ID does not exist" Feb 25 08:06:15 crc kubenswrapper[4978]: I0225 08:06:15.343260 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e25f757-2727-4824-bb91-e4d02aff8738" path="/var/lib/kubelet/pods/6e25f757-2727-4824-bb91-e4d02aff8738/volumes" Feb 25 08:06:24 crc kubenswrapper[4978]: I0225 08:06:24.924788 4978 scope.go:117] "RemoveContainer" containerID="606c397a14c6e7e3ca5ff2984d2a4feb70ede057f97489d17ca3129fd105261a" Feb 25 08:07:46 crc kubenswrapper[4978]: I0225 08:07:46.540425 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:07:46 crc kubenswrapper[4978]: I0225 08:07:46.542604 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.188666 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533448-f58x8"] Feb 25 08:08:00 crc kubenswrapper[4978]: E0225 08:08:00.189602 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e25f757-2727-4824-bb91-e4d02aff8738" containerName="extract-content" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.189619 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e25f757-2727-4824-bb91-e4d02aff8738" containerName="extract-content" Feb 25 08:08:00 crc kubenswrapper[4978]: E0225 08:08:00.189641 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67638e0a-8134-4a91-bb00-a356a33843b0" containerName="oc" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.189649 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="67638e0a-8134-4a91-bb00-a356a33843b0" containerName="oc" Feb 25 08:08:00 crc kubenswrapper[4978]: E0225 08:08:00.189668 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e25f757-2727-4824-bb91-e4d02aff8738" containerName="registry-server" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.189676 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e25f757-2727-4824-bb91-e4d02aff8738" containerName="registry-server" Feb 25 08:08:00 crc kubenswrapper[4978]: E0225 08:08:00.189694 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e25f757-2727-4824-bb91-e4d02aff8738" containerName="extract-utilities" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.189702 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e25f757-2727-4824-bb91-e4d02aff8738" containerName="extract-utilities" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.189860 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="67638e0a-8134-4a91-bb00-a356a33843b0" containerName="oc" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.189883 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e25f757-2727-4824-bb91-e4d02aff8738" containerName="registry-server" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.190899 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533448-f58x8" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.194136 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.194729 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.201384 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.201566 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533448-f58x8"] Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.376850 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxr9p\" (UniqueName: \"kubernetes.io/projected/5f47615c-d56d-45a6-a490-6cac1a4d09cb-kube-api-access-hxr9p\") pod \"auto-csr-approver-29533448-f58x8\" (UID: \"5f47615c-d56d-45a6-a490-6cac1a4d09cb\") " pod="openshift-infra/auto-csr-approver-29533448-f58x8" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.479042 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxr9p\" (UniqueName: \"kubernetes.io/projected/5f47615c-d56d-45a6-a490-6cac1a4d09cb-kube-api-access-hxr9p\") pod \"auto-csr-approver-29533448-f58x8\" (UID: \"5f47615c-d56d-45a6-a490-6cac1a4d09cb\") " pod="openshift-infra/auto-csr-approver-29533448-f58x8" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.521200 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxr9p\" (UniqueName: \"kubernetes.io/projected/5f47615c-d56d-45a6-a490-6cac1a4d09cb-kube-api-access-hxr9p\") pod \"auto-csr-approver-29533448-f58x8\" (UID: \"5f47615c-d56d-45a6-a490-6cac1a4d09cb\") " pod="openshift-infra/auto-csr-approver-29533448-f58x8" Feb 25 08:08:00 crc kubenswrapper[4978]: I0225 08:08:00.819044 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533448-f58x8" Feb 25 08:08:01 crc kubenswrapper[4978]: I0225 08:08:01.041278 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533448-f58x8"] Feb 25 08:08:02 crc kubenswrapper[4978]: I0225 08:08:02.036544 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533448-f58x8" event={"ID":"5f47615c-d56d-45a6-a490-6cac1a4d09cb","Type":"ContainerStarted","Data":"bacd1c865a872f087a090f558790421ba389cb9d68c484caf04f2b2051356251"} Feb 25 08:08:03 crc kubenswrapper[4978]: I0225 08:08:03.050875 4978 generic.go:334] "Generic (PLEG): container finished" podID="5f47615c-d56d-45a6-a490-6cac1a4d09cb" containerID="047a36f280dc81e299179e96ee49dbd63b1b67cd9a65aef29288ffd5074c87b0" exitCode=0 Feb 25 08:08:03 crc kubenswrapper[4978]: I0225 08:08:03.050982 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533448-f58x8" event={"ID":"5f47615c-d56d-45a6-a490-6cac1a4d09cb","Type":"ContainerDied","Data":"047a36f280dc81e299179e96ee49dbd63b1b67cd9a65aef29288ffd5074c87b0"} Feb 25 08:08:04 crc kubenswrapper[4978]: I0225 08:08:04.446896 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533448-f58x8" Feb 25 08:08:04 crc kubenswrapper[4978]: I0225 08:08:04.543879 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxr9p\" (UniqueName: \"kubernetes.io/projected/5f47615c-d56d-45a6-a490-6cac1a4d09cb-kube-api-access-hxr9p\") pod \"5f47615c-d56d-45a6-a490-6cac1a4d09cb\" (UID: \"5f47615c-d56d-45a6-a490-6cac1a4d09cb\") " Feb 25 08:08:04 crc kubenswrapper[4978]: I0225 08:08:04.551331 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f47615c-d56d-45a6-a490-6cac1a4d09cb-kube-api-access-hxr9p" (OuterVolumeSpecName: "kube-api-access-hxr9p") pod "5f47615c-d56d-45a6-a490-6cac1a4d09cb" (UID: "5f47615c-d56d-45a6-a490-6cac1a4d09cb"). InnerVolumeSpecName "kube-api-access-hxr9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:08:04 crc kubenswrapper[4978]: I0225 08:08:04.645578 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxr9p\" (UniqueName: \"kubernetes.io/projected/5f47615c-d56d-45a6-a490-6cac1a4d09cb-kube-api-access-hxr9p\") on node \"crc\" DevicePath \"\"" Feb 25 08:08:05 crc kubenswrapper[4978]: I0225 08:08:05.074098 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533448-f58x8" event={"ID":"5f47615c-d56d-45a6-a490-6cac1a4d09cb","Type":"ContainerDied","Data":"bacd1c865a872f087a090f558790421ba389cb9d68c484caf04f2b2051356251"} Feb 25 08:08:05 crc kubenswrapper[4978]: I0225 08:08:05.074663 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bacd1c865a872f087a090f558790421ba389cb9d68c484caf04f2b2051356251" Feb 25 08:08:05 crc kubenswrapper[4978]: I0225 08:08:05.074180 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533448-f58x8" Feb 25 08:08:05 crc kubenswrapper[4978]: I0225 08:08:05.533864 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533442-wx27s"] Feb 25 08:08:05 crc kubenswrapper[4978]: I0225 08:08:05.552197 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533442-wx27s"] Feb 25 08:08:07 crc kubenswrapper[4978]: I0225 08:08:07.347023 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18188410-1214-4177-83a5-210c22331997" path="/var/lib/kubelet/pods/18188410-1214-4177-83a5-210c22331997/volumes" Feb 25 08:08:16 crc kubenswrapper[4978]: I0225 08:08:16.541438 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:08:16 crc kubenswrapper[4978]: I0225 08:08:16.542183 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:08:25 crc kubenswrapper[4978]: I0225 08:08:25.030924 4978 scope.go:117] "RemoveContainer" containerID="69d320b9375b16f69e47c396d05ef08887258369cd56bdb619b07612d3e620bc" Feb 25 08:08:46 crc kubenswrapper[4978]: I0225 08:08:46.540438 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:08:46 crc kubenswrapper[4978]: I0225 08:08:46.541275 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:08:46 crc kubenswrapper[4978]: I0225 08:08:46.541355 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 08:08:46 crc kubenswrapper[4978]: I0225 08:08:46.542431 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 08:08:46 crc kubenswrapper[4978]: I0225 08:08:46.542533 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" gracePeriod=600 Feb 25 08:08:46 crc kubenswrapper[4978]: E0225 08:08:46.689978 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:08:47 crc kubenswrapper[4978]: I0225 08:08:47.485128 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" exitCode=0 Feb 25 08:08:47 crc kubenswrapper[4978]: I0225 08:08:47.485199 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34"} Feb 25 08:08:47 crc kubenswrapper[4978]: I0225 08:08:47.485249 4978 scope.go:117] "RemoveContainer" containerID="f9dec921ad75d0b8998d2608dca1c1f1323cda0691df3e981653b48923c773f6" Feb 25 08:08:47 crc kubenswrapper[4978]: I0225 08:08:47.486188 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:08:47 crc kubenswrapper[4978]: E0225 08:08:47.486842 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:08:59 crc kubenswrapper[4978]: I0225 08:08:59.328267 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:08:59 crc kubenswrapper[4978]: E0225 08:08:59.329772 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:09:11 crc kubenswrapper[4978]: I0225 08:09:11.327815 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:09:11 crc kubenswrapper[4978]: E0225 08:09:11.328606 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:09:25 crc kubenswrapper[4978]: I0225 08:09:25.327652 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:09:25 crc kubenswrapper[4978]: E0225 08:09:25.330673 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:09:39 crc kubenswrapper[4978]: I0225 08:09:39.328531 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:09:39 crc kubenswrapper[4978]: E0225 08:09:39.330015 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:09:54 crc kubenswrapper[4978]: I0225 08:09:54.328136 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:09:54 crc kubenswrapper[4978]: E0225 08:09:54.329412 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.169881 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533450-2zpsl"] Feb 25 08:10:00 crc kubenswrapper[4978]: E0225 08:10:00.171020 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f47615c-d56d-45a6-a490-6cac1a4d09cb" containerName="oc" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.171045 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f47615c-d56d-45a6-a490-6cac1a4d09cb" containerName="oc" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.171456 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f47615c-d56d-45a6-a490-6cac1a4d09cb" containerName="oc" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.172473 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533450-2zpsl" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.174956 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.175261 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.175761 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.182419 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533450-2zpsl"] Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.287721 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjlwj\" (UniqueName: \"kubernetes.io/projected/3e6928e0-72e4-4fea-8d60-8a5cd3514b7c-kube-api-access-vjlwj\") pod \"auto-csr-approver-29533450-2zpsl\" (UID: \"3e6928e0-72e4-4fea-8d60-8a5cd3514b7c\") " pod="openshift-infra/auto-csr-approver-29533450-2zpsl" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.389012 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjlwj\" (UniqueName: \"kubernetes.io/projected/3e6928e0-72e4-4fea-8d60-8a5cd3514b7c-kube-api-access-vjlwj\") pod \"auto-csr-approver-29533450-2zpsl\" (UID: \"3e6928e0-72e4-4fea-8d60-8a5cd3514b7c\") " pod="openshift-infra/auto-csr-approver-29533450-2zpsl" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.424635 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjlwj\" (UniqueName: \"kubernetes.io/projected/3e6928e0-72e4-4fea-8d60-8a5cd3514b7c-kube-api-access-vjlwj\") pod \"auto-csr-approver-29533450-2zpsl\" (UID: \"3e6928e0-72e4-4fea-8d60-8a5cd3514b7c\") " pod="openshift-infra/auto-csr-approver-29533450-2zpsl" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.503010 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533450-2zpsl" Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.982175 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533450-2zpsl"] Feb 25 08:10:00 crc kubenswrapper[4978]: I0225 08:10:00.993634 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 08:10:01 crc kubenswrapper[4978]: I0225 08:10:01.218988 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533450-2zpsl" event={"ID":"3e6928e0-72e4-4fea-8d60-8a5cd3514b7c","Type":"ContainerStarted","Data":"9b116211ce01e4de1d9a4289baa7fda5fec5f8c85fd399f9575818cc26c875a2"} Feb 25 08:10:03 crc kubenswrapper[4978]: I0225 08:10:03.241074 4978 generic.go:334] "Generic (PLEG): container finished" podID="3e6928e0-72e4-4fea-8d60-8a5cd3514b7c" containerID="967881ccc94747d8934b48502ab9aafe74b12c2396af224232eb915899e98c22" exitCode=0 Feb 25 08:10:03 crc kubenswrapper[4978]: I0225 08:10:03.241320 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533450-2zpsl" event={"ID":"3e6928e0-72e4-4fea-8d60-8a5cd3514b7c","Type":"ContainerDied","Data":"967881ccc94747d8934b48502ab9aafe74b12c2396af224232eb915899e98c22"} Feb 25 08:10:04 crc kubenswrapper[4978]: I0225 08:10:04.653707 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533450-2zpsl" Feb 25 08:10:04 crc kubenswrapper[4978]: I0225 08:10:04.779874 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjlwj\" (UniqueName: \"kubernetes.io/projected/3e6928e0-72e4-4fea-8d60-8a5cd3514b7c-kube-api-access-vjlwj\") pod \"3e6928e0-72e4-4fea-8d60-8a5cd3514b7c\" (UID: \"3e6928e0-72e4-4fea-8d60-8a5cd3514b7c\") " Feb 25 08:10:04 crc kubenswrapper[4978]: I0225 08:10:04.791638 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e6928e0-72e4-4fea-8d60-8a5cd3514b7c-kube-api-access-vjlwj" (OuterVolumeSpecName: "kube-api-access-vjlwj") pod "3e6928e0-72e4-4fea-8d60-8a5cd3514b7c" (UID: "3e6928e0-72e4-4fea-8d60-8a5cd3514b7c"). InnerVolumeSpecName "kube-api-access-vjlwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:10:04 crc kubenswrapper[4978]: I0225 08:10:04.882831 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjlwj\" (UniqueName: \"kubernetes.io/projected/3e6928e0-72e4-4fea-8d60-8a5cd3514b7c-kube-api-access-vjlwj\") on node \"crc\" DevicePath \"\"" Feb 25 08:10:05 crc kubenswrapper[4978]: I0225 08:10:05.269038 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533450-2zpsl" event={"ID":"3e6928e0-72e4-4fea-8d60-8a5cd3514b7c","Type":"ContainerDied","Data":"9b116211ce01e4de1d9a4289baa7fda5fec5f8c85fd399f9575818cc26c875a2"} Feb 25 08:10:05 crc kubenswrapper[4978]: I0225 08:10:05.269091 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b116211ce01e4de1d9a4289baa7fda5fec5f8c85fd399f9575818cc26c875a2" Feb 25 08:10:05 crc kubenswrapper[4978]: I0225 08:10:05.269209 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533450-2zpsl" Feb 25 08:10:05 crc kubenswrapper[4978]: I0225 08:10:05.751114 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533444-m5mhx"] Feb 25 08:10:05 crc kubenswrapper[4978]: I0225 08:10:05.761544 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533444-m5mhx"] Feb 25 08:10:07 crc kubenswrapper[4978]: I0225 08:10:07.343526 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ff2384d-779b-4524-a068-60bc02c36c5c" path="/var/lib/kubelet/pods/3ff2384d-779b-4524-a068-60bc02c36c5c/volumes" Feb 25 08:10:08 crc kubenswrapper[4978]: I0225 08:10:08.328044 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:10:08 crc kubenswrapper[4978]: E0225 08:10:08.328841 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:10:21 crc kubenswrapper[4978]: I0225 08:10:21.328419 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:10:21 crc kubenswrapper[4978]: E0225 08:10:21.329996 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:10:25 crc kubenswrapper[4978]: I0225 08:10:25.136934 4978 scope.go:117] "RemoveContainer" containerID="152f4ea577273890a1e5add29f61fd5aaa807674afaed0100a6d7e282f474d0f" Feb 25 08:10:36 crc kubenswrapper[4978]: I0225 08:10:36.327927 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:10:36 crc kubenswrapper[4978]: E0225 08:10:36.330044 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:10:49 crc kubenswrapper[4978]: I0225 08:10:49.328405 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:10:49 crc kubenswrapper[4978]: E0225 08:10:49.329345 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:11:00 crc kubenswrapper[4978]: I0225 08:11:00.327508 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:11:00 crc kubenswrapper[4978]: E0225 08:11:00.328172 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:11:14 crc kubenswrapper[4978]: I0225 08:11:14.328204 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:11:14 crc kubenswrapper[4978]: E0225 08:11:14.329089 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:11:29 crc kubenswrapper[4978]: I0225 08:11:29.328992 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:11:29 crc kubenswrapper[4978]: E0225 08:11:29.329850 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:11:42 crc kubenswrapper[4978]: I0225 08:11:42.328190 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:11:42 crc kubenswrapper[4978]: E0225 08:11:42.329191 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:11:53 crc kubenswrapper[4978]: I0225 08:11:53.328188 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:11:53 crc kubenswrapper[4978]: E0225 08:11:53.329189 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.159041 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533452-rrccs"] Feb 25 08:12:00 crc kubenswrapper[4978]: E0225 08:12:00.159940 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e6928e0-72e4-4fea-8d60-8a5cd3514b7c" containerName="oc" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.159957 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e6928e0-72e4-4fea-8d60-8a5cd3514b7c" containerName="oc" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.160158 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e6928e0-72e4-4fea-8d60-8a5cd3514b7c" containerName="oc" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.160736 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533452-rrccs" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.167495 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.167709 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.167794 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.172355 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533452-rrccs"] Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.292696 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8zgk\" (UniqueName: \"kubernetes.io/projected/f483cd75-845c-40cf-9dde-8ef1c61ee613-kube-api-access-r8zgk\") pod \"auto-csr-approver-29533452-rrccs\" (UID: \"f483cd75-845c-40cf-9dde-8ef1c61ee613\") " pod="openshift-infra/auto-csr-approver-29533452-rrccs" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.394119 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8zgk\" (UniqueName: \"kubernetes.io/projected/f483cd75-845c-40cf-9dde-8ef1c61ee613-kube-api-access-r8zgk\") pod \"auto-csr-approver-29533452-rrccs\" (UID: \"f483cd75-845c-40cf-9dde-8ef1c61ee613\") " pod="openshift-infra/auto-csr-approver-29533452-rrccs" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.431664 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8zgk\" (UniqueName: \"kubernetes.io/projected/f483cd75-845c-40cf-9dde-8ef1c61ee613-kube-api-access-r8zgk\") pod \"auto-csr-approver-29533452-rrccs\" (UID: \"f483cd75-845c-40cf-9dde-8ef1c61ee613\") " pod="openshift-infra/auto-csr-approver-29533452-rrccs" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.482459 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533452-rrccs" Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.821167 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533452-rrccs"] Feb 25 08:12:00 crc kubenswrapper[4978]: I0225 08:12:00.939001 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533452-rrccs" event={"ID":"f483cd75-845c-40cf-9dde-8ef1c61ee613","Type":"ContainerStarted","Data":"826d3a3aa1a18af58bced8b47d6fcfb3ff48c9db17fa6adc985037ecd620dbef"} Feb 25 08:12:02 crc kubenswrapper[4978]: I0225 08:12:02.963471 4978 generic.go:334] "Generic (PLEG): container finished" podID="f483cd75-845c-40cf-9dde-8ef1c61ee613" containerID="649aa004fc1a04e1abbd9c952d73f4cceb71fddbf6838243206fb376651d6e8f" exitCode=0 Feb 25 08:12:02 crc kubenswrapper[4978]: I0225 08:12:02.963568 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533452-rrccs" event={"ID":"f483cd75-845c-40cf-9dde-8ef1c61ee613","Type":"ContainerDied","Data":"649aa004fc1a04e1abbd9c952d73f4cceb71fddbf6838243206fb376651d6e8f"} Feb 25 08:12:04 crc kubenswrapper[4978]: I0225 08:12:04.321098 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533452-rrccs" Feb 25 08:12:04 crc kubenswrapper[4978]: I0225 08:12:04.458732 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8zgk\" (UniqueName: \"kubernetes.io/projected/f483cd75-845c-40cf-9dde-8ef1c61ee613-kube-api-access-r8zgk\") pod \"f483cd75-845c-40cf-9dde-8ef1c61ee613\" (UID: \"f483cd75-845c-40cf-9dde-8ef1c61ee613\") " Feb 25 08:12:04 crc kubenswrapper[4978]: I0225 08:12:04.467987 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f483cd75-845c-40cf-9dde-8ef1c61ee613-kube-api-access-r8zgk" (OuterVolumeSpecName: "kube-api-access-r8zgk") pod "f483cd75-845c-40cf-9dde-8ef1c61ee613" (UID: "f483cd75-845c-40cf-9dde-8ef1c61ee613"). InnerVolumeSpecName "kube-api-access-r8zgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:12:04 crc kubenswrapper[4978]: I0225 08:12:04.562133 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8zgk\" (UniqueName: \"kubernetes.io/projected/f483cd75-845c-40cf-9dde-8ef1c61ee613-kube-api-access-r8zgk\") on node \"crc\" DevicePath \"\"" Feb 25 08:12:04 crc kubenswrapper[4978]: I0225 08:12:04.985324 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533452-rrccs" event={"ID":"f483cd75-845c-40cf-9dde-8ef1c61ee613","Type":"ContainerDied","Data":"826d3a3aa1a18af58bced8b47d6fcfb3ff48c9db17fa6adc985037ecd620dbef"} Feb 25 08:12:04 crc kubenswrapper[4978]: I0225 08:12:04.985432 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="826d3a3aa1a18af58bced8b47d6fcfb3ff48c9db17fa6adc985037ecd620dbef" Feb 25 08:12:04 crc kubenswrapper[4978]: I0225 08:12:04.985345 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533452-rrccs" Feb 25 08:12:05 crc kubenswrapper[4978]: I0225 08:12:05.414503 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533446-48jcj"] Feb 25 08:12:05 crc kubenswrapper[4978]: I0225 08:12:05.424180 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533446-48jcj"] Feb 25 08:12:07 crc kubenswrapper[4978]: I0225 08:12:07.333583 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:12:07 crc kubenswrapper[4978]: E0225 08:12:07.334271 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:12:07 crc kubenswrapper[4978]: I0225 08:12:07.343858 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67638e0a-8134-4a91-bb00-a356a33843b0" path="/var/lib/kubelet/pods/67638e0a-8134-4a91-bb00-a356a33843b0/volumes" Feb 25 08:12:20 crc kubenswrapper[4978]: I0225 08:12:20.328036 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:12:20 crc kubenswrapper[4978]: E0225 08:12:20.328973 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:12:25 crc kubenswrapper[4978]: I0225 08:12:25.238026 4978 scope.go:117] "RemoveContainer" containerID="b7b9d5122c29649e67a9a7d62c49880b078f8d6250b9dd5e072a4b9e51327290" Feb 25 08:12:31 crc kubenswrapper[4978]: I0225 08:12:31.328235 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:12:31 crc kubenswrapper[4978]: E0225 08:12:31.328896 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:12:36 crc kubenswrapper[4978]: I0225 08:12:36.858202 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b84cj"] Feb 25 08:12:36 crc kubenswrapper[4978]: E0225 08:12:36.859047 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f483cd75-845c-40cf-9dde-8ef1c61ee613" containerName="oc" Feb 25 08:12:36 crc kubenswrapper[4978]: I0225 08:12:36.859069 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f483cd75-845c-40cf-9dde-8ef1c61ee613" containerName="oc" Feb 25 08:12:36 crc kubenswrapper[4978]: I0225 08:12:36.859343 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f483cd75-845c-40cf-9dde-8ef1c61ee613" containerName="oc" Feb 25 08:12:36 crc kubenswrapper[4978]: I0225 08:12:36.861056 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:36 crc kubenswrapper[4978]: I0225 08:12:36.880274 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b84cj"] Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.020541 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-utilities\") pod \"redhat-operators-b84cj\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.020598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptqdz\" (UniqueName: \"kubernetes.io/projected/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-kube-api-access-ptqdz\") pod \"redhat-operators-b84cj\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.021209 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-catalog-content\") pod \"redhat-operators-b84cj\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.122388 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-catalog-content\") pod \"redhat-operators-b84cj\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.122496 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-utilities\") pod \"redhat-operators-b84cj\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.122535 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptqdz\" (UniqueName: \"kubernetes.io/projected/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-kube-api-access-ptqdz\") pod \"redhat-operators-b84cj\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.123171 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-catalog-content\") pod \"redhat-operators-b84cj\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.123200 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-utilities\") pod \"redhat-operators-b84cj\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.146289 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptqdz\" (UniqueName: \"kubernetes.io/projected/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-kube-api-access-ptqdz\") pod \"redhat-operators-b84cj\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.233327 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:37 crc kubenswrapper[4978]: I0225 08:12:37.715071 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b84cj"] Feb 25 08:12:37 crc kubenswrapper[4978]: W0225 08:12:37.723488 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9ac1f1f_f52e_45a8_9a0a_2bfb6f176537.slice/crio-7841f86c7e14a4900d5968acdf4cef8d915765240feae172eedf72f5aba6a985 WatchSource:0}: Error finding container 7841f86c7e14a4900d5968acdf4cef8d915765240feae172eedf72f5aba6a985: Status 404 returned error can't find the container with id 7841f86c7e14a4900d5968acdf4cef8d915765240feae172eedf72f5aba6a985 Feb 25 08:12:38 crc kubenswrapper[4978]: I0225 08:12:38.284910 4978 generic.go:334] "Generic (PLEG): container finished" podID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerID="5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6" exitCode=0 Feb 25 08:12:38 crc kubenswrapper[4978]: I0225 08:12:38.284972 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b84cj" event={"ID":"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537","Type":"ContainerDied","Data":"5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6"} Feb 25 08:12:38 crc kubenswrapper[4978]: I0225 08:12:38.285270 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b84cj" event={"ID":"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537","Type":"ContainerStarted","Data":"7841f86c7e14a4900d5968acdf4cef8d915765240feae172eedf72f5aba6a985"} Feb 25 08:12:39 crc kubenswrapper[4978]: I0225 08:12:39.294165 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b84cj" event={"ID":"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537","Type":"ContainerStarted","Data":"6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf"} Feb 25 08:12:40 crc kubenswrapper[4978]: I0225 08:12:40.304693 4978 generic.go:334] "Generic (PLEG): container finished" podID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerID="6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf" exitCode=0 Feb 25 08:12:40 crc kubenswrapper[4978]: I0225 08:12:40.304800 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b84cj" event={"ID":"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537","Type":"ContainerDied","Data":"6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf"} Feb 25 08:12:40 crc kubenswrapper[4978]: I0225 08:12:40.856530 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tshsf"] Feb 25 08:12:40 crc kubenswrapper[4978]: I0225 08:12:40.857966 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:40 crc kubenswrapper[4978]: I0225 08:12:40.865753 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tshsf"] Feb 25 08:12:40 crc kubenswrapper[4978]: I0225 08:12:40.986902 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-utilities\") pod \"certified-operators-tshsf\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:40 crc kubenswrapper[4978]: I0225 08:12:40.987068 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmgpq\" (UniqueName: \"kubernetes.io/projected/3908c4a7-4880-47ee-83fa-0d68a57d589a-kube-api-access-bmgpq\") pod \"certified-operators-tshsf\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:40 crc kubenswrapper[4978]: I0225 08:12:40.987160 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-catalog-content\") pod \"certified-operators-tshsf\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.089098 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-catalog-content\") pod \"certified-operators-tshsf\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.089179 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-utilities\") pod \"certified-operators-tshsf\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.089256 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmgpq\" (UniqueName: \"kubernetes.io/projected/3908c4a7-4880-47ee-83fa-0d68a57d589a-kube-api-access-bmgpq\") pod \"certified-operators-tshsf\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.090520 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-catalog-content\") pod \"certified-operators-tshsf\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.090573 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-utilities\") pod \"certified-operators-tshsf\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.116010 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmgpq\" (UniqueName: \"kubernetes.io/projected/3908c4a7-4880-47ee-83fa-0d68a57d589a-kube-api-access-bmgpq\") pod \"certified-operators-tshsf\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.196628 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.314681 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b84cj" event={"ID":"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537","Type":"ContainerStarted","Data":"58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775"} Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.353681 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b84cj" podStartSLOduration=2.903822016 podStartE2EDuration="5.353667507s" podCreationTimestamp="2026-02-25 08:12:36 +0000 UTC" firstStartedPulling="2026-02-25 08:12:38.287662301 +0000 UTC m=+5251.726918780" lastFinishedPulling="2026-02-25 08:12:40.737507812 +0000 UTC m=+5254.176764271" observedRunningTime="2026-02-25 08:12:41.351903962 +0000 UTC m=+5254.791160431" watchObservedRunningTime="2026-02-25 08:12:41.353667507 +0000 UTC m=+5254.792923956" Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.531066 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tshsf"] Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.837573 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9qlnl"] Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.839585 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:41 crc kubenswrapper[4978]: I0225 08:12:41.853774 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9qlnl"] Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.009893 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfqn2\" (UniqueName: \"kubernetes.io/projected/d75f7033-ac22-4627-a53a-615b153de379-kube-api-access-cfqn2\") pod \"community-operators-9qlnl\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.009984 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-catalog-content\") pod \"community-operators-9qlnl\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.010028 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-utilities\") pod \"community-operators-9qlnl\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.111179 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-utilities\") pod \"community-operators-9qlnl\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.111276 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfqn2\" (UniqueName: \"kubernetes.io/projected/d75f7033-ac22-4627-a53a-615b153de379-kube-api-access-cfqn2\") pod \"community-operators-9qlnl\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.111334 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-catalog-content\") pod \"community-operators-9qlnl\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.111863 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-catalog-content\") pod \"community-operators-9qlnl\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.111870 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-utilities\") pod \"community-operators-9qlnl\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.136763 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfqn2\" (UniqueName: \"kubernetes.io/projected/d75f7033-ac22-4627-a53a-615b153de379-kube-api-access-cfqn2\") pod \"community-operators-9qlnl\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.153061 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.324307 4978 generic.go:334] "Generic (PLEG): container finished" podID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerID="341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a" exitCode=0 Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.325032 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshsf" event={"ID":"3908c4a7-4880-47ee-83fa-0d68a57d589a","Type":"ContainerDied","Data":"341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a"} Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.325068 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshsf" event={"ID":"3908c4a7-4880-47ee-83fa-0d68a57d589a","Type":"ContainerStarted","Data":"f91875571e21754528073e73f5411f2f80212b46a81e5c16a563d3e3d223f388"} Feb 25 08:12:42 crc kubenswrapper[4978]: I0225 08:12:42.647692 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9qlnl"] Feb 25 08:12:43 crc kubenswrapper[4978]: I0225 08:12:43.327493 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:12:43 crc kubenswrapper[4978]: E0225 08:12:43.328785 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:12:43 crc kubenswrapper[4978]: I0225 08:12:43.334913 4978 generic.go:334] "Generic (PLEG): container finished" podID="d75f7033-ac22-4627-a53a-615b153de379" containerID="6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51" exitCode=0 Feb 25 08:12:43 crc kubenswrapper[4978]: I0225 08:12:43.343134 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshsf" event={"ID":"3908c4a7-4880-47ee-83fa-0d68a57d589a","Type":"ContainerStarted","Data":"5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d"} Feb 25 08:12:43 crc kubenswrapper[4978]: I0225 08:12:43.343171 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qlnl" event={"ID":"d75f7033-ac22-4627-a53a-615b153de379","Type":"ContainerDied","Data":"6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51"} Feb 25 08:12:43 crc kubenswrapper[4978]: I0225 08:12:43.343185 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qlnl" event={"ID":"d75f7033-ac22-4627-a53a-615b153de379","Type":"ContainerStarted","Data":"4fcdc778968bfa3d7aadeeaf94cc9abaaafaeb3700f41e7aae5133591b426179"} Feb 25 08:12:44 crc kubenswrapper[4978]: I0225 08:12:44.345837 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qlnl" event={"ID":"d75f7033-ac22-4627-a53a-615b153de379","Type":"ContainerStarted","Data":"d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3"} Feb 25 08:12:44 crc kubenswrapper[4978]: I0225 08:12:44.350630 4978 generic.go:334] "Generic (PLEG): container finished" podID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerID="5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d" exitCode=0 Feb 25 08:12:44 crc kubenswrapper[4978]: I0225 08:12:44.350781 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshsf" event={"ID":"3908c4a7-4880-47ee-83fa-0d68a57d589a","Type":"ContainerDied","Data":"5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d"} Feb 25 08:12:45 crc kubenswrapper[4978]: I0225 08:12:45.361321 4978 generic.go:334] "Generic (PLEG): container finished" podID="d75f7033-ac22-4627-a53a-615b153de379" containerID="d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3" exitCode=0 Feb 25 08:12:45 crc kubenswrapper[4978]: I0225 08:12:45.361397 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qlnl" event={"ID":"d75f7033-ac22-4627-a53a-615b153de379","Type":"ContainerDied","Data":"d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3"} Feb 25 08:12:45 crc kubenswrapper[4978]: I0225 08:12:45.361722 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qlnl" event={"ID":"d75f7033-ac22-4627-a53a-615b153de379","Type":"ContainerStarted","Data":"ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0"} Feb 25 08:12:45 crc kubenswrapper[4978]: I0225 08:12:45.364475 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshsf" event={"ID":"3908c4a7-4880-47ee-83fa-0d68a57d589a","Type":"ContainerStarted","Data":"f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11"} Feb 25 08:12:45 crc kubenswrapper[4978]: I0225 08:12:45.396567 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9qlnl" podStartSLOduration=2.877341651 podStartE2EDuration="4.396540371s" podCreationTimestamp="2026-02-25 08:12:41 +0000 UTC" firstStartedPulling="2026-02-25 08:12:43.338025773 +0000 UTC m=+5256.777282272" lastFinishedPulling="2026-02-25 08:12:44.857224533 +0000 UTC m=+5258.296480992" observedRunningTime="2026-02-25 08:12:45.386115419 +0000 UTC m=+5258.825371898" watchObservedRunningTime="2026-02-25 08:12:45.396540371 +0000 UTC m=+5258.835796870" Feb 25 08:12:45 crc kubenswrapper[4978]: I0225 08:12:45.408084 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tshsf" podStartSLOduration=2.9625224 podStartE2EDuration="5.408064858s" podCreationTimestamp="2026-02-25 08:12:40 +0000 UTC" firstStartedPulling="2026-02-25 08:12:42.325761548 +0000 UTC m=+5255.765018007" lastFinishedPulling="2026-02-25 08:12:44.771304006 +0000 UTC m=+5258.210560465" observedRunningTime="2026-02-25 08:12:45.404787957 +0000 UTC m=+5258.844044476" watchObservedRunningTime="2026-02-25 08:12:45.408064858 +0000 UTC m=+5258.847321327" Feb 25 08:12:47 crc kubenswrapper[4978]: I0225 08:12:47.233920 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:47 crc kubenswrapper[4978]: I0225 08:12:47.234459 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:48 crc kubenswrapper[4978]: I0225 08:12:48.310347 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b84cj" podUID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerName="registry-server" probeResult="failure" output=< Feb 25 08:12:48 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 08:12:48 crc kubenswrapper[4978]: > Feb 25 08:12:51 crc kubenswrapper[4978]: I0225 08:12:51.196901 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:51 crc kubenswrapper[4978]: I0225 08:12:51.196952 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:51 crc kubenswrapper[4978]: I0225 08:12:51.255602 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:51 crc kubenswrapper[4978]: I0225 08:12:51.455298 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:52 crc kubenswrapper[4978]: I0225 08:12:52.153973 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:52 crc kubenswrapper[4978]: I0225 08:12:52.154048 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:52 crc kubenswrapper[4978]: I0225 08:12:52.249408 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:52 crc kubenswrapper[4978]: I0225 08:12:52.475119 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:53 crc kubenswrapper[4978]: I0225 08:12:53.233028 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tshsf"] Feb 25 08:12:53 crc kubenswrapper[4978]: I0225 08:12:53.430171 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tshsf" podUID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerName="registry-server" containerID="cri-o://f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11" gracePeriod=2 Feb 25 08:12:53 crc kubenswrapper[4978]: I0225 08:12:53.828407 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9qlnl"] Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.379262 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.396043 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-catalog-content\") pod \"3908c4a7-4880-47ee-83fa-0d68a57d589a\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.396097 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-utilities\") pod \"3908c4a7-4880-47ee-83fa-0d68a57d589a\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.396152 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmgpq\" (UniqueName: \"kubernetes.io/projected/3908c4a7-4880-47ee-83fa-0d68a57d589a-kube-api-access-bmgpq\") pod \"3908c4a7-4880-47ee-83fa-0d68a57d589a\" (UID: \"3908c4a7-4880-47ee-83fa-0d68a57d589a\") " Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.398235 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-utilities" (OuterVolumeSpecName: "utilities") pod "3908c4a7-4880-47ee-83fa-0d68a57d589a" (UID: "3908c4a7-4880-47ee-83fa-0d68a57d589a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.410330 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3908c4a7-4880-47ee-83fa-0d68a57d589a-kube-api-access-bmgpq" (OuterVolumeSpecName: "kube-api-access-bmgpq") pod "3908c4a7-4880-47ee-83fa-0d68a57d589a" (UID: "3908c4a7-4880-47ee-83fa-0d68a57d589a"). InnerVolumeSpecName "kube-api-access-bmgpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.455815 4978 generic.go:334] "Generic (PLEG): container finished" podID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerID="f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11" exitCode=0 Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.457685 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshsf" event={"ID":"3908c4a7-4880-47ee-83fa-0d68a57d589a","Type":"ContainerDied","Data":"f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11"} Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.457817 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshsf" event={"ID":"3908c4a7-4880-47ee-83fa-0d68a57d589a","Type":"ContainerDied","Data":"f91875571e21754528073e73f5411f2f80212b46a81e5c16a563d3e3d223f388"} Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.457842 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tshsf" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.457974 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9qlnl" podUID="d75f7033-ac22-4627-a53a-615b153de379" containerName="registry-server" containerID="cri-o://ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0" gracePeriod=2 Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.457861 4978 scope.go:117] "RemoveContainer" containerID="f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.492580 4978 scope.go:117] "RemoveContainer" containerID="5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.500498 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.500564 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmgpq\" (UniqueName: \"kubernetes.io/projected/3908c4a7-4880-47ee-83fa-0d68a57d589a-kube-api-access-bmgpq\") on node \"crc\" DevicePath \"\"" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.511617 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3908c4a7-4880-47ee-83fa-0d68a57d589a" (UID: "3908c4a7-4880-47ee-83fa-0d68a57d589a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.522790 4978 scope.go:117] "RemoveContainer" containerID="341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.558581 4978 scope.go:117] "RemoveContainer" containerID="f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11" Feb 25 08:12:54 crc kubenswrapper[4978]: E0225 08:12:54.560155 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11\": container with ID starting with f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11 not found: ID does not exist" containerID="f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.560202 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11"} err="failed to get container status \"f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11\": rpc error: code = NotFound desc = could not find container \"f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11\": container with ID starting with f69a2698fe6ff08aa7eb53b3d67a8c880abf3f0bb90b8a609bdccdb9a992fe11 not found: ID does not exist" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.560231 4978 scope.go:117] "RemoveContainer" containerID="5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d" Feb 25 08:12:54 crc kubenswrapper[4978]: E0225 08:12:54.560848 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d\": container with ID starting with 5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d not found: ID does not exist" containerID="5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.560885 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d"} err="failed to get container status \"5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d\": rpc error: code = NotFound desc = could not find container \"5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d\": container with ID starting with 5c80bb9c297ecf8915960a52e61c871840e51ee5a74a96cfde9cf8668cfc4a5d not found: ID does not exist" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.560910 4978 scope.go:117] "RemoveContainer" containerID="341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a" Feb 25 08:12:54 crc kubenswrapper[4978]: E0225 08:12:54.562256 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a\": container with ID starting with 341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a not found: ID does not exist" containerID="341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.562285 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a"} err="failed to get container status \"341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a\": rpc error: code = NotFound desc = could not find container \"341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a\": container with ID starting with 341cc712d038233d225dcb52d7284045e99e6f25092ead78305bd61463fca53a not found: ID does not exist" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.601983 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3908c4a7-4880-47ee-83fa-0d68a57d589a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.792131 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tshsf"] Feb 25 08:12:54 crc kubenswrapper[4978]: I0225 08:12:54.800193 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tshsf"] Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.338426 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3908c4a7-4880-47ee-83fa-0d68a57d589a" path="/var/lib/kubelet/pods/3908c4a7-4880-47ee-83fa-0d68a57d589a/volumes" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.439297 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.471236 4978 generic.go:334] "Generic (PLEG): container finished" podID="d75f7033-ac22-4627-a53a-615b153de379" containerID="ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0" exitCode=0 Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.471271 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qlnl" event={"ID":"d75f7033-ac22-4627-a53a-615b153de379","Type":"ContainerDied","Data":"ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0"} Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.471294 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9qlnl" event={"ID":"d75f7033-ac22-4627-a53a-615b153de379","Type":"ContainerDied","Data":"4fcdc778968bfa3d7aadeeaf94cc9abaaafaeb3700f41e7aae5133591b426179"} Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.471310 4978 scope.go:117] "RemoveContainer" containerID="ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.471422 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9qlnl" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.493116 4978 scope.go:117] "RemoveContainer" containerID="d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.516037 4978 scope.go:117] "RemoveContainer" containerID="6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.517454 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-utilities\") pod \"d75f7033-ac22-4627-a53a-615b153de379\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.517576 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfqn2\" (UniqueName: \"kubernetes.io/projected/d75f7033-ac22-4627-a53a-615b153de379-kube-api-access-cfqn2\") pod \"d75f7033-ac22-4627-a53a-615b153de379\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.517607 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-catalog-content\") pod \"d75f7033-ac22-4627-a53a-615b153de379\" (UID: \"d75f7033-ac22-4627-a53a-615b153de379\") " Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.518798 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-utilities" (OuterVolumeSpecName: "utilities") pod "d75f7033-ac22-4627-a53a-615b153de379" (UID: "d75f7033-ac22-4627-a53a-615b153de379"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.520729 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d75f7033-ac22-4627-a53a-615b153de379-kube-api-access-cfqn2" (OuterVolumeSpecName: "kube-api-access-cfqn2") pod "d75f7033-ac22-4627-a53a-615b153de379" (UID: "d75f7033-ac22-4627-a53a-615b153de379"). InnerVolumeSpecName "kube-api-access-cfqn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.543882 4978 scope.go:117] "RemoveContainer" containerID="ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0" Feb 25 08:12:55 crc kubenswrapper[4978]: E0225 08:12:55.544526 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0\": container with ID starting with ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0 not found: ID does not exist" containerID="ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.544577 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0"} err="failed to get container status \"ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0\": rpc error: code = NotFound desc = could not find container \"ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0\": container with ID starting with ff5a53ae81900180ac7a3787b9d35660644ce5ed58f3eb9c2bba3cd975dc5ce0 not found: ID does not exist" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.544611 4978 scope.go:117] "RemoveContainer" containerID="d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3" Feb 25 08:12:55 crc kubenswrapper[4978]: E0225 08:12:55.545166 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3\": container with ID starting with d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3 not found: ID does not exist" containerID="d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.545204 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3"} err="failed to get container status \"d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3\": rpc error: code = NotFound desc = could not find container \"d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3\": container with ID starting with d63522ed2bd563d071fa16e6dff68dab2cc503fa98b79c15ddd5aa0f4ae2b0e3 not found: ID does not exist" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.545227 4978 scope.go:117] "RemoveContainer" containerID="6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51" Feb 25 08:12:55 crc kubenswrapper[4978]: E0225 08:12:55.545530 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51\": container with ID starting with 6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51 not found: ID does not exist" containerID="6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.545572 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51"} err="failed to get container status \"6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51\": rpc error: code = NotFound desc = could not find container \"6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51\": container with ID starting with 6b3db8a08a41ef440ab7ba7f49fb4f587edfd6e8fc09475af9bcdd4253b7be51 not found: ID does not exist" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.576072 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d75f7033-ac22-4627-a53a-615b153de379" (UID: "d75f7033-ac22-4627-a53a-615b153de379"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.619891 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.619934 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfqn2\" (UniqueName: \"kubernetes.io/projected/d75f7033-ac22-4627-a53a-615b153de379-kube-api-access-cfqn2\") on node \"crc\" DevicePath \"\"" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.619951 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d75f7033-ac22-4627-a53a-615b153de379-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.815444 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9qlnl"] Feb 25 08:12:55 crc kubenswrapper[4978]: I0225 08:12:55.829982 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9qlnl"] Feb 25 08:12:57 crc kubenswrapper[4978]: I0225 08:12:57.302273 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:57 crc kubenswrapper[4978]: I0225 08:12:57.335620 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:12:57 crc kubenswrapper[4978]: E0225 08:12:57.338755 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:12:57 crc kubenswrapper[4978]: I0225 08:12:57.340793 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d75f7033-ac22-4627-a53a-615b153de379" path="/var/lib/kubelet/pods/d75f7033-ac22-4627-a53a-615b153de379/volumes" Feb 25 08:12:57 crc kubenswrapper[4978]: I0225 08:12:57.369916 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:58 crc kubenswrapper[4978]: I0225 08:12:58.834605 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b84cj"] Feb 25 08:12:58 crc kubenswrapper[4978]: I0225 08:12:58.835967 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b84cj" podUID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerName="registry-server" containerID="cri-o://58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775" gracePeriod=2 Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.319188 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.388713 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-catalog-content\") pod \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.388805 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptqdz\" (UniqueName: \"kubernetes.io/projected/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-kube-api-access-ptqdz\") pod \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.388850 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-utilities\") pod \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\" (UID: \"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537\") " Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.390418 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-utilities" (OuterVolumeSpecName: "utilities") pod "c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" (UID: "c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.395136 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-kube-api-access-ptqdz" (OuterVolumeSpecName: "kube-api-access-ptqdz") pod "c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" (UID: "c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537"). InnerVolumeSpecName "kube-api-access-ptqdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.490672 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.490748 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptqdz\" (UniqueName: \"kubernetes.io/projected/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-kube-api-access-ptqdz\") on node \"crc\" DevicePath \"\"" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.510803 4978 generic.go:334] "Generic (PLEG): container finished" podID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerID="58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775" exitCode=0 Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.510856 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b84cj" event={"ID":"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537","Type":"ContainerDied","Data":"58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775"} Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.510886 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b84cj" event={"ID":"c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537","Type":"ContainerDied","Data":"7841f86c7e14a4900d5968acdf4cef8d915765240feae172eedf72f5aba6a985"} Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.510907 4978 scope.go:117] "RemoveContainer" containerID="58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.511040 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b84cj" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.546801 4978 scope.go:117] "RemoveContainer" containerID="6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.550795 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" (UID: "c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.572137 4978 scope.go:117] "RemoveContainer" containerID="5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.592237 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.612659 4978 scope.go:117] "RemoveContainer" containerID="58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775" Feb 25 08:12:59 crc kubenswrapper[4978]: E0225 08:12:59.613052 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775\": container with ID starting with 58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775 not found: ID does not exist" containerID="58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.613093 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775"} err="failed to get container status \"58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775\": rpc error: code = NotFound desc = could not find container \"58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775\": container with ID starting with 58934a2779c3921302ec27dc500f106e3b3bfa99164e992a37bc93c621463775 not found: ID does not exist" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.613121 4978 scope.go:117] "RemoveContainer" containerID="6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf" Feb 25 08:12:59 crc kubenswrapper[4978]: E0225 08:12:59.613724 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf\": container with ID starting with 6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf not found: ID does not exist" containerID="6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.613776 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf"} err="failed to get container status \"6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf\": rpc error: code = NotFound desc = could not find container \"6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf\": container with ID starting with 6694857c3c527476c6e6aa06f523cf15b4b4b79dbbe22287f57fa74538c98bcf not found: ID does not exist" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.613807 4978 scope.go:117] "RemoveContainer" containerID="5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6" Feb 25 08:12:59 crc kubenswrapper[4978]: E0225 08:12:59.614450 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6\": container with ID starting with 5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6 not found: ID does not exist" containerID="5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.614505 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6"} err="failed to get container status \"5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6\": rpc error: code = NotFound desc = could not find container \"5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6\": container with ID starting with 5c2e501fd27f3e392622463a3dbe1fc022c3a2e5954725e6d2a651c5c8bc17d6 not found: ID does not exist" Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.864321 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b84cj"] Feb 25 08:12:59 crc kubenswrapper[4978]: I0225 08:12:59.877225 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b84cj"] Feb 25 08:13:01 crc kubenswrapper[4978]: I0225 08:13:01.344853 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" path="/var/lib/kubelet/pods/c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537/volumes" Feb 25 08:13:09 crc kubenswrapper[4978]: I0225 08:13:09.328445 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:13:09 crc kubenswrapper[4978]: E0225 08:13:09.329455 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:13:22 crc kubenswrapper[4978]: I0225 08:13:22.328995 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:13:22 crc kubenswrapper[4978]: E0225 08:13:22.330060 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:13:37 crc kubenswrapper[4978]: I0225 08:13:37.331709 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:13:37 crc kubenswrapper[4978]: E0225 08:13:37.332665 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:13:48 crc kubenswrapper[4978]: I0225 08:13:48.328722 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:13:48 crc kubenswrapper[4978]: I0225 08:13:48.992031 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"861a0fde7761354b7edadbf02215304f5389b12a48ceeb6beeb07e0a89855ed3"} Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.167733 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533454-h7xtm"] Feb 25 08:14:00 crc kubenswrapper[4978]: E0225 08:14:00.168868 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerName="registry-server" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.168888 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerName="registry-server" Feb 25 08:14:00 crc kubenswrapper[4978]: E0225 08:14:00.168910 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerName="extract-content" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.168919 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerName="extract-content" Feb 25 08:14:00 crc kubenswrapper[4978]: E0225 08:14:00.168951 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerName="extract-utilities" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.168961 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerName="extract-utilities" Feb 25 08:14:00 crc kubenswrapper[4978]: E0225 08:14:00.168982 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerName="extract-content" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.168990 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerName="extract-content" Feb 25 08:14:00 crc kubenswrapper[4978]: E0225 08:14:00.169007 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d75f7033-ac22-4627-a53a-615b153de379" containerName="extract-content" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.169036 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d75f7033-ac22-4627-a53a-615b153de379" containerName="extract-content" Feb 25 08:14:00 crc kubenswrapper[4978]: E0225 08:14:00.169052 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d75f7033-ac22-4627-a53a-615b153de379" containerName="registry-server" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.169060 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d75f7033-ac22-4627-a53a-615b153de379" containerName="registry-server" Feb 25 08:14:00 crc kubenswrapper[4978]: E0225 08:14:00.169075 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerName="registry-server" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.169083 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerName="registry-server" Feb 25 08:14:00 crc kubenswrapper[4978]: E0225 08:14:00.169095 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerName="extract-utilities" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.169126 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerName="extract-utilities" Feb 25 08:14:00 crc kubenswrapper[4978]: E0225 08:14:00.169143 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d75f7033-ac22-4627-a53a-615b153de379" containerName="extract-utilities" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.169151 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d75f7033-ac22-4627-a53a-615b153de379" containerName="extract-utilities" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.169393 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9ac1f1f-f52e-45a8-9a0a-2bfb6f176537" containerName="registry-server" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.169415 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3908c4a7-4880-47ee-83fa-0d68a57d589a" containerName="registry-server" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.169455 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d75f7033-ac22-4627-a53a-615b153de379" containerName="registry-server" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.170167 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533454-h7xtm" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.175435 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.175537 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.175703 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.179955 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533454-h7xtm"] Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.186552 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4c5l\" (UniqueName: \"kubernetes.io/projected/a128a7aa-f607-43bc-a2b0-a676531ffb51-kube-api-access-z4c5l\") pod \"auto-csr-approver-29533454-h7xtm\" (UID: \"a128a7aa-f607-43bc-a2b0-a676531ffb51\") " pod="openshift-infra/auto-csr-approver-29533454-h7xtm" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.287689 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4c5l\" (UniqueName: \"kubernetes.io/projected/a128a7aa-f607-43bc-a2b0-a676531ffb51-kube-api-access-z4c5l\") pod \"auto-csr-approver-29533454-h7xtm\" (UID: \"a128a7aa-f607-43bc-a2b0-a676531ffb51\") " pod="openshift-infra/auto-csr-approver-29533454-h7xtm" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.310433 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4c5l\" (UniqueName: \"kubernetes.io/projected/a128a7aa-f607-43bc-a2b0-a676531ffb51-kube-api-access-z4c5l\") pod \"auto-csr-approver-29533454-h7xtm\" (UID: \"a128a7aa-f607-43bc-a2b0-a676531ffb51\") " pod="openshift-infra/auto-csr-approver-29533454-h7xtm" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.488449 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533454-h7xtm" Feb 25 08:14:00 crc kubenswrapper[4978]: I0225 08:14:00.806031 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533454-h7xtm"] Feb 25 08:14:00 crc kubenswrapper[4978]: W0225 08:14:00.817875 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda128a7aa_f607_43bc_a2b0_a676531ffb51.slice/crio-6cf4763d3d9dfd6cc9611fe3d06c7178f8cc81d8a9f940a6db914f4e65173f0d WatchSource:0}: Error finding container 6cf4763d3d9dfd6cc9611fe3d06c7178f8cc81d8a9f940a6db914f4e65173f0d: Status 404 returned error can't find the container with id 6cf4763d3d9dfd6cc9611fe3d06c7178f8cc81d8a9f940a6db914f4e65173f0d Feb 25 08:14:01 crc kubenswrapper[4978]: I0225 08:14:01.151301 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533454-h7xtm" event={"ID":"a128a7aa-f607-43bc-a2b0-a676531ffb51","Type":"ContainerStarted","Data":"6cf4763d3d9dfd6cc9611fe3d06c7178f8cc81d8a9f940a6db914f4e65173f0d"} Feb 25 08:14:02 crc kubenswrapper[4978]: I0225 08:14:02.159343 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533454-h7xtm" event={"ID":"a128a7aa-f607-43bc-a2b0-a676531ffb51","Type":"ContainerStarted","Data":"112e6e23840eb8f02a4f31874a2d519d065aa086a035368f7e870e96fc92feaa"} Feb 25 08:14:02 crc kubenswrapper[4978]: I0225 08:14:02.184486 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533454-h7xtm" podStartSLOduration=1.258424264 podStartE2EDuration="2.184457802s" podCreationTimestamp="2026-02-25 08:14:00 +0000 UTC" firstStartedPulling="2026-02-25 08:14:00.820021957 +0000 UTC m=+5334.259278466" lastFinishedPulling="2026-02-25 08:14:01.746055475 +0000 UTC m=+5335.185312004" observedRunningTime="2026-02-25 08:14:02.177552218 +0000 UTC m=+5335.616808687" watchObservedRunningTime="2026-02-25 08:14:02.184457802 +0000 UTC m=+5335.623714291" Feb 25 08:14:03 crc kubenswrapper[4978]: I0225 08:14:03.171666 4978 generic.go:334] "Generic (PLEG): container finished" podID="a128a7aa-f607-43bc-a2b0-a676531ffb51" containerID="112e6e23840eb8f02a4f31874a2d519d065aa086a035368f7e870e96fc92feaa" exitCode=0 Feb 25 08:14:03 crc kubenswrapper[4978]: I0225 08:14:03.171784 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533454-h7xtm" event={"ID":"a128a7aa-f607-43bc-a2b0-a676531ffb51","Type":"ContainerDied","Data":"112e6e23840eb8f02a4f31874a2d519d065aa086a035368f7e870e96fc92feaa"} Feb 25 08:14:04 crc kubenswrapper[4978]: I0225 08:14:04.538080 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533454-h7xtm" Feb 25 08:14:04 crc kubenswrapper[4978]: I0225 08:14:04.569094 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4c5l\" (UniqueName: \"kubernetes.io/projected/a128a7aa-f607-43bc-a2b0-a676531ffb51-kube-api-access-z4c5l\") pod \"a128a7aa-f607-43bc-a2b0-a676531ffb51\" (UID: \"a128a7aa-f607-43bc-a2b0-a676531ffb51\") " Feb 25 08:14:04 crc kubenswrapper[4978]: I0225 08:14:04.575948 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a128a7aa-f607-43bc-a2b0-a676531ffb51-kube-api-access-z4c5l" (OuterVolumeSpecName: "kube-api-access-z4c5l") pod "a128a7aa-f607-43bc-a2b0-a676531ffb51" (UID: "a128a7aa-f607-43bc-a2b0-a676531ffb51"). InnerVolumeSpecName "kube-api-access-z4c5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:14:04 crc kubenswrapper[4978]: I0225 08:14:04.670796 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4c5l\" (UniqueName: \"kubernetes.io/projected/a128a7aa-f607-43bc-a2b0-a676531ffb51-kube-api-access-z4c5l\") on node \"crc\" DevicePath \"\"" Feb 25 08:14:05 crc kubenswrapper[4978]: I0225 08:14:05.193362 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533454-h7xtm" event={"ID":"a128a7aa-f607-43bc-a2b0-a676531ffb51","Type":"ContainerDied","Data":"6cf4763d3d9dfd6cc9611fe3d06c7178f8cc81d8a9f940a6db914f4e65173f0d"} Feb 25 08:14:05 crc kubenswrapper[4978]: I0225 08:14:05.193445 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cf4763d3d9dfd6cc9611fe3d06c7178f8cc81d8a9f940a6db914f4e65173f0d" Feb 25 08:14:05 crc kubenswrapper[4978]: I0225 08:14:05.193462 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533454-h7xtm" Feb 25 08:14:05 crc kubenswrapper[4978]: I0225 08:14:05.299776 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533448-f58x8"] Feb 25 08:14:05 crc kubenswrapper[4978]: I0225 08:14:05.299850 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533448-f58x8"] Feb 25 08:14:05 crc kubenswrapper[4978]: I0225 08:14:05.354049 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f47615c-d56d-45a6-a490-6cac1a4d09cb" path="/var/lib/kubelet/pods/5f47615c-d56d-45a6-a490-6cac1a4d09cb/volumes" Feb 25 08:14:25 crc kubenswrapper[4978]: I0225 08:14:25.354846 4978 scope.go:117] "RemoveContainer" containerID="047a36f280dc81e299179e96ee49dbd63b1b67cd9a65aef29288ffd5074c87b0" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.166944 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7"] Feb 25 08:15:00 crc kubenswrapper[4978]: E0225 08:15:00.167858 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a128a7aa-f607-43bc-a2b0-a676531ffb51" containerName="oc" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.167876 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a128a7aa-f607-43bc-a2b0-a676531ffb51" containerName="oc" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.168035 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a128a7aa-f607-43bc-a2b0-a676531ffb51" containerName="oc" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.168684 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.172993 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.173197 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.188680 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7"] Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.197775 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/302baf2f-e0af-4f5e-88dd-36682e379d58-config-volume\") pod \"collect-profiles-29533455-v2rl7\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.197955 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b85j\" (UniqueName: \"kubernetes.io/projected/302baf2f-e0af-4f5e-88dd-36682e379d58-kube-api-access-4b85j\") pod \"collect-profiles-29533455-v2rl7\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.198026 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/302baf2f-e0af-4f5e-88dd-36682e379d58-secret-volume\") pod \"collect-profiles-29533455-v2rl7\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.298663 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b85j\" (UniqueName: \"kubernetes.io/projected/302baf2f-e0af-4f5e-88dd-36682e379d58-kube-api-access-4b85j\") pod \"collect-profiles-29533455-v2rl7\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.298719 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/302baf2f-e0af-4f5e-88dd-36682e379d58-secret-volume\") pod \"collect-profiles-29533455-v2rl7\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.298792 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/302baf2f-e0af-4f5e-88dd-36682e379d58-config-volume\") pod \"collect-profiles-29533455-v2rl7\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.299742 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/302baf2f-e0af-4f5e-88dd-36682e379d58-config-volume\") pod \"collect-profiles-29533455-v2rl7\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.314622 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/302baf2f-e0af-4f5e-88dd-36682e379d58-secret-volume\") pod \"collect-profiles-29533455-v2rl7\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.320166 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b85j\" (UniqueName: \"kubernetes.io/projected/302baf2f-e0af-4f5e-88dd-36682e379d58-kube-api-access-4b85j\") pod \"collect-profiles-29533455-v2rl7\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.495788 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:00 crc kubenswrapper[4978]: I0225 08:15:00.772502 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7"] Feb 25 08:15:01 crc kubenswrapper[4978]: I0225 08:15:01.708762 4978 generic.go:334] "Generic (PLEG): container finished" podID="302baf2f-e0af-4f5e-88dd-36682e379d58" containerID="eb3b71bb78df7a040c81acbd60e451a59b5384296b041111b8b5140598dfe1a6" exitCode=0 Feb 25 08:15:01 crc kubenswrapper[4978]: I0225 08:15:01.708832 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" event={"ID":"302baf2f-e0af-4f5e-88dd-36682e379d58","Type":"ContainerDied","Data":"eb3b71bb78df7a040c81acbd60e451a59b5384296b041111b8b5140598dfe1a6"} Feb 25 08:15:01 crc kubenswrapper[4978]: I0225 08:15:01.709695 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" event={"ID":"302baf2f-e0af-4f5e-88dd-36682e379d58","Type":"ContainerStarted","Data":"e3e5d8c62bc9c9a76d4b509fc20537bd617c5ec6161486a4212bc74121ad14bd"} Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.023411 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.146706 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b85j\" (UniqueName: \"kubernetes.io/projected/302baf2f-e0af-4f5e-88dd-36682e379d58-kube-api-access-4b85j\") pod \"302baf2f-e0af-4f5e-88dd-36682e379d58\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.147517 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/302baf2f-e0af-4f5e-88dd-36682e379d58-secret-volume\") pod \"302baf2f-e0af-4f5e-88dd-36682e379d58\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.147584 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/302baf2f-e0af-4f5e-88dd-36682e379d58-config-volume\") pod \"302baf2f-e0af-4f5e-88dd-36682e379d58\" (UID: \"302baf2f-e0af-4f5e-88dd-36682e379d58\") " Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.148849 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302baf2f-e0af-4f5e-88dd-36682e379d58-config-volume" (OuterVolumeSpecName: "config-volume") pod "302baf2f-e0af-4f5e-88dd-36682e379d58" (UID: "302baf2f-e0af-4f5e-88dd-36682e379d58"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.153647 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/302baf2f-e0af-4f5e-88dd-36682e379d58-kube-api-access-4b85j" (OuterVolumeSpecName: "kube-api-access-4b85j") pod "302baf2f-e0af-4f5e-88dd-36682e379d58" (UID: "302baf2f-e0af-4f5e-88dd-36682e379d58"). InnerVolumeSpecName "kube-api-access-4b85j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.155596 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/302baf2f-e0af-4f5e-88dd-36682e379d58-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "302baf2f-e0af-4f5e-88dd-36682e379d58" (UID: "302baf2f-e0af-4f5e-88dd-36682e379d58"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.249189 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/302baf2f-e0af-4f5e-88dd-36682e379d58-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.249236 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/302baf2f-e0af-4f5e-88dd-36682e379d58-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.249248 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b85j\" (UniqueName: \"kubernetes.io/projected/302baf2f-e0af-4f5e-88dd-36682e379d58-kube-api-access-4b85j\") on node \"crc\" DevicePath \"\"" Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.730856 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" event={"ID":"302baf2f-e0af-4f5e-88dd-36682e379d58","Type":"ContainerDied","Data":"e3e5d8c62bc9c9a76d4b509fc20537bd617c5ec6161486a4212bc74121ad14bd"} Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.730914 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3e5d8c62bc9c9a76d4b509fc20537bd617c5ec6161486a4212bc74121ad14bd" Feb 25 08:15:03 crc kubenswrapper[4978]: I0225 08:15:03.731339 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7" Feb 25 08:15:04 crc kubenswrapper[4978]: I0225 08:15:04.122341 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx"] Feb 25 08:15:04 crc kubenswrapper[4978]: I0225 08:15:04.133211 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533410-rzrpx"] Feb 25 08:15:05 crc kubenswrapper[4978]: I0225 08:15:05.345753 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3fbc3ff-fd50-4030-be63-750709b2f3fb" path="/var/lib/kubelet/pods/c3fbc3ff-fd50-4030-be63-750709b2f3fb/volumes" Feb 25 08:15:25 crc kubenswrapper[4978]: I0225 08:15:25.451418 4978 scope.go:117] "RemoveContainer" containerID="2c0820c4fa4f3842a8ee5ce9b50c8d18b1c11f736387c8957a132224ad5854c1" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.164365 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533456-24g78"] Feb 25 08:16:00 crc kubenswrapper[4978]: E0225 08:16:00.165822 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302baf2f-e0af-4f5e-88dd-36682e379d58" containerName="collect-profiles" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.165853 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="302baf2f-e0af-4f5e-88dd-36682e379d58" containerName="collect-profiles" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.166346 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="302baf2f-e0af-4f5e-88dd-36682e379d58" containerName="collect-profiles" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.167436 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533456-24g78" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.175114 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.175149 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.175146 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.185914 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533456-24g78"] Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.205872 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd7mp\" (UniqueName: \"kubernetes.io/projected/93f60f49-8d56-4d2d-b73f-40e5a5aa6d03-kube-api-access-jd7mp\") pod \"auto-csr-approver-29533456-24g78\" (UID: \"93f60f49-8d56-4d2d-b73f-40e5a5aa6d03\") " pod="openshift-infra/auto-csr-approver-29533456-24g78" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.310115 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd7mp\" (UniqueName: \"kubernetes.io/projected/93f60f49-8d56-4d2d-b73f-40e5a5aa6d03-kube-api-access-jd7mp\") pod \"auto-csr-approver-29533456-24g78\" (UID: \"93f60f49-8d56-4d2d-b73f-40e5a5aa6d03\") " pod="openshift-infra/auto-csr-approver-29533456-24g78" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.336421 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd7mp\" (UniqueName: \"kubernetes.io/projected/93f60f49-8d56-4d2d-b73f-40e5a5aa6d03-kube-api-access-jd7mp\") pod \"auto-csr-approver-29533456-24g78\" (UID: \"93f60f49-8d56-4d2d-b73f-40e5a5aa6d03\") " pod="openshift-infra/auto-csr-approver-29533456-24g78" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.506491 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533456-24g78" Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.962746 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533456-24g78"] Feb 25 08:16:00 crc kubenswrapper[4978]: I0225 08:16:00.971064 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 08:16:01 crc kubenswrapper[4978]: I0225 08:16:01.276337 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533456-24g78" event={"ID":"93f60f49-8d56-4d2d-b73f-40e5a5aa6d03","Type":"ContainerStarted","Data":"cd8a3defa77fd5a19b01b1ac9226e9ab29746104d7b21c9cb1efc7b763e6548b"} Feb 25 08:16:02 crc kubenswrapper[4978]: I0225 08:16:02.287916 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533456-24g78" event={"ID":"93f60f49-8d56-4d2d-b73f-40e5a5aa6d03","Type":"ContainerStarted","Data":"8f4b03a6bf001d971345160007c85b199ea7395f34df46f7c9fcb54e4b40774d"} Feb 25 08:16:02 crc kubenswrapper[4978]: I0225 08:16:02.312999 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533456-24g78" podStartSLOduration=1.486188374 podStartE2EDuration="2.312978111s" podCreationTimestamp="2026-02-25 08:16:00 +0000 UTC" firstStartedPulling="2026-02-25 08:16:00.970743243 +0000 UTC m=+5454.409999712" lastFinishedPulling="2026-02-25 08:16:01.79753298 +0000 UTC m=+5455.236789449" observedRunningTime="2026-02-25 08:16:02.305412697 +0000 UTC m=+5455.744669176" watchObservedRunningTime="2026-02-25 08:16:02.312978111 +0000 UTC m=+5455.752234590" Feb 25 08:16:03 crc kubenswrapper[4978]: I0225 08:16:03.298827 4978 generic.go:334] "Generic (PLEG): container finished" podID="93f60f49-8d56-4d2d-b73f-40e5a5aa6d03" containerID="8f4b03a6bf001d971345160007c85b199ea7395f34df46f7c9fcb54e4b40774d" exitCode=0 Feb 25 08:16:03 crc kubenswrapper[4978]: I0225 08:16:03.298891 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533456-24g78" event={"ID":"93f60f49-8d56-4d2d-b73f-40e5a5aa6d03","Type":"ContainerDied","Data":"8f4b03a6bf001d971345160007c85b199ea7395f34df46f7c9fcb54e4b40774d"} Feb 25 08:16:04 crc kubenswrapper[4978]: I0225 08:16:04.661126 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533456-24g78" Feb 25 08:16:04 crc kubenswrapper[4978]: I0225 08:16:04.687207 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd7mp\" (UniqueName: \"kubernetes.io/projected/93f60f49-8d56-4d2d-b73f-40e5a5aa6d03-kube-api-access-jd7mp\") pod \"93f60f49-8d56-4d2d-b73f-40e5a5aa6d03\" (UID: \"93f60f49-8d56-4d2d-b73f-40e5a5aa6d03\") " Feb 25 08:16:04 crc kubenswrapper[4978]: I0225 08:16:04.696738 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93f60f49-8d56-4d2d-b73f-40e5a5aa6d03-kube-api-access-jd7mp" (OuterVolumeSpecName: "kube-api-access-jd7mp") pod "93f60f49-8d56-4d2d-b73f-40e5a5aa6d03" (UID: "93f60f49-8d56-4d2d-b73f-40e5a5aa6d03"). InnerVolumeSpecName "kube-api-access-jd7mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:16:04 crc kubenswrapper[4978]: I0225 08:16:04.793791 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd7mp\" (UniqueName: \"kubernetes.io/projected/93f60f49-8d56-4d2d-b73f-40e5a5aa6d03-kube-api-access-jd7mp\") on node \"crc\" DevicePath \"\"" Feb 25 08:16:05 crc kubenswrapper[4978]: I0225 08:16:05.321426 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533456-24g78" event={"ID":"93f60f49-8d56-4d2d-b73f-40e5a5aa6d03","Type":"ContainerDied","Data":"cd8a3defa77fd5a19b01b1ac9226e9ab29746104d7b21c9cb1efc7b763e6548b"} Feb 25 08:16:05 crc kubenswrapper[4978]: I0225 08:16:05.321493 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd8a3defa77fd5a19b01b1ac9226e9ab29746104d7b21c9cb1efc7b763e6548b" Feb 25 08:16:05 crc kubenswrapper[4978]: I0225 08:16:05.321586 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533456-24g78" Feb 25 08:16:05 crc kubenswrapper[4978]: I0225 08:16:05.386030 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533450-2zpsl"] Feb 25 08:16:05 crc kubenswrapper[4978]: I0225 08:16:05.392403 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533450-2zpsl"] Feb 25 08:16:07 crc kubenswrapper[4978]: I0225 08:16:07.345591 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e6928e0-72e4-4fea-8d60-8a5cd3514b7c" path="/var/lib/kubelet/pods/3e6928e0-72e4-4fea-8d60-8a5cd3514b7c/volumes" Feb 25 08:16:16 crc kubenswrapper[4978]: I0225 08:16:16.540148 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:16:16 crc kubenswrapper[4978]: I0225 08:16:16.540948 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.009506 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wk7r5"] Feb 25 08:16:19 crc kubenswrapper[4978]: E0225 08:16:19.010085 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93f60f49-8d56-4d2d-b73f-40e5a5aa6d03" containerName="oc" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.010103 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="93f60f49-8d56-4d2d-b73f-40e5a5aa6d03" containerName="oc" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.010260 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="93f60f49-8d56-4d2d-b73f-40e5a5aa6d03" containerName="oc" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.011450 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.017585 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-utilities\") pod \"redhat-marketplace-wk7r5\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.017826 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-catalog-content\") pod \"redhat-marketplace-wk7r5\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.027710 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wk7r5"] Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.119460 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-utilities\") pod \"redhat-marketplace-wk7r5\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.119651 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfr4h\" (UniqueName: \"kubernetes.io/projected/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-kube-api-access-zfr4h\") pod \"redhat-marketplace-wk7r5\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.119910 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-catalog-content\") pod \"redhat-marketplace-wk7r5\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.120014 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-utilities\") pod \"redhat-marketplace-wk7r5\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.120585 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-catalog-content\") pod \"redhat-marketplace-wk7r5\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.221070 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfr4h\" (UniqueName: \"kubernetes.io/projected/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-kube-api-access-zfr4h\") pod \"redhat-marketplace-wk7r5\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.251162 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfr4h\" (UniqueName: \"kubernetes.io/projected/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-kube-api-access-zfr4h\") pod \"redhat-marketplace-wk7r5\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.335316 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:19 crc kubenswrapper[4978]: I0225 08:16:19.591541 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wk7r5"] Feb 25 08:16:19 crc kubenswrapper[4978]: W0225 08:16:19.600114 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod799eb6b5_1baf_47ae_8dc5_041b2ba1db4b.slice/crio-6b2944674cf1f69a49dd34a8026d2083ea87db0588b1b424b296d89ba0655b7c WatchSource:0}: Error finding container 6b2944674cf1f69a49dd34a8026d2083ea87db0588b1b424b296d89ba0655b7c: Status 404 returned error can't find the container with id 6b2944674cf1f69a49dd34a8026d2083ea87db0588b1b424b296d89ba0655b7c Feb 25 08:16:20 crc kubenswrapper[4978]: I0225 08:16:20.469022 4978 generic.go:334] "Generic (PLEG): container finished" podID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerID="5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e" exitCode=0 Feb 25 08:16:20 crc kubenswrapper[4978]: I0225 08:16:20.469166 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk7r5" event={"ID":"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b","Type":"ContainerDied","Data":"5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e"} Feb 25 08:16:20 crc kubenswrapper[4978]: I0225 08:16:20.469502 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk7r5" event={"ID":"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b","Type":"ContainerStarted","Data":"6b2944674cf1f69a49dd34a8026d2083ea87db0588b1b424b296d89ba0655b7c"} Feb 25 08:16:21 crc kubenswrapper[4978]: I0225 08:16:21.482167 4978 generic.go:334] "Generic (PLEG): container finished" podID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerID="e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165" exitCode=0 Feb 25 08:16:21 crc kubenswrapper[4978]: I0225 08:16:21.482395 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk7r5" event={"ID":"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b","Type":"ContainerDied","Data":"e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165"} Feb 25 08:16:22 crc kubenswrapper[4978]: I0225 08:16:22.498468 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk7r5" event={"ID":"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b","Type":"ContainerStarted","Data":"1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83"} Feb 25 08:16:22 crc kubenswrapper[4978]: I0225 08:16:22.539819 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wk7r5" podStartSLOduration=3.127087467 podStartE2EDuration="4.539793314s" podCreationTimestamp="2026-02-25 08:16:18 +0000 UTC" firstStartedPulling="2026-02-25 08:16:20.472114352 +0000 UTC m=+5473.911370811" lastFinishedPulling="2026-02-25 08:16:21.884820159 +0000 UTC m=+5475.324076658" observedRunningTime="2026-02-25 08:16:22.530868869 +0000 UTC m=+5475.970125418" watchObservedRunningTime="2026-02-25 08:16:22.539793314 +0000 UTC m=+5475.979049833" Feb 25 08:16:25 crc kubenswrapper[4978]: I0225 08:16:25.521991 4978 scope.go:117] "RemoveContainer" containerID="967881ccc94747d8934b48502ab9aafe74b12c2396af224232eb915899e98c22" Feb 25 08:16:29 crc kubenswrapper[4978]: I0225 08:16:29.344797 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:29 crc kubenswrapper[4978]: I0225 08:16:29.345125 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:29 crc kubenswrapper[4978]: I0225 08:16:29.410573 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:29 crc kubenswrapper[4978]: I0225 08:16:29.634548 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:29 crc kubenswrapper[4978]: I0225 08:16:29.698146 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wk7r5"] Feb 25 08:16:31 crc kubenswrapper[4978]: I0225 08:16:31.581301 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wk7r5" podUID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerName="registry-server" containerID="cri-o://1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83" gracePeriod=2 Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.114739 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.228893 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-utilities\") pod \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.229026 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfr4h\" (UniqueName: \"kubernetes.io/projected/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-kube-api-access-zfr4h\") pod \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.229324 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-catalog-content\") pod \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\" (UID: \"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b\") " Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.231336 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-utilities" (OuterVolumeSpecName: "utilities") pod "799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" (UID: "799eb6b5-1baf-47ae-8dc5-041b2ba1db4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.236475 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.239042 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-kube-api-access-zfr4h" (OuterVolumeSpecName: "kube-api-access-zfr4h") pod "799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" (UID: "799eb6b5-1baf-47ae-8dc5-041b2ba1db4b"). InnerVolumeSpecName "kube-api-access-zfr4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.276852 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" (UID: "799eb6b5-1baf-47ae-8dc5-041b2ba1db4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.337871 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.337920 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfr4h\" (UniqueName: \"kubernetes.io/projected/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b-kube-api-access-zfr4h\") on node \"crc\" DevicePath \"\"" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.595612 4978 generic.go:334] "Generic (PLEG): container finished" podID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerID="1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83" exitCode=0 Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.595670 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk7r5" event={"ID":"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b","Type":"ContainerDied","Data":"1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83"} Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.595699 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wk7r5" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.595721 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk7r5" event={"ID":"799eb6b5-1baf-47ae-8dc5-041b2ba1db4b","Type":"ContainerDied","Data":"6b2944674cf1f69a49dd34a8026d2083ea87db0588b1b424b296d89ba0655b7c"} Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.595759 4978 scope.go:117] "RemoveContainer" containerID="1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.625818 4978 scope.go:117] "RemoveContainer" containerID="e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.655308 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wk7r5"] Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.673187 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wk7r5"] Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.674936 4978 scope.go:117] "RemoveContainer" containerID="5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.702588 4978 scope.go:117] "RemoveContainer" containerID="1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83" Feb 25 08:16:32 crc kubenswrapper[4978]: E0225 08:16:32.703162 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83\": container with ID starting with 1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83 not found: ID does not exist" containerID="1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.703233 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83"} err="failed to get container status \"1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83\": rpc error: code = NotFound desc = could not find container \"1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83\": container with ID starting with 1a95c31b5a1736e51bd4b93bdb2e9f3162401c84323cbf8ea6b110d35d2ffb83 not found: ID does not exist" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.703279 4978 scope.go:117] "RemoveContainer" containerID="e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165" Feb 25 08:16:32 crc kubenswrapper[4978]: E0225 08:16:32.704084 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165\": container with ID starting with e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165 not found: ID does not exist" containerID="e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.704149 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165"} err="failed to get container status \"e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165\": rpc error: code = NotFound desc = could not find container \"e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165\": container with ID starting with e8e6d2e71fcba4e1a049e83c11e9a04c41506572498e9aac62075184d2baf165 not found: ID does not exist" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.704193 4978 scope.go:117] "RemoveContainer" containerID="5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e" Feb 25 08:16:32 crc kubenswrapper[4978]: E0225 08:16:32.704690 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e\": container with ID starting with 5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e not found: ID does not exist" containerID="5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e" Feb 25 08:16:32 crc kubenswrapper[4978]: I0225 08:16:32.704746 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e"} err="failed to get container status \"5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e\": rpc error: code = NotFound desc = could not find container \"5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e\": container with ID starting with 5fbef98084a14946ebf89025a0e11056d4107396b55879ab933eb03ca2a3154e not found: ID does not exist" Feb 25 08:16:33 crc kubenswrapper[4978]: I0225 08:16:33.342608 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" path="/var/lib/kubelet/pods/799eb6b5-1baf-47ae-8dc5-041b2ba1db4b/volumes" Feb 25 08:16:46 crc kubenswrapper[4978]: I0225 08:16:46.540479 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:16:46 crc kubenswrapper[4978]: I0225 08:16:46.540927 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:17:16 crc kubenswrapper[4978]: I0225 08:17:16.540917 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:17:16 crc kubenswrapper[4978]: I0225 08:17:16.541719 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:17:16 crc kubenswrapper[4978]: I0225 08:17:16.541790 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 08:17:16 crc kubenswrapper[4978]: I0225 08:17:16.542734 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"861a0fde7761354b7edadbf02215304f5389b12a48ceeb6beeb07e0a89855ed3"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 08:17:16 crc kubenswrapper[4978]: I0225 08:17:16.542834 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://861a0fde7761354b7edadbf02215304f5389b12a48ceeb6beeb07e0a89855ed3" gracePeriod=600 Feb 25 08:17:17 crc kubenswrapper[4978]: I0225 08:17:17.033441 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="861a0fde7761354b7edadbf02215304f5389b12a48ceeb6beeb07e0a89855ed3" exitCode=0 Feb 25 08:17:17 crc kubenswrapper[4978]: I0225 08:17:17.033521 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"861a0fde7761354b7edadbf02215304f5389b12a48ceeb6beeb07e0a89855ed3"} Feb 25 08:17:17 crc kubenswrapper[4978]: I0225 08:17:17.033836 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6"} Feb 25 08:17:17 crc kubenswrapper[4978]: I0225 08:17:17.033875 4978 scope.go:117] "RemoveContainer" containerID="dcdf3f14625e80b7836bbcbf3946bab7e5c36314f53068e44f54ab94dafe1a34" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.164728 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533458-92945"] Feb 25 08:18:00 crc kubenswrapper[4978]: E0225 08:18:00.165666 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerName="registry-server" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.165683 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerName="registry-server" Feb 25 08:18:00 crc kubenswrapper[4978]: E0225 08:18:00.165712 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerName="extract-content" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.165719 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerName="extract-content" Feb 25 08:18:00 crc kubenswrapper[4978]: E0225 08:18:00.165730 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerName="extract-utilities" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.165737 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerName="extract-utilities" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.165898 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="799eb6b5-1baf-47ae-8dc5-041b2ba1db4b" containerName="registry-server" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.166538 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533458-92945" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.169133 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.169339 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.169406 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.173956 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533458-92945"] Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.226043 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4w2z\" (UniqueName: \"kubernetes.io/projected/6cac4081-5b91-4683-890f-81b8901bd5db-kube-api-access-k4w2z\") pod \"auto-csr-approver-29533458-92945\" (UID: \"6cac4081-5b91-4683-890f-81b8901bd5db\") " pod="openshift-infra/auto-csr-approver-29533458-92945" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.327408 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4w2z\" (UniqueName: \"kubernetes.io/projected/6cac4081-5b91-4683-890f-81b8901bd5db-kube-api-access-k4w2z\") pod \"auto-csr-approver-29533458-92945\" (UID: \"6cac4081-5b91-4683-890f-81b8901bd5db\") " pod="openshift-infra/auto-csr-approver-29533458-92945" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.357050 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4w2z\" (UniqueName: \"kubernetes.io/projected/6cac4081-5b91-4683-890f-81b8901bd5db-kube-api-access-k4w2z\") pod \"auto-csr-approver-29533458-92945\" (UID: \"6cac4081-5b91-4683-890f-81b8901bd5db\") " pod="openshift-infra/auto-csr-approver-29533458-92945" Feb 25 08:18:00 crc kubenswrapper[4978]: I0225 08:18:00.488045 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533458-92945" Feb 25 08:18:01 crc kubenswrapper[4978]: I0225 08:18:01.007473 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533458-92945"] Feb 25 08:18:01 crc kubenswrapper[4978]: I0225 08:18:01.455026 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533458-92945" event={"ID":"6cac4081-5b91-4683-890f-81b8901bd5db","Type":"ContainerStarted","Data":"6b87884f1a417fc6d1202f85475e558889bccceb0bf759eb325255eed6fa3eca"} Feb 25 08:18:02 crc kubenswrapper[4978]: I0225 08:18:02.461618 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533458-92945" event={"ID":"6cac4081-5b91-4683-890f-81b8901bd5db","Type":"ContainerStarted","Data":"9fa9965d85d7a05a3b923998112a98762766c61e7431e2a5765ef9bfd9aa8f79"} Feb 25 08:18:02 crc kubenswrapper[4978]: I0225 08:18:02.480357 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533458-92945" podStartSLOduration=1.467500612 podStartE2EDuration="2.480336454s" podCreationTimestamp="2026-02-25 08:18:00 +0000 UTC" firstStartedPulling="2026-02-25 08:18:01.011920653 +0000 UTC m=+5574.451177142" lastFinishedPulling="2026-02-25 08:18:02.024756525 +0000 UTC m=+5575.464012984" observedRunningTime="2026-02-25 08:18:02.475806403 +0000 UTC m=+5575.915062892" watchObservedRunningTime="2026-02-25 08:18:02.480336454 +0000 UTC m=+5575.919592933" Feb 25 08:18:03 crc kubenswrapper[4978]: I0225 08:18:03.472474 4978 generic.go:334] "Generic (PLEG): container finished" podID="6cac4081-5b91-4683-890f-81b8901bd5db" containerID="9fa9965d85d7a05a3b923998112a98762766c61e7431e2a5765ef9bfd9aa8f79" exitCode=0 Feb 25 08:18:03 crc kubenswrapper[4978]: I0225 08:18:03.472533 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533458-92945" event={"ID":"6cac4081-5b91-4683-890f-81b8901bd5db","Type":"ContainerDied","Data":"9fa9965d85d7a05a3b923998112a98762766c61e7431e2a5765ef9bfd9aa8f79"} Feb 25 08:18:04 crc kubenswrapper[4978]: I0225 08:18:04.863877 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533458-92945" Feb 25 08:18:04 crc kubenswrapper[4978]: I0225 08:18:04.996509 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4w2z\" (UniqueName: \"kubernetes.io/projected/6cac4081-5b91-4683-890f-81b8901bd5db-kube-api-access-k4w2z\") pod \"6cac4081-5b91-4683-890f-81b8901bd5db\" (UID: \"6cac4081-5b91-4683-890f-81b8901bd5db\") " Feb 25 08:18:05 crc kubenswrapper[4978]: I0225 08:18:05.008680 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cac4081-5b91-4683-890f-81b8901bd5db-kube-api-access-k4w2z" (OuterVolumeSpecName: "kube-api-access-k4w2z") pod "6cac4081-5b91-4683-890f-81b8901bd5db" (UID: "6cac4081-5b91-4683-890f-81b8901bd5db"). InnerVolumeSpecName "kube-api-access-k4w2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:18:05 crc kubenswrapper[4978]: I0225 08:18:05.098526 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4w2z\" (UniqueName: \"kubernetes.io/projected/6cac4081-5b91-4683-890f-81b8901bd5db-kube-api-access-k4w2z\") on node \"crc\" DevicePath \"\"" Feb 25 08:18:05 crc kubenswrapper[4978]: I0225 08:18:05.493478 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533458-92945" event={"ID":"6cac4081-5b91-4683-890f-81b8901bd5db","Type":"ContainerDied","Data":"6b87884f1a417fc6d1202f85475e558889bccceb0bf759eb325255eed6fa3eca"} Feb 25 08:18:05 crc kubenswrapper[4978]: I0225 08:18:05.493966 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b87884f1a417fc6d1202f85475e558889bccceb0bf759eb325255eed6fa3eca" Feb 25 08:18:05 crc kubenswrapper[4978]: I0225 08:18:05.493588 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533458-92945" Feb 25 08:18:05 crc kubenswrapper[4978]: I0225 08:18:05.563422 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533452-rrccs"] Feb 25 08:18:05 crc kubenswrapper[4978]: I0225 08:18:05.571081 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533452-rrccs"] Feb 25 08:18:07 crc kubenswrapper[4978]: I0225 08:18:07.334835 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f483cd75-845c-40cf-9dde-8ef1c61ee613" path="/var/lib/kubelet/pods/f483cd75-845c-40cf-9dde-8ef1c61ee613/volumes" Feb 25 08:18:25 crc kubenswrapper[4978]: I0225 08:18:25.676834 4978 scope.go:117] "RemoveContainer" containerID="649aa004fc1a04e1abbd9c952d73f4cceb71fddbf6838243206fb376651d6e8f" Feb 25 08:19:16 crc kubenswrapper[4978]: I0225 08:19:16.540464 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:19:16 crc kubenswrapper[4978]: I0225 08:19:16.541057 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:19:46 crc kubenswrapper[4978]: I0225 08:19:46.541090 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:19:46 crc kubenswrapper[4978]: I0225 08:19:46.541894 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.144759 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533460-hdnlv"] Feb 25 08:20:00 crc kubenswrapper[4978]: E0225 08:20:00.145534 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cac4081-5b91-4683-890f-81b8901bd5db" containerName="oc" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.145549 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cac4081-5b91-4683-890f-81b8901bd5db" containerName="oc" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.145730 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cac4081-5b91-4683-890f-81b8901bd5db" containerName="oc" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.146231 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533460-hdnlv" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.149899 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.150560 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.153391 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.161391 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533460-hdnlv"] Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.307518 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm566\" (UniqueName: \"kubernetes.io/projected/6dd60b00-0f0b-422a-971b-76336b1a813e-kube-api-access-hm566\") pod \"auto-csr-approver-29533460-hdnlv\" (UID: \"6dd60b00-0f0b-422a-971b-76336b1a813e\") " pod="openshift-infra/auto-csr-approver-29533460-hdnlv" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.409827 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm566\" (UniqueName: \"kubernetes.io/projected/6dd60b00-0f0b-422a-971b-76336b1a813e-kube-api-access-hm566\") pod \"auto-csr-approver-29533460-hdnlv\" (UID: \"6dd60b00-0f0b-422a-971b-76336b1a813e\") " pod="openshift-infra/auto-csr-approver-29533460-hdnlv" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.445564 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm566\" (UniqueName: \"kubernetes.io/projected/6dd60b00-0f0b-422a-971b-76336b1a813e-kube-api-access-hm566\") pod \"auto-csr-approver-29533460-hdnlv\" (UID: \"6dd60b00-0f0b-422a-971b-76336b1a813e\") " pod="openshift-infra/auto-csr-approver-29533460-hdnlv" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.475981 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533460-hdnlv" Feb 25 08:20:00 crc kubenswrapper[4978]: I0225 08:20:00.959516 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533460-hdnlv"] Feb 25 08:20:01 crc kubenswrapper[4978]: I0225 08:20:01.592521 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533460-hdnlv" event={"ID":"6dd60b00-0f0b-422a-971b-76336b1a813e","Type":"ContainerStarted","Data":"74361adf7fd3d57593cc6bf20dcc87bb574f3b58f26c97b7fcc0c2a8571e0f7a"} Feb 25 08:20:02 crc kubenswrapper[4978]: I0225 08:20:02.603455 4978 generic.go:334] "Generic (PLEG): container finished" podID="6dd60b00-0f0b-422a-971b-76336b1a813e" containerID="2e06c9d8615ca9e9cd9423b8ed2ac3cf958eed8793d36af984ed86cde1ea82e5" exitCode=0 Feb 25 08:20:02 crc kubenswrapper[4978]: I0225 08:20:02.603524 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533460-hdnlv" event={"ID":"6dd60b00-0f0b-422a-971b-76336b1a813e","Type":"ContainerDied","Data":"2e06c9d8615ca9e9cd9423b8ed2ac3cf958eed8793d36af984ed86cde1ea82e5"} Feb 25 08:20:03 crc kubenswrapper[4978]: I0225 08:20:03.940931 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533460-hdnlv" Feb 25 08:20:04 crc kubenswrapper[4978]: I0225 08:20:04.071469 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm566\" (UniqueName: \"kubernetes.io/projected/6dd60b00-0f0b-422a-971b-76336b1a813e-kube-api-access-hm566\") pod \"6dd60b00-0f0b-422a-971b-76336b1a813e\" (UID: \"6dd60b00-0f0b-422a-971b-76336b1a813e\") " Feb 25 08:20:04 crc kubenswrapper[4978]: I0225 08:20:04.077672 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd60b00-0f0b-422a-971b-76336b1a813e-kube-api-access-hm566" (OuterVolumeSpecName: "kube-api-access-hm566") pod "6dd60b00-0f0b-422a-971b-76336b1a813e" (UID: "6dd60b00-0f0b-422a-971b-76336b1a813e"). InnerVolumeSpecName "kube-api-access-hm566". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:20:04 crc kubenswrapper[4978]: I0225 08:20:04.174415 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm566\" (UniqueName: \"kubernetes.io/projected/6dd60b00-0f0b-422a-971b-76336b1a813e-kube-api-access-hm566\") on node \"crc\" DevicePath \"\"" Feb 25 08:20:04 crc kubenswrapper[4978]: I0225 08:20:04.629546 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533460-hdnlv" event={"ID":"6dd60b00-0f0b-422a-971b-76336b1a813e","Type":"ContainerDied","Data":"74361adf7fd3d57593cc6bf20dcc87bb574f3b58f26c97b7fcc0c2a8571e0f7a"} Feb 25 08:20:04 crc kubenswrapper[4978]: I0225 08:20:04.629659 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74361adf7fd3d57593cc6bf20dcc87bb574f3b58f26c97b7fcc0c2a8571e0f7a" Feb 25 08:20:04 crc kubenswrapper[4978]: I0225 08:20:04.629669 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533460-hdnlv" Feb 25 08:20:05 crc kubenswrapper[4978]: I0225 08:20:05.027518 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533454-h7xtm"] Feb 25 08:20:05 crc kubenswrapper[4978]: I0225 08:20:05.038058 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533454-h7xtm"] Feb 25 08:20:05 crc kubenswrapper[4978]: I0225 08:20:05.342585 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a128a7aa-f607-43bc-a2b0-a676531ffb51" path="/var/lib/kubelet/pods/a128a7aa-f607-43bc-a2b0-a676531ffb51/volumes" Feb 25 08:20:16 crc kubenswrapper[4978]: I0225 08:20:16.540617 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:20:16 crc kubenswrapper[4978]: I0225 08:20:16.541349 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:20:16 crc kubenswrapper[4978]: I0225 08:20:16.541476 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 08:20:16 crc kubenswrapper[4978]: I0225 08:20:16.542619 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 08:20:16 crc kubenswrapper[4978]: I0225 08:20:16.542887 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" gracePeriod=600 Feb 25 08:20:16 crc kubenswrapper[4978]: E0225 08:20:16.677234 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:20:16 crc kubenswrapper[4978]: I0225 08:20:16.767911 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" exitCode=0 Feb 25 08:20:16 crc kubenswrapper[4978]: I0225 08:20:16.767992 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6"} Feb 25 08:20:16 crc kubenswrapper[4978]: I0225 08:20:16.768048 4978 scope.go:117] "RemoveContainer" containerID="861a0fde7761354b7edadbf02215304f5389b12a48ceeb6beeb07e0a89855ed3" Feb 25 08:20:16 crc kubenswrapper[4978]: I0225 08:20:16.769235 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:20:16 crc kubenswrapper[4978]: E0225 08:20:16.770263 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:20:25 crc kubenswrapper[4978]: I0225 08:20:25.759125 4978 scope.go:117] "RemoveContainer" containerID="112e6e23840eb8f02a4f31874a2d519d065aa086a035368f7e870e96fc92feaa" Feb 25 08:20:30 crc kubenswrapper[4978]: I0225 08:20:30.328473 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:20:30 crc kubenswrapper[4978]: E0225 08:20:30.330870 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:20:41 crc kubenswrapper[4978]: I0225 08:20:41.328893 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:20:41 crc kubenswrapper[4978]: E0225 08:20:41.329890 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:20:54 crc kubenswrapper[4978]: I0225 08:20:54.328216 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:20:54 crc kubenswrapper[4978]: E0225 08:20:54.329197 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:21:06 crc kubenswrapper[4978]: I0225 08:21:06.328729 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:21:06 crc kubenswrapper[4978]: E0225 08:21:06.329490 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:21:20 crc kubenswrapper[4978]: I0225 08:21:20.328215 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:21:20 crc kubenswrapper[4978]: E0225 08:21:20.329624 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:21:31 crc kubenswrapper[4978]: I0225 08:21:31.327919 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:21:31 crc kubenswrapper[4978]: E0225 08:21:31.328932 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:21:43 crc kubenswrapper[4978]: I0225 08:21:43.327467 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:21:43 crc kubenswrapper[4978]: E0225 08:21:43.328148 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:21:54 crc kubenswrapper[4978]: I0225 08:21:54.327032 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:21:54 crc kubenswrapper[4978]: E0225 08:21:54.328139 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.161298 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533462-krmmf"] Feb 25 08:22:00 crc kubenswrapper[4978]: E0225 08:22:00.162345 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dd60b00-0f0b-422a-971b-76336b1a813e" containerName="oc" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.162361 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dd60b00-0f0b-422a-971b-76336b1a813e" containerName="oc" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.162595 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dd60b00-0f0b-422a-971b-76336b1a813e" containerName="oc" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.163114 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533462-krmmf" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.166982 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.167165 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.168319 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533462-krmmf"] Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.170734 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.331741 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jn9g9\" (UniqueName: \"kubernetes.io/projected/e000858b-4d20-478f-ba95-0b7cedad86af-kube-api-access-jn9g9\") pod \"auto-csr-approver-29533462-krmmf\" (UID: \"e000858b-4d20-478f-ba95-0b7cedad86af\") " pod="openshift-infra/auto-csr-approver-29533462-krmmf" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.433079 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jn9g9\" (UniqueName: \"kubernetes.io/projected/e000858b-4d20-478f-ba95-0b7cedad86af-kube-api-access-jn9g9\") pod \"auto-csr-approver-29533462-krmmf\" (UID: \"e000858b-4d20-478f-ba95-0b7cedad86af\") " pod="openshift-infra/auto-csr-approver-29533462-krmmf" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.457147 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jn9g9\" (UniqueName: \"kubernetes.io/projected/e000858b-4d20-478f-ba95-0b7cedad86af-kube-api-access-jn9g9\") pod \"auto-csr-approver-29533462-krmmf\" (UID: \"e000858b-4d20-478f-ba95-0b7cedad86af\") " pod="openshift-infra/auto-csr-approver-29533462-krmmf" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.490045 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533462-krmmf" Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.768239 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533462-krmmf"] Feb 25 08:22:00 crc kubenswrapper[4978]: I0225 08:22:00.788108 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 08:22:01 crc kubenswrapper[4978]: I0225 08:22:01.740316 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533462-krmmf" event={"ID":"e000858b-4d20-478f-ba95-0b7cedad86af","Type":"ContainerStarted","Data":"de22f896942a8f5f4cba130792ead2e603f27f0b3f8d93741ad5a19ba2fe6543"} Feb 25 08:22:02 crc kubenswrapper[4978]: I0225 08:22:02.750339 4978 generic.go:334] "Generic (PLEG): container finished" podID="e000858b-4d20-478f-ba95-0b7cedad86af" containerID="1b4554bf179c6a667b09d97ee14377c48477c4b5a8ea767a09366545ceaf3df8" exitCode=0 Feb 25 08:22:02 crc kubenswrapper[4978]: I0225 08:22:02.750446 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533462-krmmf" event={"ID":"e000858b-4d20-478f-ba95-0b7cedad86af","Type":"ContainerDied","Data":"1b4554bf179c6a667b09d97ee14377c48477c4b5a8ea767a09366545ceaf3df8"} Feb 25 08:22:04 crc kubenswrapper[4978]: I0225 08:22:04.138096 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533462-krmmf" Feb 25 08:22:04 crc kubenswrapper[4978]: I0225 08:22:04.309044 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jn9g9\" (UniqueName: \"kubernetes.io/projected/e000858b-4d20-478f-ba95-0b7cedad86af-kube-api-access-jn9g9\") pod \"e000858b-4d20-478f-ba95-0b7cedad86af\" (UID: \"e000858b-4d20-478f-ba95-0b7cedad86af\") " Feb 25 08:22:04 crc kubenswrapper[4978]: I0225 08:22:04.318525 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e000858b-4d20-478f-ba95-0b7cedad86af-kube-api-access-jn9g9" (OuterVolumeSpecName: "kube-api-access-jn9g9") pod "e000858b-4d20-478f-ba95-0b7cedad86af" (UID: "e000858b-4d20-478f-ba95-0b7cedad86af"). InnerVolumeSpecName "kube-api-access-jn9g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:22:04 crc kubenswrapper[4978]: I0225 08:22:04.411545 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jn9g9\" (UniqueName: \"kubernetes.io/projected/e000858b-4d20-478f-ba95-0b7cedad86af-kube-api-access-jn9g9\") on node \"crc\" DevicePath \"\"" Feb 25 08:22:04 crc kubenswrapper[4978]: I0225 08:22:04.768152 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533462-krmmf" event={"ID":"e000858b-4d20-478f-ba95-0b7cedad86af","Type":"ContainerDied","Data":"de22f896942a8f5f4cba130792ead2e603f27f0b3f8d93741ad5a19ba2fe6543"} Feb 25 08:22:04 crc kubenswrapper[4978]: I0225 08:22:04.768193 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de22f896942a8f5f4cba130792ead2e603f27f0b3f8d93741ad5a19ba2fe6543" Feb 25 08:22:04 crc kubenswrapper[4978]: I0225 08:22:04.768247 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533462-krmmf" Feb 25 08:22:05 crc kubenswrapper[4978]: I0225 08:22:05.233301 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533456-24g78"] Feb 25 08:22:05 crc kubenswrapper[4978]: I0225 08:22:05.239680 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533456-24g78"] Feb 25 08:22:05 crc kubenswrapper[4978]: I0225 08:22:05.336558 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93f60f49-8d56-4d2d-b73f-40e5a5aa6d03" path="/var/lib/kubelet/pods/93f60f49-8d56-4d2d-b73f-40e5a5aa6d03/volumes" Feb 25 08:22:08 crc kubenswrapper[4978]: I0225 08:22:08.327610 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:22:08 crc kubenswrapper[4978]: E0225 08:22:08.328440 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:22:23 crc kubenswrapper[4978]: I0225 08:22:23.328339 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:22:23 crc kubenswrapper[4978]: E0225 08:22:23.331459 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:22:25 crc kubenswrapper[4978]: I0225 08:22:25.862046 4978 scope.go:117] "RemoveContainer" containerID="8f4b03a6bf001d971345160007c85b199ea7395f34df46f7c9fcb54e4b40774d" Feb 25 08:22:35 crc kubenswrapper[4978]: I0225 08:22:35.328351 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:22:35 crc kubenswrapper[4978]: E0225 08:22:35.329349 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.460252 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qrfg4"] Feb 25 08:22:41 crc kubenswrapper[4978]: E0225 08:22:41.461321 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e000858b-4d20-478f-ba95-0b7cedad86af" containerName="oc" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.461344 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e000858b-4d20-478f-ba95-0b7cedad86af" containerName="oc" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.461704 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e000858b-4d20-478f-ba95-0b7cedad86af" containerName="oc" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.463622 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.481643 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qrfg4"] Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.604919 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-utilities\") pod \"certified-operators-qrfg4\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.605042 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxxp9\" (UniqueName: \"kubernetes.io/projected/a7f7bc69-4268-4b78-8c53-b8c41a30a221-kube-api-access-fxxp9\") pod \"certified-operators-qrfg4\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.605091 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-catalog-content\") pod \"certified-operators-qrfg4\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.706404 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-utilities\") pod \"certified-operators-qrfg4\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.706473 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxxp9\" (UniqueName: \"kubernetes.io/projected/a7f7bc69-4268-4b78-8c53-b8c41a30a221-kube-api-access-fxxp9\") pod \"certified-operators-qrfg4\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.706506 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-catalog-content\") pod \"certified-operators-qrfg4\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.706979 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-catalog-content\") pod \"certified-operators-qrfg4\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.707071 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-utilities\") pod \"certified-operators-qrfg4\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.739325 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxxp9\" (UniqueName: \"kubernetes.io/projected/a7f7bc69-4268-4b78-8c53-b8c41a30a221-kube-api-access-fxxp9\") pod \"certified-operators-qrfg4\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:41 crc kubenswrapper[4978]: I0225 08:22:41.806619 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:42 crc kubenswrapper[4978]: I0225 08:22:42.225092 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qrfg4"] Feb 25 08:22:43 crc kubenswrapper[4978]: I0225 08:22:43.135140 4978 generic.go:334] "Generic (PLEG): container finished" podID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerID="e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58" exitCode=0 Feb 25 08:22:43 crc kubenswrapper[4978]: I0225 08:22:43.135220 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrfg4" event={"ID":"a7f7bc69-4268-4b78-8c53-b8c41a30a221","Type":"ContainerDied","Data":"e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58"} Feb 25 08:22:43 crc kubenswrapper[4978]: I0225 08:22:43.135668 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrfg4" event={"ID":"a7f7bc69-4268-4b78-8c53-b8c41a30a221","Type":"ContainerStarted","Data":"20d469882bf2f50e4059f9c857a917f5b07f7a5d3a3255de28dab34c7f136478"} Feb 25 08:22:44 crc kubenswrapper[4978]: I0225 08:22:44.159733 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrfg4" event={"ID":"a7f7bc69-4268-4b78-8c53-b8c41a30a221","Type":"ContainerStarted","Data":"152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad"} Feb 25 08:22:45 crc kubenswrapper[4978]: I0225 08:22:45.174048 4978 generic.go:334] "Generic (PLEG): container finished" podID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerID="152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad" exitCode=0 Feb 25 08:22:45 crc kubenswrapper[4978]: I0225 08:22:45.174113 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrfg4" event={"ID":"a7f7bc69-4268-4b78-8c53-b8c41a30a221","Type":"ContainerDied","Data":"152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad"} Feb 25 08:22:46 crc kubenswrapper[4978]: I0225 08:22:46.186457 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrfg4" event={"ID":"a7f7bc69-4268-4b78-8c53-b8c41a30a221","Type":"ContainerStarted","Data":"fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32"} Feb 25 08:22:46 crc kubenswrapper[4978]: I0225 08:22:46.216279 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qrfg4" podStartSLOduration=2.726642976 podStartE2EDuration="5.216252566s" podCreationTimestamp="2026-02-25 08:22:41 +0000 UTC" firstStartedPulling="2026-02-25 08:22:43.139104446 +0000 UTC m=+5856.578360945" lastFinishedPulling="2026-02-25 08:22:45.628714046 +0000 UTC m=+5859.067970535" observedRunningTime="2026-02-25 08:22:46.210674933 +0000 UTC m=+5859.649931402" watchObservedRunningTime="2026-02-25 08:22:46.216252566 +0000 UTC m=+5859.655509065" Feb 25 08:22:48 crc kubenswrapper[4978]: I0225 08:22:48.861083 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8fbks"] Feb 25 08:22:48 crc kubenswrapper[4978]: I0225 08:22:48.863588 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:48 crc kubenswrapper[4978]: I0225 08:22:48.879630 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8fbks"] Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.019858 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-catalog-content\") pod \"redhat-operators-8fbks\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.019898 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-utilities\") pod \"redhat-operators-8fbks\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.019937 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfcnv\" (UniqueName: \"kubernetes.io/projected/e5924b1b-8fef-40dd-98bb-678564d6c80f-kube-api-access-mfcnv\") pod \"redhat-operators-8fbks\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.121545 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-catalog-content\") pod \"redhat-operators-8fbks\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.121597 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-utilities\") pod \"redhat-operators-8fbks\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.121634 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfcnv\" (UniqueName: \"kubernetes.io/projected/e5924b1b-8fef-40dd-98bb-678564d6c80f-kube-api-access-mfcnv\") pod \"redhat-operators-8fbks\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.122472 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-catalog-content\") pod \"redhat-operators-8fbks\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.123240 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-utilities\") pod \"redhat-operators-8fbks\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.154424 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfcnv\" (UniqueName: \"kubernetes.io/projected/e5924b1b-8fef-40dd-98bb-678564d6c80f-kube-api-access-mfcnv\") pod \"redhat-operators-8fbks\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.189700 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:49 crc kubenswrapper[4978]: I0225 08:22:49.431365 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8fbks"] Feb 25 08:22:50 crc kubenswrapper[4978]: I0225 08:22:50.225153 4978 generic.go:334] "Generic (PLEG): container finished" podID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerID="ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9" exitCode=0 Feb 25 08:22:50 crc kubenswrapper[4978]: I0225 08:22:50.225198 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fbks" event={"ID":"e5924b1b-8fef-40dd-98bb-678564d6c80f","Type":"ContainerDied","Data":"ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9"} Feb 25 08:22:50 crc kubenswrapper[4978]: I0225 08:22:50.225238 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fbks" event={"ID":"e5924b1b-8fef-40dd-98bb-678564d6c80f","Type":"ContainerStarted","Data":"98323940f278a9bec235073c77f9da3c863736a80a845db890cbc2014309114f"} Feb 25 08:22:50 crc kubenswrapper[4978]: I0225 08:22:50.329094 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:22:50 crc kubenswrapper[4978]: E0225 08:22:50.330114 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:22:51 crc kubenswrapper[4978]: I0225 08:22:51.234234 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fbks" event={"ID":"e5924b1b-8fef-40dd-98bb-678564d6c80f","Type":"ContainerStarted","Data":"49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853"} Feb 25 08:22:51 crc kubenswrapper[4978]: I0225 08:22:51.807577 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:51 crc kubenswrapper[4978]: I0225 08:22:51.807644 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:51 crc kubenswrapper[4978]: I0225 08:22:51.861250 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:52 crc kubenswrapper[4978]: I0225 08:22:52.247349 4978 generic.go:334] "Generic (PLEG): container finished" podID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerID="49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853" exitCode=0 Feb 25 08:22:52 crc kubenswrapper[4978]: I0225 08:22:52.247445 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fbks" event={"ID":"e5924b1b-8fef-40dd-98bb-678564d6c80f","Type":"ContainerDied","Data":"49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853"} Feb 25 08:22:52 crc kubenswrapper[4978]: I0225 08:22:52.326967 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:53 crc kubenswrapper[4978]: I0225 08:22:53.259510 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fbks" event={"ID":"e5924b1b-8fef-40dd-98bb-678564d6c80f","Type":"ContainerStarted","Data":"bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce"} Feb 25 08:22:53 crc kubenswrapper[4978]: I0225 08:22:53.302468 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8fbks" podStartSLOduration=2.863557165 podStartE2EDuration="5.302365795s" podCreationTimestamp="2026-02-25 08:22:48 +0000 UTC" firstStartedPulling="2026-02-25 08:22:50.226935035 +0000 UTC m=+5863.666191494" lastFinishedPulling="2026-02-25 08:22:52.665743665 +0000 UTC m=+5866.105000124" observedRunningTime="2026-02-25 08:22:53.295316546 +0000 UTC m=+5866.734573005" watchObservedRunningTime="2026-02-25 08:22:53.302365795 +0000 UTC m=+5866.741622284" Feb 25 08:22:54 crc kubenswrapper[4978]: I0225 08:22:54.237780 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qrfg4"] Feb 25 08:22:54 crc kubenswrapper[4978]: I0225 08:22:54.269535 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qrfg4" podUID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerName="registry-server" containerID="cri-o://fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32" gracePeriod=2 Feb 25 08:22:54 crc kubenswrapper[4978]: I0225 08:22:54.806064 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:54 crc kubenswrapper[4978]: I0225 08:22:54.919605 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxxp9\" (UniqueName: \"kubernetes.io/projected/a7f7bc69-4268-4b78-8c53-b8c41a30a221-kube-api-access-fxxp9\") pod \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " Feb 25 08:22:54 crc kubenswrapper[4978]: I0225 08:22:54.919768 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-utilities\") pod \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " Feb 25 08:22:54 crc kubenswrapper[4978]: I0225 08:22:54.919973 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-catalog-content\") pod \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\" (UID: \"a7f7bc69-4268-4b78-8c53-b8c41a30a221\") " Feb 25 08:22:54 crc kubenswrapper[4978]: I0225 08:22:54.920781 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-utilities" (OuterVolumeSpecName: "utilities") pod "a7f7bc69-4268-4b78-8c53-b8c41a30a221" (UID: "a7f7bc69-4268-4b78-8c53-b8c41a30a221"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:22:54 crc kubenswrapper[4978]: I0225 08:22:54.930730 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f7bc69-4268-4b78-8c53-b8c41a30a221-kube-api-access-fxxp9" (OuterVolumeSpecName: "kube-api-access-fxxp9") pod "a7f7bc69-4268-4b78-8c53-b8c41a30a221" (UID: "a7f7bc69-4268-4b78-8c53-b8c41a30a221"). InnerVolumeSpecName "kube-api-access-fxxp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:22:54 crc kubenswrapper[4978]: I0225 08:22:54.993950 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7f7bc69-4268-4b78-8c53-b8c41a30a221" (UID: "a7f7bc69-4268-4b78-8c53-b8c41a30a221"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.022495 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxxp9\" (UniqueName: \"kubernetes.io/projected/a7f7bc69-4268-4b78-8c53-b8c41a30a221-kube-api-access-fxxp9\") on node \"crc\" DevicePath \"\"" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.022544 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.022564 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f7bc69-4268-4b78-8c53-b8c41a30a221-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.282508 4978 generic.go:334] "Generic (PLEG): container finished" podID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerID="fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32" exitCode=0 Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.282580 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrfg4" event={"ID":"a7f7bc69-4268-4b78-8c53-b8c41a30a221","Type":"ContainerDied","Data":"fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32"} Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.282620 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qrfg4" event={"ID":"a7f7bc69-4268-4b78-8c53-b8c41a30a221","Type":"ContainerDied","Data":"20d469882bf2f50e4059f9c857a917f5b07f7a5d3a3255de28dab34c7f136478"} Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.282648 4978 scope.go:117] "RemoveContainer" containerID="fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.282843 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qrfg4" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.314897 4978 scope.go:117] "RemoveContainer" containerID="152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.348530 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qrfg4"] Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.348587 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qrfg4"] Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.350008 4978 scope.go:117] "RemoveContainer" containerID="e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.386720 4978 scope.go:117] "RemoveContainer" containerID="fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32" Feb 25 08:22:55 crc kubenswrapper[4978]: E0225 08:22:55.387273 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32\": container with ID starting with fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32 not found: ID does not exist" containerID="fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.387323 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32"} err="failed to get container status \"fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32\": rpc error: code = NotFound desc = could not find container \"fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32\": container with ID starting with fdca79938235cbccc44a99d20f51b21ccd451d1119e5654b2fd76ae5201e4c32 not found: ID does not exist" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.387353 4978 scope.go:117] "RemoveContainer" containerID="152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad" Feb 25 08:22:55 crc kubenswrapper[4978]: E0225 08:22:55.387802 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad\": container with ID starting with 152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad not found: ID does not exist" containerID="152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.387829 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad"} err="failed to get container status \"152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad\": rpc error: code = NotFound desc = could not find container \"152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad\": container with ID starting with 152da4f338dc036b6fa612b627704b05b239845ce4a973e05228784d891dedad not found: ID does not exist" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.387845 4978 scope.go:117] "RemoveContainer" containerID="e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58" Feb 25 08:22:55 crc kubenswrapper[4978]: E0225 08:22:55.388123 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58\": container with ID starting with e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58 not found: ID does not exist" containerID="e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58" Feb 25 08:22:55 crc kubenswrapper[4978]: I0225 08:22:55.388148 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58"} err="failed to get container status \"e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58\": rpc error: code = NotFound desc = could not find container \"e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58\": container with ID starting with e35fdc3c55390cc54663e85660af44f5ddf09e861b21a2c937e99d02e72bee58 not found: ID does not exist" Feb 25 08:22:57 crc kubenswrapper[4978]: I0225 08:22:57.340863 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" path="/var/lib/kubelet/pods/a7f7bc69-4268-4b78-8c53-b8c41a30a221/volumes" Feb 25 08:22:59 crc kubenswrapper[4978]: I0225 08:22:59.190318 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:22:59 crc kubenswrapper[4978]: I0225 08:22:59.190857 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:23:00 crc kubenswrapper[4978]: I0225 08:23:00.255982 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8fbks" podUID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerName="registry-server" probeResult="failure" output=< Feb 25 08:23:00 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 08:23:00 crc kubenswrapper[4978]: > Feb 25 08:23:05 crc kubenswrapper[4978]: I0225 08:23:05.328631 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:23:05 crc kubenswrapper[4978]: E0225 08:23:05.329024 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:23:09 crc kubenswrapper[4978]: I0225 08:23:09.254110 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:23:09 crc kubenswrapper[4978]: I0225 08:23:09.322766 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:23:09 crc kubenswrapper[4978]: I0225 08:23:09.503979 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8fbks"] Feb 25 08:23:10 crc kubenswrapper[4978]: I0225 08:23:10.438168 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8fbks" podUID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerName="registry-server" containerID="cri-o://bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce" gracePeriod=2 Feb 25 08:23:10 crc kubenswrapper[4978]: I0225 08:23:10.887562 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:23:10 crc kubenswrapper[4978]: I0225 08:23:10.982748 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-catalog-content\") pod \"e5924b1b-8fef-40dd-98bb-678564d6c80f\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " Feb 25 08:23:10 crc kubenswrapper[4978]: I0225 08:23:10.982858 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfcnv\" (UniqueName: \"kubernetes.io/projected/e5924b1b-8fef-40dd-98bb-678564d6c80f-kube-api-access-mfcnv\") pod \"e5924b1b-8fef-40dd-98bb-678564d6c80f\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " Feb 25 08:23:10 crc kubenswrapper[4978]: I0225 08:23:10.982892 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-utilities\") pod \"e5924b1b-8fef-40dd-98bb-678564d6c80f\" (UID: \"e5924b1b-8fef-40dd-98bb-678564d6c80f\") " Feb 25 08:23:10 crc kubenswrapper[4978]: I0225 08:23:10.984683 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-utilities" (OuterVolumeSpecName: "utilities") pod "e5924b1b-8fef-40dd-98bb-678564d6c80f" (UID: "e5924b1b-8fef-40dd-98bb-678564d6c80f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:23:10 crc kubenswrapper[4978]: I0225 08:23:10.988449 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5924b1b-8fef-40dd-98bb-678564d6c80f-kube-api-access-mfcnv" (OuterVolumeSpecName: "kube-api-access-mfcnv") pod "e5924b1b-8fef-40dd-98bb-678564d6c80f" (UID: "e5924b1b-8fef-40dd-98bb-678564d6c80f"). InnerVolumeSpecName "kube-api-access-mfcnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.085950 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfcnv\" (UniqueName: \"kubernetes.io/projected/e5924b1b-8fef-40dd-98bb-678564d6c80f-kube-api-access-mfcnv\") on node \"crc\" DevicePath \"\"" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.086475 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.189355 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5924b1b-8fef-40dd-98bb-678564d6c80f" (UID: "e5924b1b-8fef-40dd-98bb-678564d6c80f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.289145 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5924b1b-8fef-40dd-98bb-678564d6c80f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.453666 4978 generic.go:334] "Generic (PLEG): container finished" podID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerID="bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce" exitCode=0 Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.453728 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fbks" event={"ID":"e5924b1b-8fef-40dd-98bb-678564d6c80f","Type":"ContainerDied","Data":"bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce"} Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.453768 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8fbks" event={"ID":"e5924b1b-8fef-40dd-98bb-678564d6c80f","Type":"ContainerDied","Data":"98323940f278a9bec235073c77f9da3c863736a80a845db890cbc2014309114f"} Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.453797 4978 scope.go:117] "RemoveContainer" containerID="bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.455110 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8fbks" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.495247 4978 scope.go:117] "RemoveContainer" containerID="49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.504461 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8fbks"] Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.513975 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8fbks"] Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.529330 4978 scope.go:117] "RemoveContainer" containerID="ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9" Feb 25 08:23:11 crc kubenswrapper[4978]: E0225 08:23:11.539806 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5924b1b_8fef_40dd_98bb_678564d6c80f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5924b1b_8fef_40dd_98bb_678564d6c80f.slice/crio-98323940f278a9bec235073c77f9da3c863736a80a845db890cbc2014309114f\": RecentStats: unable to find data in memory cache]" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.571690 4978 scope.go:117] "RemoveContainer" containerID="bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce" Feb 25 08:23:11 crc kubenswrapper[4978]: E0225 08:23:11.573218 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce\": container with ID starting with bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce not found: ID does not exist" containerID="bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.573266 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce"} err="failed to get container status \"bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce\": rpc error: code = NotFound desc = could not find container \"bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce\": container with ID starting with bd2f7847bf262fbe81ef10055d5ccdc1e224ce2fb9e9ce34b87656207da9ddce not found: ID does not exist" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.573297 4978 scope.go:117] "RemoveContainer" containerID="49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853" Feb 25 08:23:11 crc kubenswrapper[4978]: E0225 08:23:11.574202 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853\": container with ID starting with 49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853 not found: ID does not exist" containerID="49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.574244 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853"} err="failed to get container status \"49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853\": rpc error: code = NotFound desc = could not find container \"49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853\": container with ID starting with 49888eef545e00e5ae10ccff1c8a1c7da6c6d0fc7217c02c1963b0f3352c2853 not found: ID does not exist" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.574271 4978 scope.go:117] "RemoveContainer" containerID="ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9" Feb 25 08:23:11 crc kubenswrapper[4978]: E0225 08:23:11.574978 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9\": container with ID starting with ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9 not found: ID does not exist" containerID="ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9" Feb 25 08:23:11 crc kubenswrapper[4978]: I0225 08:23:11.575003 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9"} err="failed to get container status \"ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9\": rpc error: code = NotFound desc = could not find container \"ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9\": container with ID starting with ec731149a3be3214271e6ec9f2266d19a10ad0e217cc23f03c47f11fee2d72f9 not found: ID does not exist" Feb 25 08:23:13 crc kubenswrapper[4978]: I0225 08:23:13.358613 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5924b1b-8fef-40dd-98bb-678564d6c80f" path="/var/lib/kubelet/pods/e5924b1b-8fef-40dd-98bb-678564d6c80f/volumes" Feb 25 08:23:18 crc kubenswrapper[4978]: I0225 08:23:18.327563 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:23:18 crc kubenswrapper[4978]: E0225 08:23:18.328543 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:23:29 crc kubenswrapper[4978]: I0225 08:23:29.328409 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:23:29 crc kubenswrapper[4978]: E0225 08:23:29.329988 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:23:41 crc kubenswrapper[4978]: I0225 08:23:41.328227 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:23:41 crc kubenswrapper[4978]: E0225 08:23:41.329360 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:23:54 crc kubenswrapper[4978]: I0225 08:23:54.327809 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:23:54 crc kubenswrapper[4978]: E0225 08:23:54.328566 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.147867 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533464-kpfv7"] Feb 25 08:24:00 crc kubenswrapper[4978]: E0225 08:24:00.148506 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerName="extract-utilities" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.148522 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerName="extract-utilities" Feb 25 08:24:00 crc kubenswrapper[4978]: E0225 08:24:00.148539 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerName="extract-content" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.148546 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerName="extract-content" Feb 25 08:24:00 crc kubenswrapper[4978]: E0225 08:24:00.148570 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerName="extract-utilities" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.148580 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerName="extract-utilities" Feb 25 08:24:00 crc kubenswrapper[4978]: E0225 08:24:00.148599 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerName="registry-server" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.148607 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerName="registry-server" Feb 25 08:24:00 crc kubenswrapper[4978]: E0225 08:24:00.148629 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerName="extract-content" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.148640 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerName="extract-content" Feb 25 08:24:00 crc kubenswrapper[4978]: E0225 08:24:00.148657 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerName="registry-server" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.148664 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerName="registry-server" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.148829 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f7bc69-4268-4b78-8c53-b8c41a30a221" containerName="registry-server" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.148874 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5924b1b-8fef-40dd-98bb-678564d6c80f" containerName="registry-server" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.149431 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533464-kpfv7" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.151725 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.151729 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.152457 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.165273 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533464-kpfv7"] Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.251759 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjbll\" (UniqueName: \"kubernetes.io/projected/1f99497f-94dc-4976-96c9-60eba8b5c535-kube-api-access-zjbll\") pod \"auto-csr-approver-29533464-kpfv7\" (UID: \"1f99497f-94dc-4976-96c9-60eba8b5c535\") " pod="openshift-infra/auto-csr-approver-29533464-kpfv7" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.352975 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjbll\" (UniqueName: \"kubernetes.io/projected/1f99497f-94dc-4976-96c9-60eba8b5c535-kube-api-access-zjbll\") pod \"auto-csr-approver-29533464-kpfv7\" (UID: \"1f99497f-94dc-4976-96c9-60eba8b5c535\") " pod="openshift-infra/auto-csr-approver-29533464-kpfv7" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.378830 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjbll\" (UniqueName: \"kubernetes.io/projected/1f99497f-94dc-4976-96c9-60eba8b5c535-kube-api-access-zjbll\") pod \"auto-csr-approver-29533464-kpfv7\" (UID: \"1f99497f-94dc-4976-96c9-60eba8b5c535\") " pod="openshift-infra/auto-csr-approver-29533464-kpfv7" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.484610 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533464-kpfv7" Feb 25 08:24:00 crc kubenswrapper[4978]: I0225 08:24:00.988200 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533464-kpfv7"] Feb 25 08:24:01 crc kubenswrapper[4978]: W0225 08:24:01.001486 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f99497f_94dc_4976_96c9_60eba8b5c535.slice/crio-2c2b3b372242f52ac5e0b5d10c55b3a7e78d5bf2e1ddd7b60cfb043291a8e6fa WatchSource:0}: Error finding container 2c2b3b372242f52ac5e0b5d10c55b3a7e78d5bf2e1ddd7b60cfb043291a8e6fa: Status 404 returned error can't find the container with id 2c2b3b372242f52ac5e0b5d10c55b3a7e78d5bf2e1ddd7b60cfb043291a8e6fa Feb 25 08:24:01 crc kubenswrapper[4978]: I0225 08:24:01.874971 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533464-kpfv7" event={"ID":"1f99497f-94dc-4976-96c9-60eba8b5c535","Type":"ContainerStarted","Data":"2c2b3b372242f52ac5e0b5d10c55b3a7e78d5bf2e1ddd7b60cfb043291a8e6fa"} Feb 25 08:24:02 crc kubenswrapper[4978]: I0225 08:24:02.882338 4978 generic.go:334] "Generic (PLEG): container finished" podID="1f99497f-94dc-4976-96c9-60eba8b5c535" containerID="6086e2d7bd33ed66ba3c52852b533b074ebe8596f06e09263102928f250cdd62" exitCode=0 Feb 25 08:24:02 crc kubenswrapper[4978]: I0225 08:24:02.882415 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533464-kpfv7" event={"ID":"1f99497f-94dc-4976-96c9-60eba8b5c535","Type":"ContainerDied","Data":"6086e2d7bd33ed66ba3c52852b533b074ebe8596f06e09263102928f250cdd62"} Feb 25 08:24:04 crc kubenswrapper[4978]: I0225 08:24:04.182283 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533464-kpfv7" Feb 25 08:24:04 crc kubenswrapper[4978]: I0225 08:24:04.226214 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjbll\" (UniqueName: \"kubernetes.io/projected/1f99497f-94dc-4976-96c9-60eba8b5c535-kube-api-access-zjbll\") pod \"1f99497f-94dc-4976-96c9-60eba8b5c535\" (UID: \"1f99497f-94dc-4976-96c9-60eba8b5c535\") " Feb 25 08:24:04 crc kubenswrapper[4978]: I0225 08:24:04.234822 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f99497f-94dc-4976-96c9-60eba8b5c535-kube-api-access-zjbll" (OuterVolumeSpecName: "kube-api-access-zjbll") pod "1f99497f-94dc-4976-96c9-60eba8b5c535" (UID: "1f99497f-94dc-4976-96c9-60eba8b5c535"). InnerVolumeSpecName "kube-api-access-zjbll". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:24:04 crc kubenswrapper[4978]: I0225 08:24:04.328108 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjbll\" (UniqueName: \"kubernetes.io/projected/1f99497f-94dc-4976-96c9-60eba8b5c535-kube-api-access-zjbll\") on node \"crc\" DevicePath \"\"" Feb 25 08:24:04 crc kubenswrapper[4978]: I0225 08:24:04.902880 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533464-kpfv7" event={"ID":"1f99497f-94dc-4976-96c9-60eba8b5c535","Type":"ContainerDied","Data":"2c2b3b372242f52ac5e0b5d10c55b3a7e78d5bf2e1ddd7b60cfb043291a8e6fa"} Feb 25 08:24:04 crc kubenswrapper[4978]: I0225 08:24:04.902939 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c2b3b372242f52ac5e0b5d10c55b3a7e78d5bf2e1ddd7b60cfb043291a8e6fa" Feb 25 08:24:04 crc kubenswrapper[4978]: I0225 08:24:04.902964 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533464-kpfv7" Feb 25 08:24:05 crc kubenswrapper[4978]: I0225 08:24:05.312653 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533458-92945"] Feb 25 08:24:05 crc kubenswrapper[4978]: I0225 08:24:05.322981 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533458-92945"] Feb 25 08:24:05 crc kubenswrapper[4978]: I0225 08:24:05.327764 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:24:05 crc kubenswrapper[4978]: E0225 08:24:05.328100 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:24:05 crc kubenswrapper[4978]: I0225 08:24:05.339129 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cac4081-5b91-4683-890f-81b8901bd5db" path="/var/lib/kubelet/pods/6cac4081-5b91-4683-890f-81b8901bd5db/volumes" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.652869 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-5nhcj"] Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.663312 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-5nhcj"] Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.795714 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-hgb99"] Feb 25 08:24:09 crc kubenswrapper[4978]: E0225 08:24:09.796004 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f99497f-94dc-4976-96c9-60eba8b5c535" containerName="oc" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.796016 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f99497f-94dc-4976-96c9-60eba8b5c535" containerName="oc" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.796153 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f99497f-94dc-4976-96c9-60eba8b5c535" containerName="oc" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.796580 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.800424 4978 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-7m2v4" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.800664 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.800835 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.801080 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.826631 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-hgb99"] Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.916858 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/812202c5-cfd7-47c8-be68-e3467786a74e-crc-storage\") pod \"crc-storage-crc-hgb99\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.916938 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/812202c5-cfd7-47c8-be68-e3467786a74e-node-mnt\") pod \"crc-storage-crc-hgb99\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:09 crc kubenswrapper[4978]: I0225 08:24:09.916966 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92jpg\" (UniqueName: \"kubernetes.io/projected/812202c5-cfd7-47c8-be68-e3467786a74e-kube-api-access-92jpg\") pod \"crc-storage-crc-hgb99\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:10 crc kubenswrapper[4978]: I0225 08:24:10.017966 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/812202c5-cfd7-47c8-be68-e3467786a74e-crc-storage\") pod \"crc-storage-crc-hgb99\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:10 crc kubenswrapper[4978]: I0225 08:24:10.018044 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/812202c5-cfd7-47c8-be68-e3467786a74e-node-mnt\") pod \"crc-storage-crc-hgb99\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:10 crc kubenswrapper[4978]: I0225 08:24:10.018072 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92jpg\" (UniqueName: \"kubernetes.io/projected/812202c5-cfd7-47c8-be68-e3467786a74e-kube-api-access-92jpg\") pod \"crc-storage-crc-hgb99\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:10 crc kubenswrapper[4978]: I0225 08:24:10.018543 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/812202c5-cfd7-47c8-be68-e3467786a74e-node-mnt\") pod \"crc-storage-crc-hgb99\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:10 crc kubenswrapper[4978]: I0225 08:24:10.019164 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/812202c5-cfd7-47c8-be68-e3467786a74e-crc-storage\") pod \"crc-storage-crc-hgb99\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:10 crc kubenswrapper[4978]: I0225 08:24:10.040180 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92jpg\" (UniqueName: \"kubernetes.io/projected/812202c5-cfd7-47c8-be68-e3467786a74e-kube-api-access-92jpg\") pod \"crc-storage-crc-hgb99\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:10 crc kubenswrapper[4978]: I0225 08:24:10.115645 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:10 crc kubenswrapper[4978]: I0225 08:24:10.641023 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-hgb99"] Feb 25 08:24:10 crc kubenswrapper[4978]: I0225 08:24:10.949521 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-hgb99" event={"ID":"812202c5-cfd7-47c8-be68-e3467786a74e","Type":"ContainerStarted","Data":"7b19626a0e0bb0d7430d1b46ec85c9fc2995ad5f3000b8eff8e46d05de1bb87e"} Feb 25 08:24:11 crc kubenswrapper[4978]: I0225 08:24:11.354091 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ef89fc-bdd8-4513-b546-214c92481e2b" path="/var/lib/kubelet/pods/09ef89fc-bdd8-4513-b546-214c92481e2b/volumes" Feb 25 08:24:11 crc kubenswrapper[4978]: I0225 08:24:11.956312 4978 generic.go:334] "Generic (PLEG): container finished" podID="812202c5-cfd7-47c8-be68-e3467786a74e" containerID="bd179ff7d61dd9b40b85a2b223bc598ac1a749ff1926d9aae5647a98bc03c808" exitCode=0 Feb 25 08:24:11 crc kubenswrapper[4978]: I0225 08:24:11.956356 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-hgb99" event={"ID":"812202c5-cfd7-47c8-be68-e3467786a74e","Type":"ContainerDied","Data":"bd179ff7d61dd9b40b85a2b223bc598ac1a749ff1926d9aae5647a98bc03c808"} Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.807063 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.974181 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-hgb99" event={"ID":"812202c5-cfd7-47c8-be68-e3467786a74e","Type":"ContainerDied","Data":"7b19626a0e0bb0d7430d1b46ec85c9fc2995ad5f3000b8eff8e46d05de1bb87e"} Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.974664 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b19626a0e0bb0d7430d1b46ec85c9fc2995ad5f3000b8eff8e46d05de1bb87e" Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.974269 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-hgb99" Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.977649 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/812202c5-cfd7-47c8-be68-e3467786a74e-crc-storage\") pod \"812202c5-cfd7-47c8-be68-e3467786a74e\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.977891 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92jpg\" (UniqueName: \"kubernetes.io/projected/812202c5-cfd7-47c8-be68-e3467786a74e-kube-api-access-92jpg\") pod \"812202c5-cfd7-47c8-be68-e3467786a74e\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.977934 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/812202c5-cfd7-47c8-be68-e3467786a74e-node-mnt\") pod \"812202c5-cfd7-47c8-be68-e3467786a74e\" (UID: \"812202c5-cfd7-47c8-be68-e3467786a74e\") " Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.978117 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/812202c5-cfd7-47c8-be68-e3467786a74e-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "812202c5-cfd7-47c8-be68-e3467786a74e" (UID: "812202c5-cfd7-47c8-be68-e3467786a74e"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.978464 4978 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/812202c5-cfd7-47c8-be68-e3467786a74e-node-mnt\") on node \"crc\" DevicePath \"\"" Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.984338 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/812202c5-cfd7-47c8-be68-e3467786a74e-kube-api-access-92jpg" (OuterVolumeSpecName: "kube-api-access-92jpg") pod "812202c5-cfd7-47c8-be68-e3467786a74e" (UID: "812202c5-cfd7-47c8-be68-e3467786a74e"). InnerVolumeSpecName "kube-api-access-92jpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:24:13 crc kubenswrapper[4978]: I0225 08:24:13.994826 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/812202c5-cfd7-47c8-be68-e3467786a74e-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "812202c5-cfd7-47c8-be68-e3467786a74e" (UID: "812202c5-cfd7-47c8-be68-e3467786a74e"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:24:14 crc kubenswrapper[4978]: I0225 08:24:14.080441 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92jpg\" (UniqueName: \"kubernetes.io/projected/812202c5-cfd7-47c8-be68-e3467786a74e-kube-api-access-92jpg\") on node \"crc\" DevicePath \"\"" Feb 25 08:24:14 crc kubenswrapper[4978]: I0225 08:24:14.080495 4978 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/812202c5-cfd7-47c8-be68-e3467786a74e-crc-storage\") on node \"crc\" DevicePath \"\"" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.100994 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-hgb99"] Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.111538 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-hgb99"] Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.250846 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-58jng"] Feb 25 08:24:16 crc kubenswrapper[4978]: E0225 08:24:16.251191 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="812202c5-cfd7-47c8-be68-e3467786a74e" containerName="storage" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.251208 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="812202c5-cfd7-47c8-be68-e3467786a74e" containerName="storage" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.251389 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="812202c5-cfd7-47c8-be68-e3467786a74e" containerName="storage" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.251911 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.253706 4978 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-7m2v4" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.259029 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.259035 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.260576 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.273597 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-58jng"] Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.416485 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a38283e5-ec78-480b-a04b-18f039f316f5-node-mnt\") pod \"crc-storage-crc-58jng\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.417065 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a38283e5-ec78-480b-a04b-18f039f316f5-crc-storage\") pod \"crc-storage-crc-58jng\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.417327 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zbwd\" (UniqueName: \"kubernetes.io/projected/a38283e5-ec78-480b-a04b-18f039f316f5-kube-api-access-4zbwd\") pod \"crc-storage-crc-58jng\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.518665 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a38283e5-ec78-480b-a04b-18f039f316f5-crc-storage\") pod \"crc-storage-crc-58jng\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.518828 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zbwd\" (UniqueName: \"kubernetes.io/projected/a38283e5-ec78-480b-a04b-18f039f316f5-kube-api-access-4zbwd\") pod \"crc-storage-crc-58jng\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.518900 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a38283e5-ec78-480b-a04b-18f039f316f5-node-mnt\") pod \"crc-storage-crc-58jng\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.519358 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a38283e5-ec78-480b-a04b-18f039f316f5-node-mnt\") pod \"crc-storage-crc-58jng\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.520703 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a38283e5-ec78-480b-a04b-18f039f316f5-crc-storage\") pod \"crc-storage-crc-58jng\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.553236 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zbwd\" (UniqueName: \"kubernetes.io/projected/a38283e5-ec78-480b-a04b-18f039f316f5-kube-api-access-4zbwd\") pod \"crc-storage-crc-58jng\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:16 crc kubenswrapper[4978]: I0225 08:24:16.573468 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:17 crc kubenswrapper[4978]: I0225 08:24:17.119561 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-58jng"] Feb 25 08:24:17 crc kubenswrapper[4978]: I0225 08:24:17.343125 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="812202c5-cfd7-47c8-be68-e3467786a74e" path="/var/lib/kubelet/pods/812202c5-cfd7-47c8-be68-e3467786a74e/volumes" Feb 25 08:24:18 crc kubenswrapper[4978]: I0225 08:24:18.009934 4978 generic.go:334] "Generic (PLEG): container finished" podID="a38283e5-ec78-480b-a04b-18f039f316f5" containerID="e670130752c3585af207697c8110c6bb009161f85b9289a33b5f939a3e791dc0" exitCode=0 Feb 25 08:24:18 crc kubenswrapper[4978]: I0225 08:24:18.010009 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-58jng" event={"ID":"a38283e5-ec78-480b-a04b-18f039f316f5","Type":"ContainerDied","Data":"e670130752c3585af207697c8110c6bb009161f85b9289a33b5f939a3e791dc0"} Feb 25 08:24:18 crc kubenswrapper[4978]: I0225 08:24:18.010312 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-58jng" event={"ID":"a38283e5-ec78-480b-a04b-18f039f316f5","Type":"ContainerStarted","Data":"e983d40326b79a62efa6105e35ebc8dc6c8b47af9dfb9d416aba22374f51d6dd"} Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.327026 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:24:19 crc kubenswrapper[4978]: E0225 08:24:19.327477 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.422676 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.572793 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a38283e5-ec78-480b-a04b-18f039f316f5-node-mnt\") pod \"a38283e5-ec78-480b-a04b-18f039f316f5\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.572879 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zbwd\" (UniqueName: \"kubernetes.io/projected/a38283e5-ec78-480b-a04b-18f039f316f5-kube-api-access-4zbwd\") pod \"a38283e5-ec78-480b-a04b-18f039f316f5\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.572987 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a38283e5-ec78-480b-a04b-18f039f316f5-crc-storage\") pod \"a38283e5-ec78-480b-a04b-18f039f316f5\" (UID: \"a38283e5-ec78-480b-a04b-18f039f316f5\") " Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.573328 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a38283e5-ec78-480b-a04b-18f039f316f5-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "a38283e5-ec78-480b-a04b-18f039f316f5" (UID: "a38283e5-ec78-480b-a04b-18f039f316f5"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.580724 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a38283e5-ec78-480b-a04b-18f039f316f5-kube-api-access-4zbwd" (OuterVolumeSpecName: "kube-api-access-4zbwd") pod "a38283e5-ec78-480b-a04b-18f039f316f5" (UID: "a38283e5-ec78-480b-a04b-18f039f316f5"). InnerVolumeSpecName "kube-api-access-4zbwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.595706 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a38283e5-ec78-480b-a04b-18f039f316f5-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "a38283e5-ec78-480b-a04b-18f039f316f5" (UID: "a38283e5-ec78-480b-a04b-18f039f316f5"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.675421 4978 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a38283e5-ec78-480b-a04b-18f039f316f5-node-mnt\") on node \"crc\" DevicePath \"\"" Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.675468 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zbwd\" (UniqueName: \"kubernetes.io/projected/a38283e5-ec78-480b-a04b-18f039f316f5-kube-api-access-4zbwd\") on node \"crc\" DevicePath \"\"" Feb 25 08:24:19 crc kubenswrapper[4978]: I0225 08:24:19.675483 4978 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a38283e5-ec78-480b-a04b-18f039f316f5-crc-storage\") on node \"crc\" DevicePath \"\"" Feb 25 08:24:20 crc kubenswrapper[4978]: I0225 08:24:20.030730 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-58jng" event={"ID":"a38283e5-ec78-480b-a04b-18f039f316f5","Type":"ContainerDied","Data":"e983d40326b79a62efa6105e35ebc8dc6c8b47af9dfb9d416aba22374f51d6dd"} Feb 25 08:24:20 crc kubenswrapper[4978]: I0225 08:24:20.030787 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e983d40326b79a62efa6105e35ebc8dc6c8b47af9dfb9d416aba22374f51d6dd" Feb 25 08:24:20 crc kubenswrapper[4978]: I0225 08:24:20.030799 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-58jng" Feb 25 08:24:25 crc kubenswrapper[4978]: I0225 08:24:25.991341 4978 scope.go:117] "RemoveContainer" containerID="9fa9965d85d7a05a3b923998112a98762766c61e7431e2a5765ef9bfd9aa8f79" Feb 25 08:24:26 crc kubenswrapper[4978]: I0225 08:24:26.053150 4978 scope.go:117] "RemoveContainer" containerID="c0ec7df1bcc042b4a48d1cf800f5c8bb0371f7303e7803ef3bfd7cfcedcea1e5" Feb 25 08:24:34 crc kubenswrapper[4978]: I0225 08:24:34.327764 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:24:34 crc kubenswrapper[4978]: E0225 08:24:34.328784 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:24:47 crc kubenswrapper[4978]: I0225 08:24:47.335316 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:24:47 crc kubenswrapper[4978]: E0225 08:24:47.336528 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:25:01 crc kubenswrapper[4978]: I0225 08:25:01.327739 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:25:01 crc kubenswrapper[4978]: E0225 08:25:01.328363 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:25:15 crc kubenswrapper[4978]: I0225 08:25:15.328126 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:25:15 crc kubenswrapper[4978]: E0225 08:25:15.329669 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:25:28 crc kubenswrapper[4978]: I0225 08:25:28.327900 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:25:28 crc kubenswrapper[4978]: I0225 08:25:28.679914 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"ae2d008bb6171404f42a195e3e45bb3e9fb1500e39baedfad501f642abfe36c7"} Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.215951 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533466-mrwn2"] Feb 25 08:26:00 crc kubenswrapper[4978]: E0225 08:26:00.216833 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a38283e5-ec78-480b-a04b-18f039f316f5" containerName="storage" Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.216849 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a38283e5-ec78-480b-a04b-18f039f316f5" containerName="storage" Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.217082 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a38283e5-ec78-480b-a04b-18f039f316f5" containerName="storage" Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.217720 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533466-mrwn2" Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.221981 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.222189 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.222752 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.227189 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533466-mrwn2"] Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.314654 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh9q7\" (UniqueName: \"kubernetes.io/projected/7df3eae9-8600-47fc-906e-3a0adf79ad2d-kube-api-access-fh9q7\") pod \"auto-csr-approver-29533466-mrwn2\" (UID: \"7df3eae9-8600-47fc-906e-3a0adf79ad2d\") " pod="openshift-infra/auto-csr-approver-29533466-mrwn2" Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.415459 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh9q7\" (UniqueName: \"kubernetes.io/projected/7df3eae9-8600-47fc-906e-3a0adf79ad2d-kube-api-access-fh9q7\") pod \"auto-csr-approver-29533466-mrwn2\" (UID: \"7df3eae9-8600-47fc-906e-3a0adf79ad2d\") " pod="openshift-infra/auto-csr-approver-29533466-mrwn2" Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.449133 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh9q7\" (UniqueName: \"kubernetes.io/projected/7df3eae9-8600-47fc-906e-3a0adf79ad2d-kube-api-access-fh9q7\") pod \"auto-csr-approver-29533466-mrwn2\" (UID: \"7df3eae9-8600-47fc-906e-3a0adf79ad2d\") " pod="openshift-infra/auto-csr-approver-29533466-mrwn2" Feb 25 08:26:00 crc kubenswrapper[4978]: I0225 08:26:00.552322 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533466-mrwn2" Feb 25 08:26:01 crc kubenswrapper[4978]: I0225 08:26:01.041324 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533466-mrwn2"] Feb 25 08:26:02 crc kubenswrapper[4978]: I0225 08:26:02.025070 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533466-mrwn2" event={"ID":"7df3eae9-8600-47fc-906e-3a0adf79ad2d","Type":"ContainerStarted","Data":"408c274262d3ff72f945c384c7627eed9cb8ba770c7b3c692b696be75e10f946"} Feb 25 08:26:03 crc kubenswrapper[4978]: I0225 08:26:03.037311 4978 generic.go:334] "Generic (PLEG): container finished" podID="7df3eae9-8600-47fc-906e-3a0adf79ad2d" containerID="ffb5b65fe7df9dd275efa4a8e89dfe64e8a904a82ce5e1ac3ecef7ecf13320da" exitCode=0 Feb 25 08:26:03 crc kubenswrapper[4978]: I0225 08:26:03.037463 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533466-mrwn2" event={"ID":"7df3eae9-8600-47fc-906e-3a0adf79ad2d","Type":"ContainerDied","Data":"ffb5b65fe7df9dd275efa4a8e89dfe64e8a904a82ce5e1ac3ecef7ecf13320da"} Feb 25 08:26:04 crc kubenswrapper[4978]: I0225 08:26:04.458906 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533466-mrwn2" Feb 25 08:26:04 crc kubenswrapper[4978]: I0225 08:26:04.482039 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh9q7\" (UniqueName: \"kubernetes.io/projected/7df3eae9-8600-47fc-906e-3a0adf79ad2d-kube-api-access-fh9q7\") pod \"7df3eae9-8600-47fc-906e-3a0adf79ad2d\" (UID: \"7df3eae9-8600-47fc-906e-3a0adf79ad2d\") " Feb 25 08:26:04 crc kubenswrapper[4978]: I0225 08:26:04.487934 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df3eae9-8600-47fc-906e-3a0adf79ad2d-kube-api-access-fh9q7" (OuterVolumeSpecName: "kube-api-access-fh9q7") pod "7df3eae9-8600-47fc-906e-3a0adf79ad2d" (UID: "7df3eae9-8600-47fc-906e-3a0adf79ad2d"). InnerVolumeSpecName "kube-api-access-fh9q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:26:04 crc kubenswrapper[4978]: I0225 08:26:04.583622 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh9q7\" (UniqueName: \"kubernetes.io/projected/7df3eae9-8600-47fc-906e-3a0adf79ad2d-kube-api-access-fh9q7\") on node \"crc\" DevicePath \"\"" Feb 25 08:26:05 crc kubenswrapper[4978]: I0225 08:26:05.059274 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533466-mrwn2" event={"ID":"7df3eae9-8600-47fc-906e-3a0adf79ad2d","Type":"ContainerDied","Data":"408c274262d3ff72f945c384c7627eed9cb8ba770c7b3c692b696be75e10f946"} Feb 25 08:26:05 crc kubenswrapper[4978]: I0225 08:26:05.059667 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="408c274262d3ff72f945c384c7627eed9cb8ba770c7b3c692b696be75e10f946" Feb 25 08:26:05 crc kubenswrapper[4978]: I0225 08:26:05.059465 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533466-mrwn2" Feb 25 08:26:05 crc kubenswrapper[4978]: E0225 08:26:05.313656 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7df3eae9_8600_47fc_906e_3a0adf79ad2d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7df3eae9_8600_47fc_906e_3a0adf79ad2d.slice/crio-408c274262d3ff72f945c384c7627eed9cb8ba770c7b3c692b696be75e10f946\": RecentStats: unable to find data in memory cache]" Feb 25 08:26:05 crc kubenswrapper[4978]: I0225 08:26:05.545783 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533460-hdnlv"] Feb 25 08:26:05 crc kubenswrapper[4978]: I0225 08:26:05.553088 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533460-hdnlv"] Feb 25 08:26:07 crc kubenswrapper[4978]: I0225 08:26:07.342233 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dd60b00-0f0b-422a-971b-76336b1a813e" path="/var/lib/kubelet/pods/6dd60b00-0f0b-422a-971b-76336b1a813e/volumes" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.123633 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-685bff9b5f-rrwbb"] Feb 25 08:26:26 crc kubenswrapper[4978]: E0225 08:26:26.124523 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df3eae9-8600-47fc-906e-3a0adf79ad2d" containerName="oc" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.124541 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df3eae9-8600-47fc-906e-3a0adf79ad2d" containerName="oc" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.124718 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7df3eae9-8600-47fc-906e-3a0adf79ad2d" containerName="oc" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.125731 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.127800 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.128226 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-phc6l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.128297 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.128612 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.133251 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-685bff9b5f-rrwbb"] Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.145419 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6954dff675-qbt9l"] Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.146949 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-config\") pod \"dnsmasq-dns-685bff9b5f-rrwbb\" (UID: \"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2\") " pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.147091 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n52f\" (UniqueName: \"kubernetes.io/projected/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-kube-api-access-5n52f\") pod \"dnsmasq-dns-685bff9b5f-rrwbb\" (UID: \"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2\") " pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.147123 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.149410 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.160974 4978 scope.go:117] "RemoveContainer" containerID="2e06c9d8615ca9e9cd9423b8ed2ac3cf958eed8793d36af984ed86cde1ea82e5" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.201021 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6954dff675-qbt9l"] Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.251104 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-dns-svc\") pod \"dnsmasq-dns-6954dff675-qbt9l\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.253186 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-config\") pod \"dnsmasq-dns-685bff9b5f-rrwbb\" (UID: \"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2\") " pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.253311 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lcvb\" (UniqueName: \"kubernetes.io/projected/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-kube-api-access-9lcvb\") pod \"dnsmasq-dns-6954dff675-qbt9l\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.253407 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-config\") pod \"dnsmasq-dns-6954dff675-qbt9l\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.253492 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n52f\" (UniqueName: \"kubernetes.io/projected/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-kube-api-access-5n52f\") pod \"dnsmasq-dns-685bff9b5f-rrwbb\" (UID: \"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2\") " pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.254793 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-config\") pod \"dnsmasq-dns-685bff9b5f-rrwbb\" (UID: \"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2\") " pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.291978 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n52f\" (UniqueName: \"kubernetes.io/projected/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-kube-api-access-5n52f\") pod \"dnsmasq-dns-685bff9b5f-rrwbb\" (UID: \"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2\") " pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.354637 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lcvb\" (UniqueName: \"kubernetes.io/projected/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-kube-api-access-9lcvb\") pod \"dnsmasq-dns-6954dff675-qbt9l\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.354709 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-config\") pod \"dnsmasq-dns-6954dff675-qbt9l\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.354756 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-dns-svc\") pod \"dnsmasq-dns-6954dff675-qbt9l\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.355568 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-config\") pod \"dnsmasq-dns-6954dff675-qbt9l\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.355837 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-dns-svc\") pod \"dnsmasq-dns-6954dff675-qbt9l\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.361652 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-685bff9b5f-rrwbb"] Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.362079 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.393034 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lcvb\" (UniqueName: \"kubernetes.io/projected/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-kube-api-access-9lcvb\") pod \"dnsmasq-dns-6954dff675-qbt9l\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.397536 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-658cb7bfbc-bmxj9"] Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.398908 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.409542 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658cb7bfbc-bmxj9"] Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.456208 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-config\") pod \"dnsmasq-dns-658cb7bfbc-bmxj9\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.456269 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-dns-svc\") pod \"dnsmasq-dns-658cb7bfbc-bmxj9\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.456287 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hhkp\" (UniqueName: \"kubernetes.io/projected/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-kube-api-access-4hhkp\") pod \"dnsmasq-dns-658cb7bfbc-bmxj9\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.468021 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.557320 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-config\") pod \"dnsmasq-dns-658cb7bfbc-bmxj9\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.557626 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-dns-svc\") pod \"dnsmasq-dns-658cb7bfbc-bmxj9\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.557645 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hhkp\" (UniqueName: \"kubernetes.io/projected/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-kube-api-access-4hhkp\") pod \"dnsmasq-dns-658cb7bfbc-bmxj9\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.558188 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-config\") pod \"dnsmasq-dns-658cb7bfbc-bmxj9\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.558437 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-dns-svc\") pod \"dnsmasq-dns-658cb7bfbc-bmxj9\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.587735 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hhkp\" (UniqueName: \"kubernetes.io/projected/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-kube-api-access-4hhkp\") pod \"dnsmasq-dns-658cb7bfbc-bmxj9\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.672709 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658cb7bfbc-bmxj9"] Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.673201 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.699229 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54756c6d6f-gwvms"] Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.700349 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.711663 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54756c6d6f-gwvms"] Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.723867 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-685bff9b5f-rrwbb"] Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.759858 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-config\") pod \"dnsmasq-dns-54756c6d6f-gwvms\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.760155 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-dns-svc\") pod \"dnsmasq-dns-54756c6d6f-gwvms\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.760193 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nh24\" (UniqueName: \"kubernetes.io/projected/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-kube-api-access-5nh24\") pod \"dnsmasq-dns-54756c6d6f-gwvms\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.861047 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nh24\" (UniqueName: \"kubernetes.io/projected/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-kube-api-access-5nh24\") pod \"dnsmasq-dns-54756c6d6f-gwvms\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.861136 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-config\") pod \"dnsmasq-dns-54756c6d6f-gwvms\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.861195 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-dns-svc\") pod \"dnsmasq-dns-54756c6d6f-gwvms\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.862026 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-dns-svc\") pod \"dnsmasq-dns-54756c6d6f-gwvms\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.862065 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-config\") pod \"dnsmasq-dns-54756c6d6f-gwvms\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:26 crc kubenswrapper[4978]: I0225 08:26:26.881847 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nh24\" (UniqueName: \"kubernetes.io/projected/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-kube-api-access-5nh24\") pod \"dnsmasq-dns-54756c6d6f-gwvms\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.024240 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.026735 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6954dff675-qbt9l"] Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.160028 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658cb7bfbc-bmxj9"] Feb 25 08:26:27 crc kubenswrapper[4978]: W0225 08:26:27.196788 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec82c93f_3cd9_48a9_801a_1ebb3ffee6a5.slice/crio-57f686f77cd8c6b4435db33167a9bb82b41f72037938ca2e5ac15a9c8116fe6d WatchSource:0}: Error finding container 57f686f77cd8c6b4435db33167a9bb82b41f72037938ca2e5ac15a9c8116fe6d: Status 404 returned error can't find the container with id 57f686f77cd8c6b4435db33167a9bb82b41f72037938ca2e5ac15a9c8116fe6d Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.409992 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" event={"ID":"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2","Type":"ContainerStarted","Data":"1133301b539491cd803e98c1d85469f74218b16be7dce760475722b595e0e0bd"} Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.416244 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" event={"ID":"11b9ad4c-6d6e-4bd8-8d59-c77545feb841","Type":"ContainerStarted","Data":"b02f8fcf2eaa3cc9ebe642912b32ac78d6bbc613fa9b2e43bb05745921197f6d"} Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.423669 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" event={"ID":"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5","Type":"ContainerStarted","Data":"57f686f77cd8c6b4435db33167a9bb82b41f72037938ca2e5ac15a9c8116fe6d"} Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.526394 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54756c6d6f-gwvms"] Feb 25 08:26:27 crc kubenswrapper[4978]: W0225 08:26:27.536462 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod903e7f0c_1ec6_45a2_a69c_f3caed9750a3.slice/crio-7a30d587b7eb9eafd4fce6af6c433270da962ec6a54390dfe057c85753ba7c91 WatchSource:0}: Error finding container 7a30d587b7eb9eafd4fce6af6c433270da962ec6a54390dfe057c85753ba7c91: Status 404 returned error can't find the container with id 7a30d587b7eb9eafd4fce6af6c433270da962ec6a54390dfe057c85753ba7c91 Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.806786 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.807946 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.809926 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.810813 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.811192 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-7ch67" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.811428 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.811577 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.812062 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.812227 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.820395 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975591 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975651 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975680 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpqfd\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-kube-api-access-tpqfd\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975736 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975802 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/de4db333-9150-4ccd-8f4e-d2a18917bbc5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975826 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975848 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975886 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/de4db333-9150-4ccd-8f4e-d2a18917bbc5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975920 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975956 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:27 crc kubenswrapper[4978]: I0225 08:26:27.975986 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.076929 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpqfd\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-kube-api-access-tpqfd\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077000 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077042 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/de4db333-9150-4ccd-8f4e-d2a18917bbc5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077086 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077132 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077162 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/de4db333-9150-4ccd-8f4e-d2a18917bbc5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077204 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077258 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077281 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077305 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077338 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.077989 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.078070 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.080717 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.080848 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.081563 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.082057 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.082087 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/50ca25561a5c1786c943b437a4014dbd873252e15a6addc89cbb6d6bb81d4347/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.085863 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/de4db333-9150-4ccd-8f4e-d2a18917bbc5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.092806 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/de4db333-9150-4ccd-8f4e-d2a18917bbc5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.094943 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.096993 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpqfd\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-kube-api-access-tpqfd\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.097039 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.109629 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") pod \"rabbitmq-cell1-server-0\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.146911 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.432164 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" event={"ID":"903e7f0c-1ec6-45a2-a69c-f3caed9750a3","Type":"ContainerStarted","Data":"7a30d587b7eb9eafd4fce6af6c433270da962ec6a54390dfe057c85753ba7c91"} Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.571254 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 08:26:28 crc kubenswrapper[4978]: W0225 08:26:28.581208 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde4db333_9150_4ccd_8f4e_d2a18917bbc5.slice/crio-afd671d54749b27b54e3d7002bc6190b2b37e36097b296d786045ee9b80bb6b0 WatchSource:0}: Error finding container afd671d54749b27b54e3d7002bc6190b2b37e36097b296d786045ee9b80bb6b0: Status 404 returned error can't find the container with id afd671d54749b27b54e3d7002bc6190b2b37e36097b296d786045ee9b80bb6b0 Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.885520 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.886675 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.892297 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.897584 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-rxfh7" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.898514 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.898709 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.903831 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.903927 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.995711 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11a899ad-0164-4f8b-ae66-797557d34867-operator-scripts\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.995769 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a899ad-0164-4f8b-ae66-797557d34867-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.995808 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cea3032f-4e90-4507-8d22-c272512ba672\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cea3032f-4e90-4507-8d22-c272512ba672\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.995854 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/11a899ad-0164-4f8b-ae66-797557d34867-config-data-default\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.995880 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/11a899ad-0164-4f8b-ae66-797557d34867-kolla-config\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.995896 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/11a899ad-0164-4f8b-ae66-797557d34867-config-data-generated\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.995913 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/11a899ad-0164-4f8b-ae66-797557d34867-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:28 crc kubenswrapper[4978]: I0225 08:26:28.995939 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44ljn\" (UniqueName: \"kubernetes.io/projected/11a899ad-0164-4f8b-ae66-797557d34867-kube-api-access-44ljn\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.042528 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.043586 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.056547 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-wsprr" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.056551 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.056688 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.056834 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.057039 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.057609 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.057716 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.058529 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.097576 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cea3032f-4e90-4507-8d22-c272512ba672\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cea3032f-4e90-4507-8d22-c272512ba672\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.097661 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/11a899ad-0164-4f8b-ae66-797557d34867-config-data-default\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.097694 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/11a899ad-0164-4f8b-ae66-797557d34867-kolla-config\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.097713 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/11a899ad-0164-4f8b-ae66-797557d34867-config-data-generated\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.097733 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/11a899ad-0164-4f8b-ae66-797557d34867-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.098072 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44ljn\" (UniqueName: \"kubernetes.io/projected/11a899ad-0164-4f8b-ae66-797557d34867-kube-api-access-44ljn\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.098119 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11a899ad-0164-4f8b-ae66-797557d34867-operator-scripts\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.098412 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a899ad-0164-4f8b-ae66-797557d34867-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.098340 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/11a899ad-0164-4f8b-ae66-797557d34867-config-data-generated\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.100260 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.100282 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cea3032f-4e90-4507-8d22-c272512ba672\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cea3032f-4e90-4507-8d22-c272512ba672\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/40e04f71a625596904b7ea6bfc143e28d93b86836d8d7b1ee7a8bba688090ba0/globalmount\"" pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.100501 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/11a899ad-0164-4f8b-ae66-797557d34867-config-data-default\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.100573 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/11a899ad-0164-4f8b-ae66-797557d34867-kolla-config\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.102107 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11a899ad-0164-4f8b-ae66-797557d34867-operator-scripts\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.104081 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/11a899ad-0164-4f8b-ae66-797557d34867-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.108743 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a899ad-0164-4f8b-ae66-797557d34867-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.121488 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44ljn\" (UniqueName: \"kubernetes.io/projected/11a899ad-0164-4f8b-ae66-797557d34867-kube-api-access-44ljn\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.162642 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cea3032f-4e90-4507-8d22-c272512ba672\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cea3032f-4e90-4507-8d22-c272512ba672\") pod \"openstack-galera-0\" (UID: \"11a899ad-0164-4f8b-ae66-797557d34867\") " pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200105 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200159 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-389bf769-1bfe-41e5-84a3-249075245d23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200202 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200245 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200267 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200286 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jpmq\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-kube-api-access-8jpmq\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200308 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4186a72-0d11-4647-902b-3908454bf0b1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200340 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200379 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200501 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.200655 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4186a72-0d11-4647-902b-3908454bf0b1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.216504 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302629 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302661 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302689 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4186a72-0d11-4647-902b-3908454bf0b1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302706 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302731 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-389bf769-1bfe-41e5-84a3-249075245d23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302761 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302798 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302820 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302838 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jpmq\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-kube-api-access-8jpmq\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302856 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4186a72-0d11-4647-902b-3908454bf0b1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.302878 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.304349 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-config-data\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.304609 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.306346 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.308452 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.308544 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4186a72-0d11-4647-902b-3908454bf0b1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.309014 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.310416 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.324180 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.324223 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-389bf769-1bfe-41e5-84a3-249075245d23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/97ca09a128e247017af88d0c60eefc68e785640cb893abdbcf3e382fbf328a1c/globalmount\"" pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.324329 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.328656 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jpmq\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-kube-api-access-8jpmq\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.330287 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4186a72-0d11-4647-902b-3908454bf0b1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.373686 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-389bf769-1bfe-41e5-84a3-249075245d23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") pod \"rabbitmq-server-0\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.457947 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"de4db333-9150-4ccd-8f4e-d2a18917bbc5","Type":"ContainerStarted","Data":"afd671d54749b27b54e3d7002bc6190b2b37e36097b296d786045ee9b80bb6b0"} Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.679733 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 08:26:29 crc kubenswrapper[4978]: I0225 08:26:29.705781 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 25 08:26:29 crc kubenswrapper[4978]: W0225 08:26:29.717528 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11a899ad_0164_4f8b_ae66_797557d34867.slice/crio-133a977fdbd4faec9a4475faa7d1945246b10a9adeeeaf87ea5bf4f4fcdf9665 WatchSource:0}: Error finding container 133a977fdbd4faec9a4475faa7d1945246b10a9adeeeaf87ea5bf4f4fcdf9665: Status 404 returned error can't find the container with id 133a977fdbd4faec9a4475faa7d1945246b10a9adeeeaf87ea5bf4f4fcdf9665 Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.116617 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 08:26:30 crc kubenswrapper[4978]: W0225 08:26:30.125666 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4186a72_0d11_4647_902b_3908454bf0b1.slice/crio-dad6a6fbcb1bcd42bf22231fca9315d95bac52a3bdfad72ca01ef3f01123055e WatchSource:0}: Error finding container dad6a6fbcb1bcd42bf22231fca9315d95bac52a3bdfad72ca01ef3f01123055e: Status 404 returned error can't find the container with id dad6a6fbcb1bcd42bf22231fca9315d95bac52a3bdfad72ca01ef3f01123055e Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.246419 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.247937 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.251284 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.251500 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.251656 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-kx9hq" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.252792 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.274842 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.321579 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7be4547-b312-4601-93bf-9a3c62a13829-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.321626 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnj9l\" (UniqueName: \"kubernetes.io/projected/c7be4547-b312-4601-93bf-9a3c62a13829-kube-api-access-wnj9l\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.321657 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c7be4547-b312-4601-93bf-9a3c62a13829-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.321684 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-12bae305-db97-4e36-ae4e-f1b21b8e4cf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bae305-db97-4e36-ae4e-f1b21b8e4cf2\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.321706 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c7be4547-b312-4601-93bf-9a3c62a13829-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.321736 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7be4547-b312-4601-93bf-9a3c62a13829-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.321768 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7be4547-b312-4601-93bf-9a3c62a13829-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.321792 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c7be4547-b312-4601-93bf-9a3c62a13829-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.414991 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.416123 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.418794 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-hjf48" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.419100 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.419353 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.423349 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c7be4547-b312-4601-93bf-9a3c62a13829-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.423449 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7be4547-b312-4601-93bf-9a3c62a13829-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.423505 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnj9l\" (UniqueName: \"kubernetes.io/projected/c7be4547-b312-4601-93bf-9a3c62a13829-kube-api-access-wnj9l\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.423537 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c7be4547-b312-4601-93bf-9a3c62a13829-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.423567 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-12bae305-db97-4e36-ae4e-f1b21b8e4cf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bae305-db97-4e36-ae4e-f1b21b8e4cf2\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.423586 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c7be4547-b312-4601-93bf-9a3c62a13829-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.423615 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7be4547-b312-4601-93bf-9a3c62a13829-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.423677 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7be4547-b312-4601-93bf-9a3c62a13829-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.424754 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c7be4547-b312-4601-93bf-9a3c62a13829-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.424779 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c7be4547-b312-4601-93bf-9a3c62a13829-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.425317 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7be4547-b312-4601-93bf-9a3c62a13829-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.427391 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c7be4547-b312-4601-93bf-9a3c62a13829-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.432510 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.432562 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-12bae305-db97-4e36-ae4e-f1b21b8e4cf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bae305-db97-4e36-ae4e-f1b21b8e4cf2\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4b4eb021748965cf8aa86ac763c72d3db5da713df8cbeca269f3a182e2e7d4a8/globalmount\"" pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.434363 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7be4547-b312-4601-93bf-9a3c62a13829-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.437097 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7be4547-b312-4601-93bf-9a3c62a13829-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.437184 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.443974 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnj9l\" (UniqueName: \"kubernetes.io/projected/c7be4547-b312-4601-93bf-9a3c62a13829-kube-api-access-wnj9l\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.468403 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"11a899ad-0164-4f8b-ae66-797557d34867","Type":"ContainerStarted","Data":"133a977fdbd4faec9a4475faa7d1945246b10a9adeeeaf87ea5bf4f4fcdf9665"} Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.469734 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4186a72-0d11-4647-902b-3908454bf0b1","Type":"ContainerStarted","Data":"dad6a6fbcb1bcd42bf22231fca9315d95bac52a3bdfad72ca01ef3f01123055e"} Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.470072 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-12bae305-db97-4e36-ae4e-f1b21b8e4cf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-12bae305-db97-4e36-ae4e-f1b21b8e4cf2\") pod \"openstack-cell1-galera-0\" (UID: \"c7be4547-b312-4601-93bf-9a3c62a13829\") " pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.526276 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2fd170eb-5a4e-4461-9619-58f110e6e8f1-config-data\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.526354 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2fd170eb-5a4e-4461-9619-58f110e6e8f1-kolla-config\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.526398 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd170eb-5a4e-4461-9619-58f110e6e8f1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.526426 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcfpp\" (UniqueName: \"kubernetes.io/projected/2fd170eb-5a4e-4461-9619-58f110e6e8f1-kube-api-access-bcfpp\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.526464 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd170eb-5a4e-4461-9619-58f110e6e8f1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.583907 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.628073 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2fd170eb-5a4e-4461-9619-58f110e6e8f1-config-data\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.628150 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2fd170eb-5a4e-4461-9619-58f110e6e8f1-kolla-config\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.628186 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd170eb-5a4e-4461-9619-58f110e6e8f1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.628211 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcfpp\" (UniqueName: \"kubernetes.io/projected/2fd170eb-5a4e-4461-9619-58f110e6e8f1-kube-api-access-bcfpp\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.628255 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd170eb-5a4e-4461-9619-58f110e6e8f1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.629169 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2fd170eb-5a4e-4461-9619-58f110e6e8f1-kolla-config\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.629172 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2fd170eb-5a4e-4461-9619-58f110e6e8f1-config-data\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.632459 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fd170eb-5a4e-4461-9619-58f110e6e8f1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.633173 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fd170eb-5a4e-4461-9619-58f110e6e8f1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.649945 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcfpp\" (UniqueName: \"kubernetes.io/projected/2fd170eb-5a4e-4461-9619-58f110e6e8f1-kube-api-access-bcfpp\") pod \"memcached-0\" (UID: \"2fd170eb-5a4e-4461-9619-58f110e6e8f1\") " pod="openstack/memcached-0" Feb 25 08:26:30 crc kubenswrapper[4978]: I0225 08:26:30.803262 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 25 08:26:31 crc kubenswrapper[4978]: I0225 08:26:31.072869 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 25 08:26:31 crc kubenswrapper[4978]: W0225 08:26:31.084951 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7be4547_b312_4601_93bf_9a3c62a13829.slice/crio-f4c8915ba480a9d6736cc998d456cad9ad41283893e2c0e906366d692df64a2a WatchSource:0}: Error finding container f4c8915ba480a9d6736cc998d456cad9ad41283893e2c0e906366d692df64a2a: Status 404 returned error can't find the container with id f4c8915ba480a9d6736cc998d456cad9ad41283893e2c0e906366d692df64a2a Feb 25 08:26:31 crc kubenswrapper[4978]: I0225 08:26:31.222941 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 25 08:26:31 crc kubenswrapper[4978]: W0225 08:26:31.249588 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fd170eb_5a4e_4461_9619_58f110e6e8f1.slice/crio-9236e5eccaa48d96ce046d9df230b7907d8a7eba6136eb7453aa816868667c43 WatchSource:0}: Error finding container 9236e5eccaa48d96ce046d9df230b7907d8a7eba6136eb7453aa816868667c43: Status 404 returned error can't find the container with id 9236e5eccaa48d96ce046d9df230b7907d8a7eba6136eb7453aa816868667c43 Feb 25 08:26:31 crc kubenswrapper[4978]: I0225 08:26:31.477896 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2fd170eb-5a4e-4461-9619-58f110e6e8f1","Type":"ContainerStarted","Data":"9236e5eccaa48d96ce046d9df230b7907d8a7eba6136eb7453aa816868667c43"} Feb 25 08:26:31 crc kubenswrapper[4978]: I0225 08:26:31.479406 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c7be4547-b312-4601-93bf-9a3c62a13829","Type":"ContainerStarted","Data":"f4c8915ba480a9d6736cc998d456cad9ad41283893e2c0e906366d692df64a2a"} Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.661924 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"11a899ad-0164-4f8b-ae66-797557d34867","Type":"ContainerStarted","Data":"34fc60e992e78baa5a3cf4650a1088b2da5fb07125819dcd4d7b534e255a953f"} Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.666752 4978 generic.go:334] "Generic (PLEG): container finished" podID="3bba8e0a-fc78-46b0-8e3d-915c910f1ca2" containerID="efcb3b7aa7c8936aca1572869c26f4f0492ed044d2ddd3dc01c48952fb2d3226" exitCode=0 Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.666834 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" event={"ID":"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2","Type":"ContainerDied","Data":"efcb3b7aa7c8936aca1572869c26f4f0492ed044d2ddd3dc01c48952fb2d3226"} Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.669104 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2fd170eb-5a4e-4461-9619-58f110e6e8f1","Type":"ContainerStarted","Data":"380c18c12fda2cc9c7d2f03aedeaeed8c64f51394684e464c39b6ad54403f8d0"} Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.669276 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.675603 4978 generic.go:334] "Generic (PLEG): container finished" podID="903e7f0c-1ec6-45a2-a69c-f3caed9750a3" containerID="99e211e9a0378ea047273aa9024bc2e06fd87b2507530dfaa57750faf831891a" exitCode=0 Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.675647 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" event={"ID":"903e7f0c-1ec6-45a2-a69c-f3caed9750a3","Type":"ContainerDied","Data":"99e211e9a0378ea047273aa9024bc2e06fd87b2507530dfaa57750faf831891a"} Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.678588 4978 generic.go:334] "Generic (PLEG): container finished" podID="11b9ad4c-6d6e-4bd8-8d59-c77545feb841" containerID="487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4" exitCode=0 Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.678633 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" event={"ID":"11b9ad4c-6d6e-4bd8-8d59-c77545feb841","Type":"ContainerDied","Data":"487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4"} Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.713047 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c7be4547-b312-4601-93bf-9a3c62a13829","Type":"ContainerStarted","Data":"2a2fd043e6faaaf313c9f49049c0d0747ed33fc9449d39c1fadb620936d6531e"} Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.719478 4978 generic.go:334] "Generic (PLEG): container finished" podID="ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5" containerID="c3ffc95562390e44d02122b9069c0836a3f40ef86b65ef52a0a44456fbfc79dd" exitCode=0 Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.719520 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" event={"ID":"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5","Type":"ContainerDied","Data":"c3ffc95562390e44d02122b9069c0836a3f40ef86b65ef52a0a44456fbfc79dd"} Feb 25 08:26:52 crc kubenswrapper[4978]: I0225 08:26:52.741536 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.895443984 podStartE2EDuration="22.741513714s" podCreationTimestamp="2026-02-25 08:26:30 +0000 UTC" firstStartedPulling="2026-02-25 08:26:31.255353978 +0000 UTC m=+6084.694610437" lastFinishedPulling="2026-02-25 08:26:52.101423658 +0000 UTC m=+6105.540680167" observedRunningTime="2026-02-25 08:26:52.724779796 +0000 UTC m=+6106.164036265" watchObservedRunningTime="2026-02-25 08:26:52.741513714 +0000 UTC m=+6106.180770183" Feb 25 08:26:52 crc kubenswrapper[4978]: E0225 08:26:52.897703 4978 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Feb 25 08:26:52 crc kubenswrapper[4978]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/11b9ad4c-6d6e-4bd8-8d59-c77545feb841/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 25 08:26:52 crc kubenswrapper[4978]: > podSandboxID="b02f8fcf2eaa3cc9ebe642912b32ac78d6bbc613fa9b2e43bb05745921197f6d" Feb 25 08:26:52 crc kubenswrapper[4978]: E0225 08:26:52.898075 4978 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 25 08:26:52 crc kubenswrapper[4978]: container &Container{Name:dnsmasq-dns,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:4198fe0138d2c46ae51eb6e53fed5006,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n697h54dhb7h666h69h76h59ch55ch65ch596h8h79h5c8h57hc8hfch5d7h697h79h698h5fch644hf9h54chbfh655hfchcbh5f8h646h5f7h89q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9lcvb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6954dff675-qbt9l_openstack(11b9ad4c-6d6e-4bd8-8d59-c77545feb841): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/11b9ad4c-6d6e-4bd8-8d59-c77545feb841/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 25 08:26:52 crc kubenswrapper[4978]: > logger="UnhandledError" Feb 25 08:26:52 crc kubenswrapper[4978]: E0225 08:26:52.900246 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/11b9ad4c-6d6e-4bd8-8d59-c77545feb841/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" podUID="11b9ad4c-6d6e-4bd8-8d59-c77545feb841" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.025152 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.051156 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.095268 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n52f\" (UniqueName: \"kubernetes.io/projected/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-kube-api-access-5n52f\") pod \"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2\" (UID: \"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2\") " Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.095434 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-config\") pod \"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2\" (UID: \"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2\") " Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.186300 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-kube-api-access-5n52f" (OuterVolumeSpecName: "kube-api-access-5n52f") pod "3bba8e0a-fc78-46b0-8e3d-915c910f1ca2" (UID: "3bba8e0a-fc78-46b0-8e3d-915c910f1ca2"). InnerVolumeSpecName "kube-api-access-5n52f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.196390 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-dns-svc\") pod \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.196824 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-config\") pod \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.196901 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hhkp\" (UniqueName: \"kubernetes.io/projected/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-kube-api-access-4hhkp\") pod \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\" (UID: \"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5\") " Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.197247 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n52f\" (UniqueName: \"kubernetes.io/projected/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-kube-api-access-5n52f\") on node \"crc\" DevicePath \"\"" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.221626 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5" (UID: "ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.285857 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-kube-api-access-4hhkp" (OuterVolumeSpecName: "kube-api-access-4hhkp") pod "ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5" (UID: "ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5"). InnerVolumeSpecName "kube-api-access-4hhkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.285981 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-config" (OuterVolumeSpecName: "config") pod "3bba8e0a-fc78-46b0-8e3d-915c910f1ca2" (UID: "3bba8e0a-fc78-46b0-8e3d-915c910f1ca2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.298806 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.298852 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hhkp\" (UniqueName: \"kubernetes.io/projected/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-kube-api-access-4hhkp\") on node \"crc\" DevicePath \"\"" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.298873 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.312724 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-config" (OuterVolumeSpecName: "config") pod "ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5" (UID: "ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.400528 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.731823 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" event={"ID":"ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5","Type":"ContainerDied","Data":"57f686f77cd8c6b4435db33167a9bb82b41f72037938ca2e5ac15a9c8116fe6d"} Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.731891 4978 scope.go:117] "RemoveContainer" containerID="c3ffc95562390e44d02122b9069c0836a3f40ef86b65ef52a0a44456fbfc79dd" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.732071 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658cb7bfbc-bmxj9" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.734583 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" event={"ID":"3bba8e0a-fc78-46b0-8e3d-915c910f1ca2","Type":"ContainerDied","Data":"1133301b539491cd803e98c1d85469f74218b16be7dce760475722b595e0e0bd"} Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.734744 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-685bff9b5f-rrwbb" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.738385 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4186a72-0d11-4647-902b-3908454bf0b1","Type":"ContainerStarted","Data":"851781b6535eb50c799c75b76137c93e166f97d3eb4f6b3617eddc52c7e6db47"} Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.740574 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" event={"ID":"903e7f0c-1ec6-45a2-a69c-f3caed9750a3","Type":"ContainerStarted","Data":"497158ec40a3a54beb4b4eae741dd2b5b21192570a6645c182ffc39cc18a8602"} Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.741068 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.742794 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"de4db333-9150-4ccd-8f4e-d2a18917bbc5","Type":"ContainerStarted","Data":"8eb08550ac619cc24779584a8a60dad408e75a3636fff54ad14f71daa4268556"} Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.752771 4978 scope.go:117] "RemoveContainer" containerID="efcb3b7aa7c8936aca1572869c26f4f0492ed044d2ddd3dc01c48952fb2d3226" Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.792470 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658cb7bfbc-bmxj9"] Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.805334 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-658cb7bfbc-bmxj9"] Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.840682 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-685bff9b5f-rrwbb"] Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.846941 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-685bff9b5f-rrwbb"] Feb 25 08:26:53 crc kubenswrapper[4978]: I0225 08:26:53.849809 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" podStartSLOduration=3.202076721 podStartE2EDuration="27.849786157s" podCreationTimestamp="2026-02-25 08:26:26 +0000 UTC" firstStartedPulling="2026-02-25 08:26:27.538211525 +0000 UTC m=+6080.977467974" lastFinishedPulling="2026-02-25 08:26:52.185920951 +0000 UTC m=+6105.625177410" observedRunningTime="2026-02-25 08:26:53.846064981 +0000 UTC m=+6107.285321450" watchObservedRunningTime="2026-02-25 08:26:53.849786157 +0000 UTC m=+6107.289042616" Feb 25 08:26:54 crc kubenswrapper[4978]: I0225 08:26:54.760038 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" event={"ID":"11b9ad4c-6d6e-4bd8-8d59-c77545feb841","Type":"ContainerStarted","Data":"18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f"} Feb 25 08:26:54 crc kubenswrapper[4978]: I0225 08:26:54.790062 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" podStartSLOduration=3.711482103 podStartE2EDuration="28.790031083s" podCreationTimestamp="2026-02-25 08:26:26 +0000 UTC" firstStartedPulling="2026-02-25 08:26:27.061758772 +0000 UTC m=+6080.501015231" lastFinishedPulling="2026-02-25 08:26:52.140307752 +0000 UTC m=+6105.579564211" observedRunningTime="2026-02-25 08:26:54.785572276 +0000 UTC m=+6108.224828785" watchObservedRunningTime="2026-02-25 08:26:54.790031083 +0000 UTC m=+6108.229287582" Feb 25 08:26:55 crc kubenswrapper[4978]: I0225 08:26:55.344979 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bba8e0a-fc78-46b0-8e3d-915c910f1ca2" path="/var/lib/kubelet/pods/3bba8e0a-fc78-46b0-8e3d-915c910f1ca2/volumes" Feb 25 08:26:55 crc kubenswrapper[4978]: I0225 08:26:55.345981 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5" path="/var/lib/kubelet/pods/ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5/volumes" Feb 25 08:26:56 crc kubenswrapper[4978]: I0225 08:26:56.468521 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:56 crc kubenswrapper[4978]: I0225 08:26:56.782703 4978 generic.go:334] "Generic (PLEG): container finished" podID="c7be4547-b312-4601-93bf-9a3c62a13829" containerID="2a2fd043e6faaaf313c9f49049c0d0747ed33fc9449d39c1fadb620936d6531e" exitCode=0 Feb 25 08:26:56 crc kubenswrapper[4978]: I0225 08:26:56.782770 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c7be4547-b312-4601-93bf-9a3c62a13829","Type":"ContainerDied","Data":"2a2fd043e6faaaf313c9f49049c0d0747ed33fc9449d39c1fadb620936d6531e"} Feb 25 08:26:56 crc kubenswrapper[4978]: I0225 08:26:56.786325 4978 generic.go:334] "Generic (PLEG): container finished" podID="11a899ad-0164-4f8b-ae66-797557d34867" containerID="34fc60e992e78baa5a3cf4650a1088b2da5fb07125819dcd4d7b534e255a953f" exitCode=0 Feb 25 08:26:56 crc kubenswrapper[4978]: I0225 08:26:56.786449 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"11a899ad-0164-4f8b-ae66-797557d34867","Type":"ContainerDied","Data":"34fc60e992e78baa5a3cf4650a1088b2da5fb07125819dcd4d7b534e255a953f"} Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.029675 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.096800 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6954dff675-qbt9l"] Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.097036 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" podUID="11b9ad4c-6d6e-4bd8-8d59-c77545feb841" containerName="dnsmasq-dns" containerID="cri-o://18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f" gracePeriod=10 Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.456031 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.579219 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lcvb\" (UniqueName: \"kubernetes.io/projected/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-kube-api-access-9lcvb\") pod \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.579289 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-dns-svc\") pod \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.579389 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-config\") pod \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\" (UID: \"11b9ad4c-6d6e-4bd8-8d59-c77545feb841\") " Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.584510 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-kube-api-access-9lcvb" (OuterVolumeSpecName: "kube-api-access-9lcvb") pod "11b9ad4c-6d6e-4bd8-8d59-c77545feb841" (UID: "11b9ad4c-6d6e-4bd8-8d59-c77545feb841"). InnerVolumeSpecName "kube-api-access-9lcvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.609052 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11b9ad4c-6d6e-4bd8-8d59-c77545feb841" (UID: "11b9ad4c-6d6e-4bd8-8d59-c77545feb841"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.611850 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-config" (OuterVolumeSpecName: "config") pod "11b9ad4c-6d6e-4bd8-8d59-c77545feb841" (UID: "11b9ad4c-6d6e-4bd8-8d59-c77545feb841"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.680586 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lcvb\" (UniqueName: \"kubernetes.io/projected/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-kube-api-access-9lcvb\") on node \"crc\" DevicePath \"\"" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.680617 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.680627 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11b9ad4c-6d6e-4bd8-8d59-c77545feb841-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.800800 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"c7be4547-b312-4601-93bf-9a3c62a13829","Type":"ContainerStarted","Data":"9a3f2fed253e557fb7a6bc5608b7ef727568cb04ac86a524fd6337aaaa752888"} Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.808813 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"11a899ad-0164-4f8b-ae66-797557d34867","Type":"ContainerStarted","Data":"70e70609742eacc1230808b288fd3330d8a922ae9e34f13c2a5d1e1a01b1a115"} Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.817400 4978 generic.go:334] "Generic (PLEG): container finished" podID="11b9ad4c-6d6e-4bd8-8d59-c77545feb841" containerID="18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f" exitCode=0 Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.817458 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" event={"ID":"11b9ad4c-6d6e-4bd8-8d59-c77545feb841","Type":"ContainerDied","Data":"18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f"} Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.817501 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" event={"ID":"11b9ad4c-6d6e-4bd8-8d59-c77545feb841","Type":"ContainerDied","Data":"b02f8fcf2eaa3cc9ebe642912b32ac78d6bbc613fa9b2e43bb05745921197f6d"} Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.817529 4978 scope.go:117] "RemoveContainer" containerID="18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.817659 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6954dff675-qbt9l" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.839404 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.804225745 podStartE2EDuration="28.839384714s" podCreationTimestamp="2026-02-25 08:26:29 +0000 UTC" firstStartedPulling="2026-02-25 08:26:31.087309051 +0000 UTC m=+6084.526565500" lastFinishedPulling="2026-02-25 08:26:52.122468 +0000 UTC m=+6105.561724469" observedRunningTime="2026-02-25 08:26:57.837615109 +0000 UTC m=+6111.276871638" watchObservedRunningTime="2026-02-25 08:26:57.839384714 +0000 UTC m=+6111.278641173" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.892297 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.457454817 podStartE2EDuration="30.892249099s" podCreationTimestamp="2026-02-25 08:26:27 +0000 UTC" firstStartedPulling="2026-02-25 08:26:29.723995361 +0000 UTC m=+6083.163251820" lastFinishedPulling="2026-02-25 08:26:52.158789643 +0000 UTC m=+6105.598046102" observedRunningTime="2026-02-25 08:26:57.872250941 +0000 UTC m=+6111.311507440" watchObservedRunningTime="2026-02-25 08:26:57.892249099 +0000 UTC m=+6111.331505578" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.895696 4978 scope.go:117] "RemoveContainer" containerID="487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.902093 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6954dff675-qbt9l"] Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.920495 4978 scope.go:117] "RemoveContainer" containerID="18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f" Feb 25 08:26:57 crc kubenswrapper[4978]: E0225 08:26:57.920998 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f\": container with ID starting with 18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f not found: ID does not exist" containerID="18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.921031 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f"} err="failed to get container status \"18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f\": rpc error: code = NotFound desc = could not find container \"18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f\": container with ID starting with 18ab45871067c9674bf6d1441d85df2d358b624b0eeae7b36dc29bccbb10f36f not found: ID does not exist" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.921052 4978 scope.go:117] "RemoveContainer" containerID="487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4" Feb 25 08:26:57 crc kubenswrapper[4978]: E0225 08:26:57.921344 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4\": container with ID starting with 487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4 not found: ID does not exist" containerID="487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.921361 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4"} err="failed to get container status \"487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4\": rpc error: code = NotFound desc = could not find container \"487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4\": container with ID starting with 487ff0ccc79d123b3bc3519dbba9db708fa315af405ef4aeb9d64a81047454a4 not found: ID does not exist" Feb 25 08:26:57 crc kubenswrapper[4978]: I0225 08:26:57.928550 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6954dff675-qbt9l"] Feb 25 08:26:59 crc kubenswrapper[4978]: I0225 08:26:59.216964 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 25 08:26:59 crc kubenswrapper[4978]: I0225 08:26:59.217333 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 25 08:26:59 crc kubenswrapper[4978]: I0225 08:26:59.343303 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11b9ad4c-6d6e-4bd8-8d59-c77545feb841" path="/var/lib/kubelet/pods/11b9ad4c-6d6e-4bd8-8d59-c77545feb841/volumes" Feb 25 08:26:59 crc kubenswrapper[4978]: E0225 08:26:59.919613 4978 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.146:46046->38.102.83.146:35177: write tcp 38.102.83.146:46046->38.102.83.146:35177: write: broken pipe Feb 25 08:27:00 crc kubenswrapper[4978]: I0225 08:27:00.584410 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 25 08:27:00 crc kubenswrapper[4978]: I0225 08:27:00.584443 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 25 08:27:00 crc kubenswrapper[4978]: I0225 08:27:00.804561 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 25 08:27:01 crc kubenswrapper[4978]: I0225 08:27:01.865164 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 25 08:27:01 crc kubenswrapper[4978]: I0225 08:27:01.963436 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 25 08:27:04 crc kubenswrapper[4978]: I0225 08:27:04.690179 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 25 08:27:04 crc kubenswrapper[4978]: I0225 08:27:04.788049 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.887712 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-gvrv2"] Feb 25 08:27:07 crc kubenswrapper[4978]: E0225 08:27:07.889111 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b9ad4c-6d6e-4bd8-8d59-c77545feb841" containerName="dnsmasq-dns" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.889132 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b9ad4c-6d6e-4bd8-8d59-c77545feb841" containerName="dnsmasq-dns" Feb 25 08:27:07 crc kubenswrapper[4978]: E0225 08:27:07.889172 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11b9ad4c-6d6e-4bd8-8d59-c77545feb841" containerName="init" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.889185 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="11b9ad4c-6d6e-4bd8-8d59-c77545feb841" containerName="init" Feb 25 08:27:07 crc kubenswrapper[4978]: E0225 08:27:07.889211 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bba8e0a-fc78-46b0-8e3d-915c910f1ca2" containerName="init" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.889224 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bba8e0a-fc78-46b0-8e3d-915c910f1ca2" containerName="init" Feb 25 08:27:07 crc kubenswrapper[4978]: E0225 08:27:07.889263 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5" containerName="init" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.889277 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5" containerName="init" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.889561 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec82c93f-3cd9-48a9-801a-1ebb3ffee6a5" containerName="init" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.889602 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bba8e0a-fc78-46b0-8e3d-915c910f1ca2" containerName="init" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.889629 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="11b9ad4c-6d6e-4bd8-8d59-c77545feb841" containerName="dnsmasq-dns" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.890586 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gvrv2" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.893750 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.919546 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-gvrv2"] Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.996165 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a797ed47-954e-463b-a7c6-3d58e91c5cfb-operator-scripts\") pod \"root-account-create-update-gvrv2\" (UID: \"a797ed47-954e-463b-a7c6-3d58e91c5cfb\") " pod="openstack/root-account-create-update-gvrv2" Feb 25 08:27:07 crc kubenswrapper[4978]: I0225 08:27:07.996251 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6twkb\" (UniqueName: \"kubernetes.io/projected/a797ed47-954e-463b-a7c6-3d58e91c5cfb-kube-api-access-6twkb\") pod \"root-account-create-update-gvrv2\" (UID: \"a797ed47-954e-463b-a7c6-3d58e91c5cfb\") " pod="openstack/root-account-create-update-gvrv2" Feb 25 08:27:08 crc kubenswrapper[4978]: I0225 08:27:08.098717 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6twkb\" (UniqueName: \"kubernetes.io/projected/a797ed47-954e-463b-a7c6-3d58e91c5cfb-kube-api-access-6twkb\") pod \"root-account-create-update-gvrv2\" (UID: \"a797ed47-954e-463b-a7c6-3d58e91c5cfb\") " pod="openstack/root-account-create-update-gvrv2" Feb 25 08:27:08 crc kubenswrapper[4978]: I0225 08:27:08.099002 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a797ed47-954e-463b-a7c6-3d58e91c5cfb-operator-scripts\") pod \"root-account-create-update-gvrv2\" (UID: \"a797ed47-954e-463b-a7c6-3d58e91c5cfb\") " pod="openstack/root-account-create-update-gvrv2" Feb 25 08:27:08 crc kubenswrapper[4978]: I0225 08:27:08.100537 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a797ed47-954e-463b-a7c6-3d58e91c5cfb-operator-scripts\") pod \"root-account-create-update-gvrv2\" (UID: \"a797ed47-954e-463b-a7c6-3d58e91c5cfb\") " pod="openstack/root-account-create-update-gvrv2" Feb 25 08:27:08 crc kubenswrapper[4978]: I0225 08:27:08.143842 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6twkb\" (UniqueName: \"kubernetes.io/projected/a797ed47-954e-463b-a7c6-3d58e91c5cfb-kube-api-access-6twkb\") pod \"root-account-create-update-gvrv2\" (UID: \"a797ed47-954e-463b-a7c6-3d58e91c5cfb\") " pod="openstack/root-account-create-update-gvrv2" Feb 25 08:27:08 crc kubenswrapper[4978]: I0225 08:27:08.220613 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gvrv2" Feb 25 08:27:08 crc kubenswrapper[4978]: W0225 08:27:08.762959 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda797ed47_954e_463b_a7c6_3d58e91c5cfb.slice/crio-1a85634c62d1240a443de7f60e5e9e721c15cca76508d813a33b47b2f90db14e WatchSource:0}: Error finding container 1a85634c62d1240a443de7f60e5e9e721c15cca76508d813a33b47b2f90db14e: Status 404 returned error can't find the container with id 1a85634c62d1240a443de7f60e5e9e721c15cca76508d813a33b47b2f90db14e Feb 25 08:27:08 crc kubenswrapper[4978]: I0225 08:27:08.764841 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-gvrv2"] Feb 25 08:27:08 crc kubenswrapper[4978]: I0225 08:27:08.942969 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gvrv2" event={"ID":"a797ed47-954e-463b-a7c6-3d58e91c5cfb","Type":"ContainerStarted","Data":"1a85634c62d1240a443de7f60e5e9e721c15cca76508d813a33b47b2f90db14e"} Feb 25 08:27:09 crc kubenswrapper[4978]: I0225 08:27:09.955280 4978 generic.go:334] "Generic (PLEG): container finished" podID="a797ed47-954e-463b-a7c6-3d58e91c5cfb" containerID="ab593c3ff7696a5ce5d120903b7d0230eb4b7dfd21902abd1ee0048bd253f13f" exitCode=0 Feb 25 08:27:09 crc kubenswrapper[4978]: I0225 08:27:09.955358 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gvrv2" event={"ID":"a797ed47-954e-463b-a7c6-3d58e91c5cfb","Type":"ContainerDied","Data":"ab593c3ff7696a5ce5d120903b7d0230eb4b7dfd21902abd1ee0048bd253f13f"} Feb 25 08:27:11 crc kubenswrapper[4978]: I0225 08:27:11.335562 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gvrv2" Feb 25 08:27:11 crc kubenswrapper[4978]: I0225 08:27:11.454318 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6twkb\" (UniqueName: \"kubernetes.io/projected/a797ed47-954e-463b-a7c6-3d58e91c5cfb-kube-api-access-6twkb\") pod \"a797ed47-954e-463b-a7c6-3d58e91c5cfb\" (UID: \"a797ed47-954e-463b-a7c6-3d58e91c5cfb\") " Feb 25 08:27:11 crc kubenswrapper[4978]: I0225 08:27:11.454587 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a797ed47-954e-463b-a7c6-3d58e91c5cfb-operator-scripts\") pod \"a797ed47-954e-463b-a7c6-3d58e91c5cfb\" (UID: \"a797ed47-954e-463b-a7c6-3d58e91c5cfb\") " Feb 25 08:27:11 crc kubenswrapper[4978]: I0225 08:27:11.455755 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a797ed47-954e-463b-a7c6-3d58e91c5cfb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a797ed47-954e-463b-a7c6-3d58e91c5cfb" (UID: "a797ed47-954e-463b-a7c6-3d58e91c5cfb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:27:11 crc kubenswrapper[4978]: I0225 08:27:11.463727 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a797ed47-954e-463b-a7c6-3d58e91c5cfb-kube-api-access-6twkb" (OuterVolumeSpecName: "kube-api-access-6twkb") pod "a797ed47-954e-463b-a7c6-3d58e91c5cfb" (UID: "a797ed47-954e-463b-a7c6-3d58e91c5cfb"). InnerVolumeSpecName "kube-api-access-6twkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:27:11 crc kubenswrapper[4978]: I0225 08:27:11.556312 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6twkb\" (UniqueName: \"kubernetes.io/projected/a797ed47-954e-463b-a7c6-3d58e91c5cfb-kube-api-access-6twkb\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:11 crc kubenswrapper[4978]: I0225 08:27:11.556402 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a797ed47-954e-463b-a7c6-3d58e91c5cfb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:11 crc kubenswrapper[4978]: I0225 08:27:11.975754 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gvrv2" event={"ID":"a797ed47-954e-463b-a7c6-3d58e91c5cfb","Type":"ContainerDied","Data":"1a85634c62d1240a443de7f60e5e9e721c15cca76508d813a33b47b2f90db14e"} Feb 25 08:27:11 crc kubenswrapper[4978]: I0225 08:27:11.975810 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a85634c62d1240a443de7f60e5e9e721c15cca76508d813a33b47b2f90db14e" Feb 25 08:27:11 crc kubenswrapper[4978]: I0225 08:27:11.975842 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gvrv2" Feb 25 08:27:14 crc kubenswrapper[4978]: I0225 08:27:14.262230 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-gvrv2"] Feb 25 08:27:14 crc kubenswrapper[4978]: I0225 08:27:14.278389 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-gvrv2"] Feb 25 08:27:15 crc kubenswrapper[4978]: I0225 08:27:15.343951 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a797ed47-954e-463b-a7c6-3d58e91c5cfb" path="/var/lib/kubelet/pods/a797ed47-954e-463b-a7c6-3d58e91c5cfb/volumes" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.286760 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-5zkv5"] Feb 25 08:27:19 crc kubenswrapper[4978]: E0225 08:27:19.287575 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a797ed47-954e-463b-a7c6-3d58e91c5cfb" containerName="mariadb-account-create-update" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.287591 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a797ed47-954e-463b-a7c6-3d58e91c5cfb" containerName="mariadb-account-create-update" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.287753 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a797ed47-954e-463b-a7c6-3d58e91c5cfb" containerName="mariadb-account-create-update" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.288295 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5zkv5" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.291195 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.322797 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-5zkv5"] Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.388838 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd608922-a6fb-4b0a-b693-a2409079044e-operator-scripts\") pod \"root-account-create-update-5zkv5\" (UID: \"bd608922-a6fb-4b0a-b693-a2409079044e\") " pod="openstack/root-account-create-update-5zkv5" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.388985 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjr72\" (UniqueName: \"kubernetes.io/projected/bd608922-a6fb-4b0a-b693-a2409079044e-kube-api-access-jjr72\") pod \"root-account-create-update-5zkv5\" (UID: \"bd608922-a6fb-4b0a-b693-a2409079044e\") " pod="openstack/root-account-create-update-5zkv5" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.490849 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd608922-a6fb-4b0a-b693-a2409079044e-operator-scripts\") pod \"root-account-create-update-5zkv5\" (UID: \"bd608922-a6fb-4b0a-b693-a2409079044e\") " pod="openstack/root-account-create-update-5zkv5" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.490966 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjr72\" (UniqueName: \"kubernetes.io/projected/bd608922-a6fb-4b0a-b693-a2409079044e-kube-api-access-jjr72\") pod \"root-account-create-update-5zkv5\" (UID: \"bd608922-a6fb-4b0a-b693-a2409079044e\") " pod="openstack/root-account-create-update-5zkv5" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.492346 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd608922-a6fb-4b0a-b693-a2409079044e-operator-scripts\") pod \"root-account-create-update-5zkv5\" (UID: \"bd608922-a6fb-4b0a-b693-a2409079044e\") " pod="openstack/root-account-create-update-5zkv5" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.525706 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjr72\" (UniqueName: \"kubernetes.io/projected/bd608922-a6fb-4b0a-b693-a2409079044e-kube-api-access-jjr72\") pod \"root-account-create-update-5zkv5\" (UID: \"bd608922-a6fb-4b0a-b693-a2409079044e\") " pod="openstack/root-account-create-update-5zkv5" Feb 25 08:27:19 crc kubenswrapper[4978]: I0225 08:27:19.632591 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5zkv5" Feb 25 08:27:20 crc kubenswrapper[4978]: I0225 08:27:20.154766 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-5zkv5"] Feb 25 08:27:21 crc kubenswrapper[4978]: I0225 08:27:21.068261 4978 generic.go:334] "Generic (PLEG): container finished" podID="bd608922-a6fb-4b0a-b693-a2409079044e" containerID="ab6e81ad25d93636134c19ac61fed0a82345386a33ad7e66e753035460048efb" exitCode=0 Feb 25 08:27:21 crc kubenswrapper[4978]: I0225 08:27:21.068335 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5zkv5" event={"ID":"bd608922-a6fb-4b0a-b693-a2409079044e","Type":"ContainerDied","Data":"ab6e81ad25d93636134c19ac61fed0a82345386a33ad7e66e753035460048efb"} Feb 25 08:27:21 crc kubenswrapper[4978]: I0225 08:27:21.068413 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5zkv5" event={"ID":"bd608922-a6fb-4b0a-b693-a2409079044e","Type":"ContainerStarted","Data":"11cc42bdbc5ba945b3f603bbfb49d6688a703c1b14a49de255a8c59dcd27a55c"} Feb 25 08:27:22 crc kubenswrapper[4978]: I0225 08:27:22.524093 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5zkv5" Feb 25 08:27:22 crc kubenswrapper[4978]: I0225 08:27:22.643881 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd608922-a6fb-4b0a-b693-a2409079044e-operator-scripts\") pod \"bd608922-a6fb-4b0a-b693-a2409079044e\" (UID: \"bd608922-a6fb-4b0a-b693-a2409079044e\") " Feb 25 08:27:22 crc kubenswrapper[4978]: I0225 08:27:22.644147 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjr72\" (UniqueName: \"kubernetes.io/projected/bd608922-a6fb-4b0a-b693-a2409079044e-kube-api-access-jjr72\") pod \"bd608922-a6fb-4b0a-b693-a2409079044e\" (UID: \"bd608922-a6fb-4b0a-b693-a2409079044e\") " Feb 25 08:27:22 crc kubenswrapper[4978]: I0225 08:27:22.646461 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd608922-a6fb-4b0a-b693-a2409079044e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bd608922-a6fb-4b0a-b693-a2409079044e" (UID: "bd608922-a6fb-4b0a-b693-a2409079044e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:27:22 crc kubenswrapper[4978]: I0225 08:27:22.652928 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd608922-a6fb-4b0a-b693-a2409079044e-kube-api-access-jjr72" (OuterVolumeSpecName: "kube-api-access-jjr72") pod "bd608922-a6fb-4b0a-b693-a2409079044e" (UID: "bd608922-a6fb-4b0a-b693-a2409079044e"). InnerVolumeSpecName "kube-api-access-jjr72". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:27:22 crc kubenswrapper[4978]: I0225 08:27:22.746135 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjr72\" (UniqueName: \"kubernetes.io/projected/bd608922-a6fb-4b0a-b693-a2409079044e-kube-api-access-jjr72\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:22 crc kubenswrapper[4978]: I0225 08:27:22.746186 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd608922-a6fb-4b0a-b693-a2409079044e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:23 crc kubenswrapper[4978]: I0225 08:27:23.093043 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5zkv5" event={"ID":"bd608922-a6fb-4b0a-b693-a2409079044e","Type":"ContainerDied","Data":"11cc42bdbc5ba945b3f603bbfb49d6688a703c1b14a49de255a8c59dcd27a55c"} Feb 25 08:27:23 crc kubenswrapper[4978]: I0225 08:27:23.093104 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11cc42bdbc5ba945b3f603bbfb49d6688a703c1b14a49de255a8c59dcd27a55c" Feb 25 08:27:23 crc kubenswrapper[4978]: I0225 08:27:23.093132 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5zkv5" Feb 25 08:27:27 crc kubenswrapper[4978]: I0225 08:27:27.136945 4978 generic.go:334] "Generic (PLEG): container finished" podID="de4db333-9150-4ccd-8f4e-d2a18917bbc5" containerID="8eb08550ac619cc24779584a8a60dad408e75a3636fff54ad14f71daa4268556" exitCode=0 Feb 25 08:27:27 crc kubenswrapper[4978]: I0225 08:27:27.137050 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"de4db333-9150-4ccd-8f4e-d2a18917bbc5","Type":"ContainerDied","Data":"8eb08550ac619cc24779584a8a60dad408e75a3636fff54ad14f71daa4268556"} Feb 25 08:27:27 crc kubenswrapper[4978]: I0225 08:27:27.139394 4978 generic.go:334] "Generic (PLEG): container finished" podID="e4186a72-0d11-4647-902b-3908454bf0b1" containerID="851781b6535eb50c799c75b76137c93e166f97d3eb4f6b3617eddc52c7e6db47" exitCode=0 Feb 25 08:27:27 crc kubenswrapper[4978]: I0225 08:27:27.139440 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4186a72-0d11-4647-902b-3908454bf0b1","Type":"ContainerDied","Data":"851781b6535eb50c799c75b76137c93e166f97d3eb4f6b3617eddc52c7e6db47"} Feb 25 08:27:28 crc kubenswrapper[4978]: I0225 08:27:28.151544 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4186a72-0d11-4647-902b-3908454bf0b1","Type":"ContainerStarted","Data":"d05989a159a711fa2990b91a0facc0678145a2d383e9beebb9a32cc5081f789f"} Feb 25 08:27:28 crc kubenswrapper[4978]: I0225 08:27:28.152124 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 25 08:27:28 crc kubenswrapper[4978]: I0225 08:27:28.154735 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"de4db333-9150-4ccd-8f4e-d2a18917bbc5","Type":"ContainerStarted","Data":"2836e002724c7861d4a21769b1fe8f1005982b144fb7e7154401b54a1b8f113b"} Feb 25 08:27:28 crc kubenswrapper[4978]: I0225 08:27:28.155001 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:28 crc kubenswrapper[4978]: I0225 08:27:28.193050 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.219267671 podStartE2EDuration="1m2.193026306s" podCreationTimestamp="2026-02-25 08:26:26 +0000 UTC" firstStartedPulling="2026-02-25 08:26:30.128400456 +0000 UTC m=+6083.567656915" lastFinishedPulling="2026-02-25 08:26:52.102159091 +0000 UTC m=+6105.541415550" observedRunningTime="2026-02-25 08:27:28.188940819 +0000 UTC m=+6141.628197298" watchObservedRunningTime="2026-02-25 08:27:28.193026306 +0000 UTC m=+6141.632282795" Feb 25 08:27:28 crc kubenswrapper[4978]: I0225 08:27:28.231156 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.776508287 podStartE2EDuration="1m2.231127637s" podCreationTimestamp="2026-02-25 08:26:26 +0000 UTC" firstStartedPulling="2026-02-25 08:26:28.586096671 +0000 UTC m=+6082.025353130" lastFinishedPulling="2026-02-25 08:26:52.040715991 +0000 UTC m=+6105.479972480" observedRunningTime="2026-02-25 08:27:28.224768999 +0000 UTC m=+6141.664025478" watchObservedRunningTime="2026-02-25 08:27:28.231127637 +0000 UTC m=+6141.670384136" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.097019 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m96pf"] Feb 25 08:27:29 crc kubenswrapper[4978]: E0225 08:27:29.097796 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd608922-a6fb-4b0a-b693-a2409079044e" containerName="mariadb-account-create-update" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.097824 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd608922-a6fb-4b0a-b693-a2409079044e" containerName="mariadb-account-create-update" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.098047 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd608922-a6fb-4b0a-b693-a2409079044e" containerName="mariadb-account-create-update" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.099695 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.114581 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m96pf"] Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.250866 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-catalog-content\") pod \"redhat-marketplace-m96pf\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.250958 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-utilities\") pod \"redhat-marketplace-m96pf\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.251061 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4vft\" (UniqueName: \"kubernetes.io/projected/0fce40e4-c06d-4526-92b0-4544965da2c9-kube-api-access-v4vft\") pod \"redhat-marketplace-m96pf\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.352851 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4vft\" (UniqueName: \"kubernetes.io/projected/0fce40e4-c06d-4526-92b0-4544965da2c9-kube-api-access-v4vft\") pod \"redhat-marketplace-m96pf\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.353609 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-catalog-content\") pod \"redhat-marketplace-m96pf\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.353735 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-utilities\") pod \"redhat-marketplace-m96pf\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.354131 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-utilities\") pod \"redhat-marketplace-m96pf\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.354724 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-catalog-content\") pod \"redhat-marketplace-m96pf\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.378717 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4vft\" (UniqueName: \"kubernetes.io/projected/0fce40e4-c06d-4526-92b0-4544965da2c9-kube-api-access-v4vft\") pod \"redhat-marketplace-m96pf\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.423695 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:29 crc kubenswrapper[4978]: I0225 08:27:29.673485 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m96pf"] Feb 25 08:27:30 crc kubenswrapper[4978]: I0225 08:27:30.184836 4978 generic.go:334] "Generic (PLEG): container finished" podID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerID="e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e" exitCode=0 Feb 25 08:27:30 crc kubenswrapper[4978]: I0225 08:27:30.184954 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m96pf" event={"ID":"0fce40e4-c06d-4526-92b0-4544965da2c9","Type":"ContainerDied","Data":"e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e"} Feb 25 08:27:30 crc kubenswrapper[4978]: I0225 08:27:30.185345 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m96pf" event={"ID":"0fce40e4-c06d-4526-92b0-4544965da2c9","Type":"ContainerStarted","Data":"f6fab243f42c4d493e5146225a1d45da07884aacb4d59918cec64721bbfb914a"} Feb 25 08:27:30 crc kubenswrapper[4978]: I0225 08:27:30.187825 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 08:27:31 crc kubenswrapper[4978]: I0225 08:27:31.202430 4978 generic.go:334] "Generic (PLEG): container finished" podID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerID="023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571" exitCode=0 Feb 25 08:27:31 crc kubenswrapper[4978]: I0225 08:27:31.202486 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m96pf" event={"ID":"0fce40e4-c06d-4526-92b0-4544965da2c9","Type":"ContainerDied","Data":"023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571"} Feb 25 08:27:32 crc kubenswrapper[4978]: I0225 08:27:32.215597 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m96pf" event={"ID":"0fce40e4-c06d-4526-92b0-4544965da2c9","Type":"ContainerStarted","Data":"f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584"} Feb 25 08:27:32 crc kubenswrapper[4978]: I0225 08:27:32.243435 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m96pf" podStartSLOduration=1.848593786 podStartE2EDuration="3.243405393s" podCreationTimestamp="2026-02-25 08:27:29 +0000 UTC" firstStartedPulling="2026-02-25 08:27:30.187584164 +0000 UTC m=+6143.626840624" lastFinishedPulling="2026-02-25 08:27:31.582395762 +0000 UTC m=+6145.021652231" observedRunningTime="2026-02-25 08:27:32.236354754 +0000 UTC m=+6145.675611303" watchObservedRunningTime="2026-02-25 08:27:32.243405393 +0000 UTC m=+6145.682661892" Feb 25 08:27:38 crc kubenswrapper[4978]: I0225 08:27:38.154990 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:39 crc kubenswrapper[4978]: I0225 08:27:39.424780 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:39 crc kubenswrapper[4978]: I0225 08:27:39.425233 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:39 crc kubenswrapper[4978]: I0225 08:27:39.488664 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:39 crc kubenswrapper[4978]: I0225 08:27:39.683656 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 25 08:27:40 crc kubenswrapper[4978]: I0225 08:27:40.388889 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:40 crc kubenswrapper[4978]: I0225 08:27:40.432644 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m96pf"] Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.341924 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m96pf" podUID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerName="registry-server" containerID="cri-o://f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584" gracePeriod=2 Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.848792 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76d5b4c5b5-crsmw"] Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.850489 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.872693 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.882749 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76d5b4c5b5-crsmw"] Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.893940 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb9h2\" (UniqueName: \"kubernetes.io/projected/f887ca94-f583-4674-b0ed-8b45710a1804-kube-api-access-wb9h2\") pod \"dnsmasq-dns-76d5b4c5b5-crsmw\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.894051 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-config\") pod \"dnsmasq-dns-76d5b4c5b5-crsmw\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.894086 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-dns-svc\") pod \"dnsmasq-dns-76d5b4c5b5-crsmw\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.995654 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4vft\" (UniqueName: \"kubernetes.io/projected/0fce40e4-c06d-4526-92b0-4544965da2c9-kube-api-access-v4vft\") pod \"0fce40e4-c06d-4526-92b0-4544965da2c9\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.996021 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-utilities\") pod \"0fce40e4-c06d-4526-92b0-4544965da2c9\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.996053 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-catalog-content\") pod \"0fce40e4-c06d-4526-92b0-4544965da2c9\" (UID: \"0fce40e4-c06d-4526-92b0-4544965da2c9\") " Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.996364 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb9h2\" (UniqueName: \"kubernetes.io/projected/f887ca94-f583-4674-b0ed-8b45710a1804-kube-api-access-wb9h2\") pod \"dnsmasq-dns-76d5b4c5b5-crsmw\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.996436 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-config\") pod \"dnsmasq-dns-76d5b4c5b5-crsmw\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.996460 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-dns-svc\") pod \"dnsmasq-dns-76d5b4c5b5-crsmw\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.996970 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-utilities" (OuterVolumeSpecName: "utilities") pod "0fce40e4-c06d-4526-92b0-4544965da2c9" (UID: "0fce40e4-c06d-4526-92b0-4544965da2c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.997397 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-dns-svc\") pod \"dnsmasq-dns-76d5b4c5b5-crsmw\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:42 crc kubenswrapper[4978]: I0225 08:27:42.997408 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-config\") pod \"dnsmasq-dns-76d5b4c5b5-crsmw\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.009428 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fce40e4-c06d-4526-92b0-4544965da2c9-kube-api-access-v4vft" (OuterVolumeSpecName: "kube-api-access-v4vft") pod "0fce40e4-c06d-4526-92b0-4544965da2c9" (UID: "0fce40e4-c06d-4526-92b0-4544965da2c9"). InnerVolumeSpecName "kube-api-access-v4vft". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.015742 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb9h2\" (UniqueName: \"kubernetes.io/projected/f887ca94-f583-4674-b0ed-8b45710a1804-kube-api-access-wb9h2\") pod \"dnsmasq-dns-76d5b4c5b5-crsmw\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.040193 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0fce40e4-c06d-4526-92b0-4544965da2c9" (UID: "0fce40e4-c06d-4526-92b0-4544965da2c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.098268 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4vft\" (UniqueName: \"kubernetes.io/projected/0fce40e4-c06d-4526-92b0-4544965da2c9-kube-api-access-v4vft\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.098304 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.098315 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fce40e4-c06d-4526-92b0-4544965da2c9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.188125 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.350957 4978 generic.go:334] "Generic (PLEG): container finished" podID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerID="f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584" exitCode=0 Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.351008 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m96pf" event={"ID":"0fce40e4-c06d-4526-92b0-4544965da2c9","Type":"ContainerDied","Data":"f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584"} Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.351037 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m96pf" event={"ID":"0fce40e4-c06d-4526-92b0-4544965da2c9","Type":"ContainerDied","Data":"f6fab243f42c4d493e5146225a1d45da07884aacb4d59918cec64721bbfb914a"} Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.351058 4978 scope.go:117] "RemoveContainer" containerID="f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.351223 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m96pf" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.381669 4978 scope.go:117] "RemoveContainer" containerID="023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.403517 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m96pf"] Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.418802 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m96pf"] Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.428102 4978 scope.go:117] "RemoveContainer" containerID="e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.461606 4978 scope.go:117] "RemoveContainer" containerID="f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584" Feb 25 08:27:43 crc kubenswrapper[4978]: E0225 08:27:43.468466 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584\": container with ID starting with f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584 not found: ID does not exist" containerID="f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.468502 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584"} err="failed to get container status \"f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584\": rpc error: code = NotFound desc = could not find container \"f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584\": container with ID starting with f5ba8cc1d56058d84db3360249c6ba74105eab8f4397b84e0499de9d1cd82584 not found: ID does not exist" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.468530 4978 scope.go:117] "RemoveContainer" containerID="023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571" Feb 25 08:27:43 crc kubenswrapper[4978]: E0225 08:27:43.471670 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571\": container with ID starting with 023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571 not found: ID does not exist" containerID="023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.471707 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571"} err="failed to get container status \"023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571\": rpc error: code = NotFound desc = could not find container \"023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571\": container with ID starting with 023929e4bfba43f0f540cff8231d65c6186dc300cd43ff6fa2c462e077a5d571 not found: ID does not exist" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.471737 4978 scope.go:117] "RemoveContainer" containerID="e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e" Feb 25 08:27:43 crc kubenswrapper[4978]: E0225 08:27:43.474436 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e\": container with ID starting with e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e not found: ID does not exist" containerID="e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.474461 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e"} err="failed to get container status \"e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e\": rpc error: code = NotFound desc = could not find container \"e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e\": container with ID starting with e22e00809dabd4bc78bbd6cd3d6593b2f255adc90b55c233f871b5dc3847da0e not found: ID does not exist" Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.505779 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76d5b4c5b5-crsmw"] Feb 25 08:27:43 crc kubenswrapper[4978]: W0225 08:27:43.517487 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf887ca94_f583_4674_b0ed_8b45710a1804.slice/crio-f28f115feb721f8bc99ada00ef097dd49b06b27a2ea8b802c1a40a43fb2c85db WatchSource:0}: Error finding container f28f115feb721f8bc99ada00ef097dd49b06b27a2ea8b802c1a40a43fb2c85db: Status 404 returned error can't find the container with id f28f115feb721f8bc99ada00ef097dd49b06b27a2ea8b802c1a40a43fb2c85db Feb 25 08:27:43 crc kubenswrapper[4978]: I0225 08:27:43.621739 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 08:27:44 crc kubenswrapper[4978]: I0225 08:27:44.117849 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 08:27:44 crc kubenswrapper[4978]: I0225 08:27:44.358970 4978 generic.go:334] "Generic (PLEG): container finished" podID="f887ca94-f583-4674-b0ed-8b45710a1804" containerID="c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e" exitCode=0 Feb 25 08:27:44 crc kubenswrapper[4978]: I0225 08:27:44.359022 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" event={"ID":"f887ca94-f583-4674-b0ed-8b45710a1804","Type":"ContainerDied","Data":"c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e"} Feb 25 08:27:44 crc kubenswrapper[4978]: I0225 08:27:44.359235 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" event={"ID":"f887ca94-f583-4674-b0ed-8b45710a1804","Type":"ContainerStarted","Data":"f28f115feb721f8bc99ada00ef097dd49b06b27a2ea8b802c1a40a43fb2c85db"} Feb 25 08:27:45 crc kubenswrapper[4978]: I0225 08:27:45.339494 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fce40e4-c06d-4526-92b0-4544965da2c9" path="/var/lib/kubelet/pods/0fce40e4-c06d-4526-92b0-4544965da2c9/volumes" Feb 25 08:27:45 crc kubenswrapper[4978]: I0225 08:27:45.369147 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" event={"ID":"f887ca94-f583-4674-b0ed-8b45710a1804","Type":"ContainerStarted","Data":"4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7"} Feb 25 08:27:45 crc kubenswrapper[4978]: I0225 08:27:45.369317 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:45 crc kubenswrapper[4978]: I0225 08:27:45.387702 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" podStartSLOduration=3.3876803300000002 podStartE2EDuration="3.38768033s" podCreationTimestamp="2026-02-25 08:27:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:27:45.382931502 +0000 UTC m=+6158.822187961" watchObservedRunningTime="2026-02-25 08:27:45.38768033 +0000 UTC m=+6158.826936789" Feb 25 08:27:46 crc kubenswrapper[4978]: I0225 08:27:46.540984 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:27:46 crc kubenswrapper[4978]: I0225 08:27:46.541051 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:27:47 crc kubenswrapper[4978]: I0225 08:27:47.930334 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="de4db333-9150-4ccd-8f4e-d2a18917bbc5" containerName="rabbitmq" containerID="cri-o://2836e002724c7861d4a21769b1fe8f1005982b144fb7e7154401b54a1b8f113b" gracePeriod=604797 Feb 25 08:27:47 crc kubenswrapper[4978]: I0225 08:27:47.999265 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="e4186a72-0d11-4647-902b-3908454bf0b1" containerName="rabbitmq" containerID="cri-o://d05989a159a711fa2990b91a0facc0678145a2d383e9beebb9a32cc5081f789f" gracePeriod=604796 Feb 25 08:27:48 crc kubenswrapper[4978]: I0225 08:27:48.148466 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="de4db333-9150-4ccd-8f4e-d2a18917bbc5" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.41:5671: connect: connection refused" Feb 25 08:27:49 crc kubenswrapper[4978]: I0225 08:27:49.681273 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="e4186a72-0d11-4647-902b-3908454bf0b1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.43:5671: connect: connection refused" Feb 25 08:27:53 crc kubenswrapper[4978]: I0225 08:27:53.190611 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:27:53 crc kubenswrapper[4978]: I0225 08:27:53.262919 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54756c6d6f-gwvms"] Feb 25 08:27:53 crc kubenswrapper[4978]: I0225 08:27:53.263716 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" podUID="903e7f0c-1ec6-45a2-a69c-f3caed9750a3" containerName="dnsmasq-dns" containerID="cri-o://497158ec40a3a54beb4b4eae741dd2b5b21192570a6645c182ffc39cc18a8602" gracePeriod=10 Feb 25 08:27:53 crc kubenswrapper[4978]: I0225 08:27:53.444183 4978 generic.go:334] "Generic (PLEG): container finished" podID="903e7f0c-1ec6-45a2-a69c-f3caed9750a3" containerID="497158ec40a3a54beb4b4eae741dd2b5b21192570a6645c182ffc39cc18a8602" exitCode=0 Feb 25 08:27:53 crc kubenswrapper[4978]: I0225 08:27:53.444249 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" event={"ID":"903e7f0c-1ec6-45a2-a69c-f3caed9750a3","Type":"ContainerDied","Data":"497158ec40a3a54beb4b4eae741dd2b5b21192570a6645c182ffc39cc18a8602"} Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:53.749525 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:53.802987 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nh24\" (UniqueName: \"kubernetes.io/projected/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-kube-api-access-5nh24\") pod \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:53.803104 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-config\") pod \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:53.803158 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-dns-svc\") pod \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\" (UID: \"903e7f0c-1ec6-45a2-a69c-f3caed9750a3\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:53.810070 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-kube-api-access-5nh24" (OuterVolumeSpecName: "kube-api-access-5nh24") pod "903e7f0c-1ec6-45a2-a69c-f3caed9750a3" (UID: "903e7f0c-1ec6-45a2-a69c-f3caed9750a3"). InnerVolumeSpecName "kube-api-access-5nh24". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:53.846925 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "903e7f0c-1ec6-45a2-a69c-f3caed9750a3" (UID: "903e7f0c-1ec6-45a2-a69c-f3caed9750a3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:53.862259 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-config" (OuterVolumeSpecName: "config") pod "903e7f0c-1ec6-45a2-a69c-f3caed9750a3" (UID: "903e7f0c-1ec6-45a2-a69c-f3caed9750a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:53.904807 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:53.904837 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:53.904851 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nh24\" (UniqueName: \"kubernetes.io/projected/903e7f0c-1ec6-45a2-a69c-f3caed9750a3-kube-api-access-5nh24\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.474572 4978 generic.go:334] "Generic (PLEG): container finished" podID="e4186a72-0d11-4647-902b-3908454bf0b1" containerID="d05989a159a711fa2990b91a0facc0678145a2d383e9beebb9a32cc5081f789f" exitCode=0 Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.474660 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4186a72-0d11-4647-902b-3908454bf0b1","Type":"ContainerDied","Data":"d05989a159a711fa2990b91a0facc0678145a2d383e9beebb9a32cc5081f789f"} Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.478503 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" event={"ID":"903e7f0c-1ec6-45a2-a69c-f3caed9750a3","Type":"ContainerDied","Data":"7a30d587b7eb9eafd4fce6af6c433270da962ec6a54390dfe057c85753ba7c91"} Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.478610 4978 scope.go:117] "RemoveContainer" containerID="497158ec40a3a54beb4b4eae741dd2b5b21192570a6645c182ffc39cc18a8602" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.478519 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54756c6d6f-gwvms" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.481507 4978 generic.go:334] "Generic (PLEG): container finished" podID="de4db333-9150-4ccd-8f4e-d2a18917bbc5" containerID="2836e002724c7861d4a21769b1fe8f1005982b144fb7e7154401b54a1b8f113b" exitCode=0 Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.481538 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"de4db333-9150-4ccd-8f4e-d2a18917bbc5","Type":"ContainerDied","Data":"2836e002724c7861d4a21769b1fe8f1005982b144fb7e7154401b54a1b8f113b"} Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.481557 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"de4db333-9150-4ccd-8f4e-d2a18917bbc5","Type":"ContainerDied","Data":"afd671d54749b27b54e3d7002bc6190b2b37e36097b296d786045ee9b80bb6b0"} Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.481569 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afd671d54749b27b54e3d7002bc6190b2b37e36097b296d786045ee9b80bb6b0" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.513051 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.523683 4978 scope.go:117] "RemoveContainer" containerID="99e211e9a0378ea047273aa9024bc2e06fd87b2507530dfaa57750faf831891a" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.530752 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54756c6d6f-gwvms"] Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.536173 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54756c6d6f-gwvms"] Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.611244 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.617728 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.617782 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpqfd\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-kube-api-access-tpqfd\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.617841 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-config-data\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.617983 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-erlang-cookie\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.618068 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-plugins-conf\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.618120 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-plugins\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.618201 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/de4db333-9150-4ccd-8f4e-d2a18917bbc5-erlang-cookie-secret\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.618253 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-tls\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.618291 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-server-conf\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.618344 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-confd\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.618404 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/de4db333-9150-4ccd-8f4e-d2a18917bbc5-pod-info\") pod \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\" (UID: \"de4db333-9150-4ccd-8f4e-d2a18917bbc5\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.621239 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.621706 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.621714 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.623448 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4db333-9150-4ccd-8f4e-d2a18917bbc5-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.625605 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/de4db333-9150-4ccd-8f4e-d2a18917bbc5-pod-info" (OuterVolumeSpecName: "pod-info") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.626547 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.627959 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-kube-api-access-tpqfd" (OuterVolumeSpecName: "kube-api-access-tpqfd") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "kube-api-access-tpqfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.687063 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-config-data" (OuterVolumeSpecName: "config-data") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.690242 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc" (OuterVolumeSpecName: "persistence") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.705649 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-server-conf" (OuterVolumeSpecName: "server-conf") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.719586 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jpmq\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-kube-api-access-8jpmq\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.719626 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-plugins\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.719702 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-config-data\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.719736 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-tls\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.719775 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-server-conf\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.719860 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.719882 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4186a72-0d11-4647-902b-3908454bf0b1-pod-info\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.719900 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-confd\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.719922 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4186a72-0d11-4647-902b-3908454bf0b1-erlang-cookie-secret\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.719966 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-erlang-cookie\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720004 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-plugins-conf\") pod \"e4186a72-0d11-4647-902b-3908454bf0b1\" (UID: \"e4186a72-0d11-4647-902b-3908454bf0b1\") " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720227 4978 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/de4db333-9150-4ccd-8f4e-d2a18917bbc5-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720245 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720254 4978 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-server-conf\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720262 4978 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/de4db333-9150-4ccd-8f4e-d2a18917bbc5-pod-info\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720283 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") on node \"crc\" " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720294 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpqfd\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-kube-api-access-tpqfd\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720304 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720312 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720321 4978 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/de4db333-9150-4ccd-8f4e-d2a18917bbc5-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.720329 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.721061 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.721323 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.722164 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.722175 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-kube-api-access-8jpmq" (OuterVolumeSpecName: "kube-api-access-8jpmq") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "kube-api-access-8jpmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.725922 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4186a72-0d11-4647-902b-3908454bf0b1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.725933 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e4186a72-0d11-4647-902b-3908454bf0b1-pod-info" (OuterVolumeSpecName: "pod-info") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.726016 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.735090 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23" (OuterVolumeSpecName: "persistence") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "pvc-389bf769-1bfe-41e5-84a3-249075245d23". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.737632 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "de4db333-9150-4ccd-8f4e-d2a18917bbc5" (UID: "de4db333-9150-4ccd-8f4e-d2a18917bbc5"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.744736 4978 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.745662 4978 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc") on node "crc" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.746460 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-config-data" (OuterVolumeSpecName: "config-data") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.759174 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-server-conf" (OuterVolumeSpecName: "server-conf") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.817798 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e4186a72-0d11-4647-902b-3908454bf0b1" (UID: "e4186a72-0d11-4647-902b-3908454bf0b1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821497 4978 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-server-conf\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821555 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-389bf769-1bfe-41e5-84a3-249075245d23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") on node \"crc\" " Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821569 4978 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e4186a72-0d11-4647-902b-3908454bf0b1-pod-info\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821578 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821589 4978 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e4186a72-0d11-4647-902b-3908454bf0b1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821599 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/de4db333-9150-4ccd-8f4e-d2a18917bbc5-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821607 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821615 4978 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821624 4978 reconciler_common.go:293] "Volume detached for volume \"pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821634 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jpmq\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-kube-api-access-8jpmq\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821642 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821650 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e4186a72-0d11-4647-902b-3908454bf0b1-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.821657 4978 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e4186a72-0d11-4647-902b-3908454bf0b1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.838536 4978 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.838705 4978 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-389bf769-1bfe-41e5-84a3-249075245d23" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23") on node "crc" Feb 25 08:27:54 crc kubenswrapper[4978]: I0225 08:27:54.923591 4978 reconciler_common.go:293] "Volume detached for volume \"pvc-389bf769-1bfe-41e5-84a3-249075245d23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") on node \"crc\" DevicePath \"\"" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.342995 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="903e7f0c-1ec6-45a2-a69c-f3caed9750a3" path="/var/lib/kubelet/pods/903e7f0c-1ec6-45a2-a69c-f3caed9750a3/volumes" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.495994 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e4186a72-0d11-4647-902b-3908454bf0b1","Type":"ContainerDied","Data":"dad6a6fbcb1bcd42bf22231fca9315d95bac52a3bdfad72ca01ef3f01123055e"} Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.496035 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.496060 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.496259 4978 scope.go:117] "RemoveContainer" containerID="d05989a159a711fa2990b91a0facc0678145a2d383e9beebb9a32cc5081f789f" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.525385 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.531977 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.532478 4978 scope.go:117] "RemoveContainer" containerID="851781b6535eb50c799c75b76137c93e166f97d3eb4f6b3617eddc52c7e6db47" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.570651 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.584342 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.592136 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 08:27:55 crc kubenswrapper[4978]: E0225 08:27:55.593850 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerName="extract-utilities" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.593878 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerName="extract-utilities" Feb 25 08:27:55 crc kubenswrapper[4978]: E0225 08:27:55.593904 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="903e7f0c-1ec6-45a2-a69c-f3caed9750a3" containerName="init" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.593913 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="903e7f0c-1ec6-45a2-a69c-f3caed9750a3" containerName="init" Feb 25 08:27:55 crc kubenswrapper[4978]: E0225 08:27:55.593921 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerName="extract-content" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.593933 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerName="extract-content" Feb 25 08:27:55 crc kubenswrapper[4978]: E0225 08:27:55.593950 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4186a72-0d11-4647-902b-3908454bf0b1" containerName="rabbitmq" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.593958 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4186a72-0d11-4647-902b-3908454bf0b1" containerName="rabbitmq" Feb 25 08:27:55 crc kubenswrapper[4978]: E0225 08:27:55.593971 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4db333-9150-4ccd-8f4e-d2a18917bbc5" containerName="setup-container" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.593979 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4db333-9150-4ccd-8f4e-d2a18917bbc5" containerName="setup-container" Feb 25 08:27:55 crc kubenswrapper[4978]: E0225 08:27:55.593995 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4186a72-0d11-4647-902b-3908454bf0b1" containerName="setup-container" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.594002 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4186a72-0d11-4647-902b-3908454bf0b1" containerName="setup-container" Feb 25 08:27:55 crc kubenswrapper[4978]: E0225 08:27:55.594013 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerName="registry-server" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.594020 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerName="registry-server" Feb 25 08:27:55 crc kubenswrapper[4978]: E0225 08:27:55.594038 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="903e7f0c-1ec6-45a2-a69c-f3caed9750a3" containerName="dnsmasq-dns" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.594047 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="903e7f0c-1ec6-45a2-a69c-f3caed9750a3" containerName="dnsmasq-dns" Feb 25 08:27:55 crc kubenswrapper[4978]: E0225 08:27:55.594063 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4db333-9150-4ccd-8f4e-d2a18917bbc5" containerName="rabbitmq" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.594071 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4db333-9150-4ccd-8f4e-d2a18917bbc5" containerName="rabbitmq" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.594243 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="903e7f0c-1ec6-45a2-a69c-f3caed9750a3" containerName="dnsmasq-dns" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.594257 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="de4db333-9150-4ccd-8f4e-d2a18917bbc5" containerName="rabbitmq" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.594270 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4186a72-0d11-4647-902b-3908454bf0b1" containerName="rabbitmq" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.594285 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fce40e4-c06d-4526-92b0-4544965da2c9" containerName="registry-server" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.595191 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.597340 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.600238 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-wsprr" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.600565 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.600902 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.600597 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.600660 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.600767 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.606700 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.608744 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.613411 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.613683 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.614432 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.614672 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.614826 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-7ch67" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.614990 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.615102 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.620178 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636421 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmj6x\" (UniqueName: \"kubernetes.io/projected/785d8daf-cc33-4011-9767-fa6049438fa8-kube-api-access-mmj6x\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636478 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/785d8daf-cc33-4011-9767-fa6049438fa8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636536 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/785d8daf-cc33-4011-9767-fa6049438fa8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636568 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636594 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/785d8daf-cc33-4011-9767-fa6049438fa8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636626 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/785d8daf-cc33-4011-9767-fa6049438fa8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636647 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636679 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636701 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636728 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-389bf769-1bfe-41e5-84a3-249075245d23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.636751 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/785d8daf-cc33-4011-9767-fa6049438fa8-config-data\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.648398 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738356 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/787e6909-959b-475a-a300-5ff751242c20-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738417 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/785d8daf-cc33-4011-9767-fa6049438fa8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738450 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfl8w\" (UniqueName: \"kubernetes.io/projected/787e6909-959b-475a-a300-5ff751242c20-kube-api-access-pfl8w\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738471 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738489 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/785d8daf-cc33-4011-9767-fa6049438fa8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738510 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738535 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/785d8daf-cc33-4011-9767-fa6049438fa8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738551 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738574 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738588 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738606 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-389bf769-1bfe-41e5-84a3-249075245d23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738620 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738638 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/785d8daf-cc33-4011-9767-fa6049438fa8-config-data\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738655 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/787e6909-959b-475a-a300-5ff751242c20-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738676 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/787e6909-959b-475a-a300-5ff751242c20-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738692 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmj6x\" (UniqueName: \"kubernetes.io/projected/785d8daf-cc33-4011-9767-fa6049438fa8-kube-api-access-mmj6x\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738709 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/787e6909-959b-475a-a300-5ff751242c20-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738726 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738747 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/785d8daf-cc33-4011-9767-fa6049438fa8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738762 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/787e6909-959b-475a-a300-5ff751242c20-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738786 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.738813 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.740300 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.740436 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.740898 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/785d8daf-cc33-4011-9767-fa6049438fa8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.741818 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/785d8daf-cc33-4011-9767-fa6049438fa8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.743331 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/785d8daf-cc33-4011-9767-fa6049438fa8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.744105 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/785d8daf-cc33-4011-9767-fa6049438fa8-config-data\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.744349 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.747124 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/785d8daf-cc33-4011-9767-fa6049438fa8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.749501 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.749555 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-389bf769-1bfe-41e5-84a3-249075245d23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/97ca09a128e247017af88d0c60eefc68e785640cb893abdbcf3e382fbf328a1c/globalmount\"" pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.751354 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/785d8daf-cc33-4011-9767-fa6049438fa8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.760300 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmj6x\" (UniqueName: \"kubernetes.io/projected/785d8daf-cc33-4011-9767-fa6049438fa8-kube-api-access-mmj6x\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.780328 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-389bf769-1bfe-41e5-84a3-249075245d23\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-389bf769-1bfe-41e5-84a3-249075245d23\") pod \"rabbitmq-server-0\" (UID: \"785d8daf-cc33-4011-9767-fa6049438fa8\") " pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.841585 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.841679 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/787e6909-959b-475a-a300-5ff751242c20-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.841732 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/787e6909-959b-475a-a300-5ff751242c20-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.841781 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/787e6909-959b-475a-a300-5ff751242c20-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.841827 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.841879 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/787e6909-959b-475a-a300-5ff751242c20-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.841937 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.841998 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.842052 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/787e6909-959b-475a-a300-5ff751242c20-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.842115 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfl8w\" (UniqueName: \"kubernetes.io/projected/787e6909-959b-475a-a300-5ff751242c20-kube-api-access-pfl8w\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.842169 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.844341 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.845351 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.845777 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.845823 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/50ca25561a5c1786c943b437a4014dbd873252e15a6addc89cbb6d6bb81d4347/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.846140 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/787e6909-959b-475a-a300-5ff751242c20-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.848169 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/787e6909-959b-475a-a300-5ff751242c20-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.849179 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.849419 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/787e6909-959b-475a-a300-5ff751242c20-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.850447 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/787e6909-959b-475a-a300-5ff751242c20-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.852680 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/787e6909-959b-475a-a300-5ff751242c20-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.860311 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/787e6909-959b-475a-a300-5ff751242c20-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.872668 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfl8w\" (UniqueName: \"kubernetes.io/projected/787e6909-959b-475a-a300-5ff751242c20-kube-api-access-pfl8w\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.899824 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b56abf-86de-4ae8-aa2d-0f2f2d3293cc\") pod \"rabbitmq-cell1-server-0\" (UID: \"787e6909-959b-475a-a300-5ff751242c20\") " pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.913818 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 25 08:27:55 crc kubenswrapper[4978]: I0225 08:27:55.928656 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:27:56 crc kubenswrapper[4978]: W0225 08:27:56.220764 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod787e6909_959b_475a_a300_5ff751242c20.slice/crio-3c261ee07b93d07b5f2fc37d3acedb4d49caec6797a20704b85866fa9084afc8 WatchSource:0}: Error finding container 3c261ee07b93d07b5f2fc37d3acedb4d49caec6797a20704b85866fa9084afc8: Status 404 returned error can't find the container with id 3c261ee07b93d07b5f2fc37d3acedb4d49caec6797a20704b85866fa9084afc8 Feb 25 08:27:56 crc kubenswrapper[4978]: I0225 08:27:56.222276 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 25 08:27:56 crc kubenswrapper[4978]: I0225 08:27:56.373528 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 25 08:27:56 crc kubenswrapper[4978]: W0225 08:27:56.380195 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod785d8daf_cc33_4011_9767_fa6049438fa8.slice/crio-929fbb03a8e45920770cb9108c7d4178823d40152244559c8cd3694e71f19783 WatchSource:0}: Error finding container 929fbb03a8e45920770cb9108c7d4178823d40152244559c8cd3694e71f19783: Status 404 returned error can't find the container with id 929fbb03a8e45920770cb9108c7d4178823d40152244559c8cd3694e71f19783 Feb 25 08:27:56 crc kubenswrapper[4978]: I0225 08:27:56.506126 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"787e6909-959b-475a-a300-5ff751242c20","Type":"ContainerStarted","Data":"3c261ee07b93d07b5f2fc37d3acedb4d49caec6797a20704b85866fa9084afc8"} Feb 25 08:27:56 crc kubenswrapper[4978]: I0225 08:27:56.509487 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"785d8daf-cc33-4011-9767-fa6049438fa8","Type":"ContainerStarted","Data":"929fbb03a8e45920770cb9108c7d4178823d40152244559c8cd3694e71f19783"} Feb 25 08:27:57 crc kubenswrapper[4978]: I0225 08:27:57.345279 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de4db333-9150-4ccd-8f4e-d2a18917bbc5" path="/var/lib/kubelet/pods/de4db333-9150-4ccd-8f4e-d2a18917bbc5/volumes" Feb 25 08:27:57 crc kubenswrapper[4978]: I0225 08:27:57.348110 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4186a72-0d11-4647-902b-3908454bf0b1" path="/var/lib/kubelet/pods/e4186a72-0d11-4647-902b-3908454bf0b1/volumes" Feb 25 08:27:58 crc kubenswrapper[4978]: I0225 08:27:58.529996 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"787e6909-959b-475a-a300-5ff751242c20","Type":"ContainerStarted","Data":"b50510b3eb25499ff9c1a9110f7243b5fd5c08758454933e31397131806ebbab"} Feb 25 08:27:58 crc kubenswrapper[4978]: I0225 08:27:58.533732 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"785d8daf-cc33-4011-9767-fa6049438fa8","Type":"ContainerStarted","Data":"4980dd5ff664ad6f3250bd73dee1fc723cc4af3a7586ac0b79d5c927a1013b26"} Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.138235 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533468-z8znk"] Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.139094 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533468-z8znk" Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.140482 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.140803 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.141215 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.152062 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533468-z8znk"] Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.214143 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk7wl\" (UniqueName: \"kubernetes.io/projected/95b01cb1-795a-4462-b98c-669366afaa8a-kube-api-access-nk7wl\") pod \"auto-csr-approver-29533468-z8znk\" (UID: \"95b01cb1-795a-4462-b98c-669366afaa8a\") " pod="openshift-infra/auto-csr-approver-29533468-z8znk" Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.315886 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk7wl\" (UniqueName: \"kubernetes.io/projected/95b01cb1-795a-4462-b98c-669366afaa8a-kube-api-access-nk7wl\") pod \"auto-csr-approver-29533468-z8znk\" (UID: \"95b01cb1-795a-4462-b98c-669366afaa8a\") " pod="openshift-infra/auto-csr-approver-29533468-z8znk" Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.339953 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk7wl\" (UniqueName: \"kubernetes.io/projected/95b01cb1-795a-4462-b98c-669366afaa8a-kube-api-access-nk7wl\") pod \"auto-csr-approver-29533468-z8znk\" (UID: \"95b01cb1-795a-4462-b98c-669366afaa8a\") " pod="openshift-infra/auto-csr-approver-29533468-z8znk" Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.454538 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533468-z8znk" Feb 25 08:28:00 crc kubenswrapper[4978]: I0225 08:28:00.718123 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533468-z8znk"] Feb 25 08:28:01 crc kubenswrapper[4978]: I0225 08:28:01.563404 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533468-z8znk" event={"ID":"95b01cb1-795a-4462-b98c-669366afaa8a","Type":"ContainerStarted","Data":"40c6124d79c0bc1a8f8a9a910079fff13dacd195d0d6a1a95eacecbdc7e6af07"} Feb 25 08:28:02 crc kubenswrapper[4978]: I0225 08:28:02.572695 4978 generic.go:334] "Generic (PLEG): container finished" podID="95b01cb1-795a-4462-b98c-669366afaa8a" containerID="910915931f4de4652e60446aa0c6e462a3c51a2f57e5567145aa0df216734ca0" exitCode=0 Feb 25 08:28:02 crc kubenswrapper[4978]: I0225 08:28:02.572744 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533468-z8znk" event={"ID":"95b01cb1-795a-4462-b98c-669366afaa8a","Type":"ContainerDied","Data":"910915931f4de4652e60446aa0c6e462a3c51a2f57e5567145aa0df216734ca0"} Feb 25 08:28:03 crc kubenswrapper[4978]: I0225 08:28:03.978912 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533468-z8znk" Feb 25 08:28:04 crc kubenswrapper[4978]: I0225 08:28:04.080161 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk7wl\" (UniqueName: \"kubernetes.io/projected/95b01cb1-795a-4462-b98c-669366afaa8a-kube-api-access-nk7wl\") pod \"95b01cb1-795a-4462-b98c-669366afaa8a\" (UID: \"95b01cb1-795a-4462-b98c-669366afaa8a\") " Feb 25 08:28:04 crc kubenswrapper[4978]: I0225 08:28:04.089174 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95b01cb1-795a-4462-b98c-669366afaa8a-kube-api-access-nk7wl" (OuterVolumeSpecName: "kube-api-access-nk7wl") pod "95b01cb1-795a-4462-b98c-669366afaa8a" (UID: "95b01cb1-795a-4462-b98c-669366afaa8a"). InnerVolumeSpecName "kube-api-access-nk7wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:28:04 crc kubenswrapper[4978]: I0225 08:28:04.182802 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk7wl\" (UniqueName: \"kubernetes.io/projected/95b01cb1-795a-4462-b98c-669366afaa8a-kube-api-access-nk7wl\") on node \"crc\" DevicePath \"\"" Feb 25 08:28:04 crc kubenswrapper[4978]: I0225 08:28:04.609282 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533468-z8znk" event={"ID":"95b01cb1-795a-4462-b98c-669366afaa8a","Type":"ContainerDied","Data":"40c6124d79c0bc1a8f8a9a910079fff13dacd195d0d6a1a95eacecbdc7e6af07"} Feb 25 08:28:04 crc kubenswrapper[4978]: I0225 08:28:04.609358 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40c6124d79c0bc1a8f8a9a910079fff13dacd195d0d6a1a95eacecbdc7e6af07" Feb 25 08:28:04 crc kubenswrapper[4978]: I0225 08:28:04.609568 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533468-z8znk" Feb 25 08:28:05 crc kubenswrapper[4978]: I0225 08:28:05.078571 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533462-krmmf"] Feb 25 08:28:05 crc kubenswrapper[4978]: I0225 08:28:05.089576 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533462-krmmf"] Feb 25 08:28:05 crc kubenswrapper[4978]: I0225 08:28:05.343907 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e000858b-4d20-478f-ba95-0b7cedad86af" path="/var/lib/kubelet/pods/e000858b-4d20-478f-ba95-0b7cedad86af/volumes" Feb 25 08:28:16 crc kubenswrapper[4978]: I0225 08:28:16.540032 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:28:16 crc kubenswrapper[4978]: I0225 08:28:16.540907 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:28:26 crc kubenswrapper[4978]: I0225 08:28:26.334767 4978 scope.go:117] "RemoveContainer" containerID="1b4554bf179c6a667b09d97ee14377c48477c4b5a8ea767a09366545ceaf3df8" Feb 25 08:28:31 crc kubenswrapper[4978]: I0225 08:28:31.893868 4978 generic.go:334] "Generic (PLEG): container finished" podID="785d8daf-cc33-4011-9767-fa6049438fa8" containerID="4980dd5ff664ad6f3250bd73dee1fc723cc4af3a7586ac0b79d5c927a1013b26" exitCode=0 Feb 25 08:28:31 crc kubenswrapper[4978]: I0225 08:28:31.893998 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"785d8daf-cc33-4011-9767-fa6049438fa8","Type":"ContainerDied","Data":"4980dd5ff664ad6f3250bd73dee1fc723cc4af3a7586ac0b79d5c927a1013b26"} Feb 25 08:28:31 crc kubenswrapper[4978]: I0225 08:28:31.896111 4978 generic.go:334] "Generic (PLEG): container finished" podID="787e6909-959b-475a-a300-5ff751242c20" containerID="b50510b3eb25499ff9c1a9110f7243b5fd5c08758454933e31397131806ebbab" exitCode=0 Feb 25 08:28:31 crc kubenswrapper[4978]: I0225 08:28:31.896186 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"787e6909-959b-475a-a300-5ff751242c20","Type":"ContainerDied","Data":"b50510b3eb25499ff9c1a9110f7243b5fd5c08758454933e31397131806ebbab"} Feb 25 08:28:32 crc kubenswrapper[4978]: I0225 08:28:32.907563 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"787e6909-959b-475a-a300-5ff751242c20","Type":"ContainerStarted","Data":"0b6d334957eca53af1dda91fa9e25dc5c591c999128ee88cf39be971b6f69c59"} Feb 25 08:28:32 crc kubenswrapper[4978]: I0225 08:28:32.908181 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:28:32 crc kubenswrapper[4978]: I0225 08:28:32.909531 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"785d8daf-cc33-4011-9767-fa6049438fa8","Type":"ContainerStarted","Data":"52ab7890a6be5c8d51b2191950241654b8de18e719da5fb694e5589d6221ffb5"} Feb 25 08:28:32 crc kubenswrapper[4978]: I0225 08:28:32.909716 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 25 08:28:32 crc kubenswrapper[4978]: I0225 08:28:32.956548 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.956521758 podStartE2EDuration="37.956521758s" podCreationTimestamp="2026-02-25 08:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:28:32.952802734 +0000 UTC m=+6206.392059213" watchObservedRunningTime="2026-02-25 08:28:32.956521758 +0000 UTC m=+6206.395778217" Feb 25 08:28:32 crc kubenswrapper[4978]: I0225 08:28:32.959988 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.959977976 podStartE2EDuration="37.959977976s" podCreationTimestamp="2026-02-25 08:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:28:32.932991829 +0000 UTC m=+6206.372248308" watchObservedRunningTime="2026-02-25 08:28:32.959977976 +0000 UTC m=+6206.399234445" Feb 25 08:28:45 crc kubenswrapper[4978]: I0225 08:28:45.917694 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 25 08:28:45 crc kubenswrapper[4978]: I0225 08:28:45.932640 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 25 08:28:46 crc kubenswrapper[4978]: I0225 08:28:46.540556 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:28:46 crc kubenswrapper[4978]: I0225 08:28:46.540935 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:28:46 crc kubenswrapper[4978]: I0225 08:28:46.541003 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 08:28:46 crc kubenswrapper[4978]: I0225 08:28:46.541787 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae2d008bb6171404f42a195e3e45bb3e9fb1500e39baedfad501f642abfe36c7"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 08:28:46 crc kubenswrapper[4978]: I0225 08:28:46.541889 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://ae2d008bb6171404f42a195e3e45bb3e9fb1500e39baedfad501f642abfe36c7" gracePeriod=600 Feb 25 08:28:47 crc kubenswrapper[4978]: I0225 08:28:47.041633 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="ae2d008bb6171404f42a195e3e45bb3e9fb1500e39baedfad501f642abfe36c7" exitCode=0 Feb 25 08:28:47 crc kubenswrapper[4978]: I0225 08:28:47.041734 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"ae2d008bb6171404f42a195e3e45bb3e9fb1500e39baedfad501f642abfe36c7"} Feb 25 08:28:47 crc kubenswrapper[4978]: I0225 08:28:47.043766 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e"} Feb 25 08:28:47 crc kubenswrapper[4978]: I0225 08:28:47.043882 4978 scope.go:117] "RemoveContainer" containerID="244db60e39313929f85016ebb14756feda1ef26198cd379918b316328cff65a6" Feb 25 08:28:49 crc kubenswrapper[4978]: I0225 08:28:49.800077 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Feb 25 08:28:49 crc kubenswrapper[4978]: E0225 08:28:49.801207 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95b01cb1-795a-4462-b98c-669366afaa8a" containerName="oc" Feb 25 08:28:49 crc kubenswrapper[4978]: I0225 08:28:49.801230 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="95b01cb1-795a-4462-b98c-669366afaa8a" containerName="oc" Feb 25 08:28:49 crc kubenswrapper[4978]: I0225 08:28:49.801580 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="95b01cb1-795a-4462-b98c-669366afaa8a" containerName="oc" Feb 25 08:28:49 crc kubenswrapper[4978]: I0225 08:28:49.802297 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:28:49 crc kubenswrapper[4978]: I0225 08:28:49.806813 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fm969" Feb 25 08:28:49 crc kubenswrapper[4978]: I0225 08:28:49.815225 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:28:49 crc kubenswrapper[4978]: I0225 08:28:49.971936 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgf85\" (UniqueName: \"kubernetes.io/projected/68d0f022-f7ee-4335-848f-e847b80dde7c-kube-api-access-fgf85\") pod \"mariadb-client\" (UID: \"68d0f022-f7ee-4335-848f-e847b80dde7c\") " pod="openstack/mariadb-client" Feb 25 08:28:50 crc kubenswrapper[4978]: I0225 08:28:50.074244 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgf85\" (UniqueName: \"kubernetes.io/projected/68d0f022-f7ee-4335-848f-e847b80dde7c-kube-api-access-fgf85\") pod \"mariadb-client\" (UID: \"68d0f022-f7ee-4335-848f-e847b80dde7c\") " pod="openstack/mariadb-client" Feb 25 08:28:50 crc kubenswrapper[4978]: I0225 08:28:50.099890 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgf85\" (UniqueName: \"kubernetes.io/projected/68d0f022-f7ee-4335-848f-e847b80dde7c-kube-api-access-fgf85\") pod \"mariadb-client\" (UID: \"68d0f022-f7ee-4335-848f-e847b80dde7c\") " pod="openstack/mariadb-client" Feb 25 08:28:50 crc kubenswrapper[4978]: I0225 08:28:50.139038 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:28:50 crc kubenswrapper[4978]: I0225 08:28:50.445553 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:28:50 crc kubenswrapper[4978]: W0225 08:28:50.452770 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68d0f022_f7ee_4335_848f_e847b80dde7c.slice/crio-5ce926af66d79c2dd050df176f6944680a77bcd7924999813d3832f67940b83b WatchSource:0}: Error finding container 5ce926af66d79c2dd050df176f6944680a77bcd7924999813d3832f67940b83b: Status 404 returned error can't find the container with id 5ce926af66d79c2dd050df176f6944680a77bcd7924999813d3832f67940b83b Feb 25 08:28:51 crc kubenswrapper[4978]: I0225 08:28:51.090955 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"68d0f022-f7ee-4335-848f-e847b80dde7c","Type":"ContainerStarted","Data":"5ce926af66d79c2dd050df176f6944680a77bcd7924999813d3832f67940b83b"} Feb 25 08:28:55 crc kubenswrapper[4978]: I0225 08:28:55.125730 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"68d0f022-f7ee-4335-848f-e847b80dde7c","Type":"ContainerStarted","Data":"3bd1cba932762ab1c570c971012a01f4bab2f499e224367014b914a22b101d8d"} Feb 25 08:28:55 crc kubenswrapper[4978]: I0225 08:28:55.148902 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client" podStartSLOduration=2.002611625 podStartE2EDuration="6.148882115s" podCreationTimestamp="2026-02-25 08:28:49 +0000 UTC" firstStartedPulling="2026-02-25 08:28:50.454116493 +0000 UTC m=+6223.893372952" lastFinishedPulling="2026-02-25 08:28:54.600386973 +0000 UTC m=+6228.039643442" observedRunningTime="2026-02-25 08:28:55.144898252 +0000 UTC m=+6228.584154711" watchObservedRunningTime="2026-02-25 08:28:55.148882115 +0000 UTC m=+6228.588138574" Feb 25 08:28:56 crc kubenswrapper[4978]: I0225 08:28:56.806161 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gqx9p"] Feb 25 08:28:56 crc kubenswrapper[4978]: I0225 08:28:56.808641 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:56 crc kubenswrapper[4978]: I0225 08:28:56.822284 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gqx9p"] Feb 25 08:28:56 crc kubenswrapper[4978]: I0225 08:28:56.898735 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-utilities\") pod \"community-operators-gqx9p\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:56 crc kubenswrapper[4978]: I0225 08:28:56.898961 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-catalog-content\") pod \"community-operators-gqx9p\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:56 crc kubenswrapper[4978]: I0225 08:28:56.898997 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9thg7\" (UniqueName: \"kubernetes.io/projected/77eb447f-6ccd-48c7-8466-bb54b4a2d751-kube-api-access-9thg7\") pod \"community-operators-gqx9p\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:57 crc kubenswrapper[4978]: I0225 08:28:57.000328 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-utilities\") pod \"community-operators-gqx9p\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:57 crc kubenswrapper[4978]: I0225 08:28:57.000521 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-catalog-content\") pod \"community-operators-gqx9p\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:57 crc kubenswrapper[4978]: I0225 08:28:57.000571 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9thg7\" (UniqueName: \"kubernetes.io/projected/77eb447f-6ccd-48c7-8466-bb54b4a2d751-kube-api-access-9thg7\") pod \"community-operators-gqx9p\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:57 crc kubenswrapper[4978]: I0225 08:28:57.000884 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-utilities\") pod \"community-operators-gqx9p\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:57 crc kubenswrapper[4978]: I0225 08:28:57.001142 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-catalog-content\") pod \"community-operators-gqx9p\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:57 crc kubenswrapper[4978]: I0225 08:28:57.031110 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9thg7\" (UniqueName: \"kubernetes.io/projected/77eb447f-6ccd-48c7-8466-bb54b4a2d751-kube-api-access-9thg7\") pod \"community-operators-gqx9p\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:57 crc kubenswrapper[4978]: I0225 08:28:57.127811 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:28:57 crc kubenswrapper[4978]: I0225 08:28:57.576904 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gqx9p"] Feb 25 08:28:58 crc kubenswrapper[4978]: I0225 08:28:58.153776 4978 generic.go:334] "Generic (PLEG): container finished" podID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerID="dd199648034bbc4769b8e6a64c8c3a847c7987fcf46dc08ac6604ea45d06c954" exitCode=0 Feb 25 08:28:58 crc kubenswrapper[4978]: I0225 08:28:58.154131 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqx9p" event={"ID":"77eb447f-6ccd-48c7-8466-bb54b4a2d751","Type":"ContainerDied","Data":"dd199648034bbc4769b8e6a64c8c3a847c7987fcf46dc08ac6604ea45d06c954"} Feb 25 08:28:58 crc kubenswrapper[4978]: I0225 08:28:58.154171 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqx9p" event={"ID":"77eb447f-6ccd-48c7-8466-bb54b4a2d751","Type":"ContainerStarted","Data":"431e63793b8c46bc292f533680b62d6523da13b1e863d35bde6b0039b868f71b"} Feb 25 08:28:59 crc kubenswrapper[4978]: I0225 08:28:59.165173 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqx9p" event={"ID":"77eb447f-6ccd-48c7-8466-bb54b4a2d751","Type":"ContainerStarted","Data":"ab9d8eaab221ee43ece760a27f1f16a4fab062d3e26fdb94e566c007c755c024"} Feb 25 08:29:00 crc kubenswrapper[4978]: I0225 08:29:00.178315 4978 generic.go:334] "Generic (PLEG): container finished" podID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerID="ab9d8eaab221ee43ece760a27f1f16a4fab062d3e26fdb94e566c007c755c024" exitCode=0 Feb 25 08:29:00 crc kubenswrapper[4978]: I0225 08:29:00.178427 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqx9p" event={"ID":"77eb447f-6ccd-48c7-8466-bb54b4a2d751","Type":"ContainerDied","Data":"ab9d8eaab221ee43ece760a27f1f16a4fab062d3e26fdb94e566c007c755c024"} Feb 25 08:29:01 crc kubenswrapper[4978]: I0225 08:29:01.188670 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqx9p" event={"ID":"77eb447f-6ccd-48c7-8466-bb54b4a2d751","Type":"ContainerStarted","Data":"bdd76187cc142cf17c14b03ab13b90d311da93f39b36ae81c90fc1d34178013f"} Feb 25 08:29:01 crc kubenswrapper[4978]: I0225 08:29:01.210723 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gqx9p" podStartSLOduration=2.803623698 podStartE2EDuration="5.210699985s" podCreationTimestamp="2026-02-25 08:28:56 +0000 UTC" firstStartedPulling="2026-02-25 08:28:58.158574872 +0000 UTC m=+6231.597831361" lastFinishedPulling="2026-02-25 08:29:00.565651179 +0000 UTC m=+6234.004907648" observedRunningTime="2026-02-25 08:29:01.207875087 +0000 UTC m=+6234.647131576" watchObservedRunningTime="2026-02-25 08:29:01.210699985 +0000 UTC m=+6234.649956454" Feb 25 08:29:07 crc kubenswrapper[4978]: I0225 08:29:07.128758 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:29:07 crc kubenswrapper[4978]: I0225 08:29:07.129124 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:29:07 crc kubenswrapper[4978]: I0225 08:29:07.195180 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:29:07 crc kubenswrapper[4978]: I0225 08:29:07.308791 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:29:09 crc kubenswrapper[4978]: I0225 08:29:09.803448 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gqx9p"] Feb 25 08:29:09 crc kubenswrapper[4978]: I0225 08:29:09.804186 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gqx9p" podUID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerName="registry-server" containerID="cri-o://bdd76187cc142cf17c14b03ab13b90d311da93f39b36ae81c90fc1d34178013f" gracePeriod=2 Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.289507 4978 generic.go:334] "Generic (PLEG): container finished" podID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerID="bdd76187cc142cf17c14b03ab13b90d311da93f39b36ae81c90fc1d34178013f" exitCode=0 Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.289614 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqx9p" event={"ID":"77eb447f-6ccd-48c7-8466-bb54b4a2d751","Type":"ContainerDied","Data":"bdd76187cc142cf17c14b03ab13b90d311da93f39b36ae81c90fc1d34178013f"} Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.719185 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.837334 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-catalog-content\") pod \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.837471 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9thg7\" (UniqueName: \"kubernetes.io/projected/77eb447f-6ccd-48c7-8466-bb54b4a2d751-kube-api-access-9thg7\") pod \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.837608 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-utilities\") pod \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\" (UID: \"77eb447f-6ccd-48c7-8466-bb54b4a2d751\") " Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.838663 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-utilities" (OuterVolumeSpecName: "utilities") pod "77eb447f-6ccd-48c7-8466-bb54b4a2d751" (UID: "77eb447f-6ccd-48c7-8466-bb54b4a2d751"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.845815 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77eb447f-6ccd-48c7-8466-bb54b4a2d751-kube-api-access-9thg7" (OuterVolumeSpecName: "kube-api-access-9thg7") pod "77eb447f-6ccd-48c7-8466-bb54b4a2d751" (UID: "77eb447f-6ccd-48c7-8466-bb54b4a2d751"). InnerVolumeSpecName "kube-api-access-9thg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.898573 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77eb447f-6ccd-48c7-8466-bb54b4a2d751" (UID: "77eb447f-6ccd-48c7-8466-bb54b4a2d751"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.939721 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.939766 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77eb447f-6ccd-48c7-8466-bb54b4a2d751-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:29:10 crc kubenswrapper[4978]: I0225 08:29:10.939800 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9thg7\" (UniqueName: \"kubernetes.io/projected/77eb447f-6ccd-48c7-8466-bb54b4a2d751-kube-api-access-9thg7\") on node \"crc\" DevicePath \"\"" Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.098310 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.098512 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-client" podUID="68d0f022-f7ee-4335-848f-e847b80dde7c" containerName="mariadb-client" containerID="cri-o://3bd1cba932762ab1c570c971012a01f4bab2f499e224367014b914a22b101d8d" gracePeriod=30 Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.301721 4978 generic.go:334] "Generic (PLEG): container finished" podID="68d0f022-f7ee-4335-848f-e847b80dde7c" containerID="3bd1cba932762ab1c570c971012a01f4bab2f499e224367014b914a22b101d8d" exitCode=143 Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.301869 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"68d0f022-f7ee-4335-848f-e847b80dde7c","Type":"ContainerDied","Data":"3bd1cba932762ab1c570c971012a01f4bab2f499e224367014b914a22b101d8d"} Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.307022 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gqx9p" event={"ID":"77eb447f-6ccd-48c7-8466-bb54b4a2d751","Type":"ContainerDied","Data":"431e63793b8c46bc292f533680b62d6523da13b1e863d35bde6b0039b868f71b"} Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.307103 4978 scope.go:117] "RemoveContainer" containerID="bdd76187cc142cf17c14b03ab13b90d311da93f39b36ae81c90fc1d34178013f" Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.307130 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gqx9p" Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.332182 4978 scope.go:117] "RemoveContainer" containerID="ab9d8eaab221ee43ece760a27f1f16a4fab062d3e26fdb94e566c007c755c024" Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.358797 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gqx9p"] Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.372176 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gqx9p"] Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.386540 4978 scope.go:117] "RemoveContainer" containerID="dd199648034bbc4769b8e6a64c8c3a847c7987fcf46dc08ac6604ea45d06c954" Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.659510 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.753055 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgf85\" (UniqueName: \"kubernetes.io/projected/68d0f022-f7ee-4335-848f-e847b80dde7c-kube-api-access-fgf85\") pod \"68d0f022-f7ee-4335-848f-e847b80dde7c\" (UID: \"68d0f022-f7ee-4335-848f-e847b80dde7c\") " Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.761635 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d0f022-f7ee-4335-848f-e847b80dde7c-kube-api-access-fgf85" (OuterVolumeSpecName: "kube-api-access-fgf85") pod "68d0f022-f7ee-4335-848f-e847b80dde7c" (UID: "68d0f022-f7ee-4335-848f-e847b80dde7c"). InnerVolumeSpecName "kube-api-access-fgf85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:29:11 crc kubenswrapper[4978]: I0225 08:29:11.854775 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgf85\" (UniqueName: \"kubernetes.io/projected/68d0f022-f7ee-4335-848f-e847b80dde7c-kube-api-access-fgf85\") on node \"crc\" DevicePath \"\"" Feb 25 08:29:12 crc kubenswrapper[4978]: I0225 08:29:12.319952 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"68d0f022-f7ee-4335-848f-e847b80dde7c","Type":"ContainerDied","Data":"5ce926af66d79c2dd050df176f6944680a77bcd7924999813d3832f67940b83b"} Feb 25 08:29:12 crc kubenswrapper[4978]: I0225 08:29:12.320037 4978 scope.go:117] "RemoveContainer" containerID="3bd1cba932762ab1c570c971012a01f4bab2f499e224367014b914a22b101d8d" Feb 25 08:29:12 crc kubenswrapper[4978]: I0225 08:29:12.320208 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:29:12 crc kubenswrapper[4978]: I0225 08:29:12.375905 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:29:12 crc kubenswrapper[4978]: I0225 08:29:12.386091 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:29:13 crc kubenswrapper[4978]: I0225 08:29:13.340167 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68d0f022-f7ee-4335-848f-e847b80dde7c" path="/var/lib/kubelet/pods/68d0f022-f7ee-4335-848f-e847b80dde7c/volumes" Feb 25 08:29:13 crc kubenswrapper[4978]: I0225 08:29:13.341749 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" path="/var/lib/kubelet/pods/77eb447f-6ccd-48c7-8466-bb54b4a2d751/volumes" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.171086 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533470-5cs55"] Feb 25 08:30:00 crc kubenswrapper[4978]: E0225 08:30:00.172045 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerName="extract-content" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.172064 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerName="extract-content" Feb 25 08:30:00 crc kubenswrapper[4978]: E0225 08:30:00.172084 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d0f022-f7ee-4335-848f-e847b80dde7c" containerName="mariadb-client" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.172092 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d0f022-f7ee-4335-848f-e847b80dde7c" containerName="mariadb-client" Feb 25 08:30:00 crc kubenswrapper[4978]: E0225 08:30:00.172121 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerName="extract-utilities" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.172131 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerName="extract-utilities" Feb 25 08:30:00 crc kubenswrapper[4978]: E0225 08:30:00.172149 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerName="registry-server" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.172157 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerName="registry-server" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.172350 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d0f022-f7ee-4335-848f-e847b80dde7c" containerName="mariadb-client" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.172387 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="77eb447f-6ccd-48c7-8466-bb54b4a2d751" containerName="registry-server" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.173009 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533470-5cs55" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.175238 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.178801 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.179020 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.182699 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p"] Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.183717 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.187968 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.188131 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.193262 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p"] Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.200063 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533470-5cs55"] Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.256281 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59fb3f4d-8672-48ea-946e-5b768b1cb62a-config-volume\") pod \"collect-profiles-29533470-9w64p\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.256481 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjcs8\" (UniqueName: \"kubernetes.io/projected/a285f988-f118-4b73-aaeb-d1f36a843d85-kube-api-access-hjcs8\") pod \"auto-csr-approver-29533470-5cs55\" (UID: \"a285f988-f118-4b73-aaeb-d1f36a843d85\") " pod="openshift-infra/auto-csr-approver-29533470-5cs55" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.256595 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59fb3f4d-8672-48ea-946e-5b768b1cb62a-secret-volume\") pod \"collect-profiles-29533470-9w64p\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.256680 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnnjc\" (UniqueName: \"kubernetes.io/projected/59fb3f4d-8672-48ea-946e-5b768b1cb62a-kube-api-access-nnnjc\") pod \"collect-profiles-29533470-9w64p\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.357848 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59fb3f4d-8672-48ea-946e-5b768b1cb62a-config-volume\") pod \"collect-profiles-29533470-9w64p\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.358530 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjcs8\" (UniqueName: \"kubernetes.io/projected/a285f988-f118-4b73-aaeb-d1f36a843d85-kube-api-access-hjcs8\") pod \"auto-csr-approver-29533470-5cs55\" (UID: \"a285f988-f118-4b73-aaeb-d1f36a843d85\") " pod="openshift-infra/auto-csr-approver-29533470-5cs55" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.358588 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59fb3f4d-8672-48ea-946e-5b768b1cb62a-secret-volume\") pod \"collect-profiles-29533470-9w64p\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.358620 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnnjc\" (UniqueName: \"kubernetes.io/projected/59fb3f4d-8672-48ea-946e-5b768b1cb62a-kube-api-access-nnnjc\") pod \"collect-profiles-29533470-9w64p\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.359100 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59fb3f4d-8672-48ea-946e-5b768b1cb62a-config-volume\") pod \"collect-profiles-29533470-9w64p\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.363942 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59fb3f4d-8672-48ea-946e-5b768b1cb62a-secret-volume\") pod \"collect-profiles-29533470-9w64p\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.376666 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjcs8\" (UniqueName: \"kubernetes.io/projected/a285f988-f118-4b73-aaeb-d1f36a843d85-kube-api-access-hjcs8\") pod \"auto-csr-approver-29533470-5cs55\" (UID: \"a285f988-f118-4b73-aaeb-d1f36a843d85\") " pod="openshift-infra/auto-csr-approver-29533470-5cs55" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.384772 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnnjc\" (UniqueName: \"kubernetes.io/projected/59fb3f4d-8672-48ea-946e-5b768b1cb62a-kube-api-access-nnnjc\") pod \"collect-profiles-29533470-9w64p\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.540545 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533470-5cs55" Feb 25 08:30:00 crc kubenswrapper[4978]: I0225 08:30:00.553846 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:01 crc kubenswrapper[4978]: I0225 08:30:01.039010 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533470-5cs55"] Feb 25 08:30:01 crc kubenswrapper[4978]: I0225 08:30:01.100422 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p"] Feb 25 08:30:01 crc kubenswrapper[4978]: W0225 08:30:01.105748 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59fb3f4d_8672_48ea_946e_5b768b1cb62a.slice/crio-5b494110ce52977b6c756660e06d74c504c59763e8cd1a3f25dd094d10443abd WatchSource:0}: Error finding container 5b494110ce52977b6c756660e06d74c504c59763e8cd1a3f25dd094d10443abd: Status 404 returned error can't find the container with id 5b494110ce52977b6c756660e06d74c504c59763e8cd1a3f25dd094d10443abd Feb 25 08:30:01 crc kubenswrapper[4978]: I0225 08:30:01.812468 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533470-5cs55" event={"ID":"a285f988-f118-4b73-aaeb-d1f36a843d85","Type":"ContainerStarted","Data":"d58253107d749f353f9cf14313bf71cccee2d36fd5420c6b2d8c348ed80d09b7"} Feb 25 08:30:01 crc kubenswrapper[4978]: I0225 08:30:01.814810 4978 generic.go:334] "Generic (PLEG): container finished" podID="59fb3f4d-8672-48ea-946e-5b768b1cb62a" containerID="80cb84fc321be9bdaf63738d7cd737a060a425c1bc6c7009a312571b625f4536" exitCode=0 Feb 25 08:30:01 crc kubenswrapper[4978]: I0225 08:30:01.814887 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" event={"ID":"59fb3f4d-8672-48ea-946e-5b768b1cb62a","Type":"ContainerDied","Data":"80cb84fc321be9bdaf63738d7cd737a060a425c1bc6c7009a312571b625f4536"} Feb 25 08:30:01 crc kubenswrapper[4978]: I0225 08:30:01.815184 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" event={"ID":"59fb3f4d-8672-48ea-946e-5b768b1cb62a","Type":"ContainerStarted","Data":"5b494110ce52977b6c756660e06d74c504c59763e8cd1a3f25dd094d10443abd"} Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.189072 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.334717 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59fb3f4d-8672-48ea-946e-5b768b1cb62a-secret-volume\") pod \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.334853 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59fb3f4d-8672-48ea-946e-5b768b1cb62a-config-volume\") pod \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.334896 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnnjc\" (UniqueName: \"kubernetes.io/projected/59fb3f4d-8672-48ea-946e-5b768b1cb62a-kube-api-access-nnnjc\") pod \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\" (UID: \"59fb3f4d-8672-48ea-946e-5b768b1cb62a\") " Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.336971 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59fb3f4d-8672-48ea-946e-5b768b1cb62a-config-volume" (OuterVolumeSpecName: "config-volume") pod "59fb3f4d-8672-48ea-946e-5b768b1cb62a" (UID: "59fb3f4d-8672-48ea-946e-5b768b1cb62a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.342742 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59fb3f4d-8672-48ea-946e-5b768b1cb62a-kube-api-access-nnnjc" (OuterVolumeSpecName: "kube-api-access-nnnjc") pod "59fb3f4d-8672-48ea-946e-5b768b1cb62a" (UID: "59fb3f4d-8672-48ea-946e-5b768b1cb62a"). InnerVolumeSpecName "kube-api-access-nnnjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.346669 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59fb3f4d-8672-48ea-946e-5b768b1cb62a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "59fb3f4d-8672-48ea-946e-5b768b1cb62a" (UID: "59fb3f4d-8672-48ea-946e-5b768b1cb62a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.437039 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59fb3f4d-8672-48ea-946e-5b768b1cb62a-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.437071 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59fb3f4d-8672-48ea-946e-5b768b1cb62a-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.437086 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnnjc\" (UniqueName: \"kubernetes.io/projected/59fb3f4d-8672-48ea-946e-5b768b1cb62a-kube-api-access-nnnjc\") on node \"crc\" DevicePath \"\"" Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.840006 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" event={"ID":"59fb3f4d-8672-48ea-946e-5b768b1cb62a","Type":"ContainerDied","Data":"5b494110ce52977b6c756660e06d74c504c59763e8cd1a3f25dd094d10443abd"} Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.840045 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b494110ce52977b6c756660e06d74c504c59763e8cd1a3f25dd094d10443abd" Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.840094 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p" Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.843093 4978 generic.go:334] "Generic (PLEG): container finished" podID="a285f988-f118-4b73-aaeb-d1f36a843d85" containerID="3eea4b2ca29e6aa79cb0f7100898c21e0efced1fe6f69485bfee3cd9f0e073d9" exitCode=0 Feb 25 08:30:03 crc kubenswrapper[4978]: I0225 08:30:03.843158 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533470-5cs55" event={"ID":"a285f988-f118-4b73-aaeb-d1f36a843d85","Type":"ContainerDied","Data":"3eea4b2ca29e6aa79cb0f7100898c21e0efced1fe6f69485bfee3cd9f0e073d9"} Feb 25 08:30:04 crc kubenswrapper[4978]: I0225 08:30:04.275077 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp"] Feb 25 08:30:04 crc kubenswrapper[4978]: I0225 08:30:04.281970 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533425-4d6zp"] Feb 25 08:30:05 crc kubenswrapper[4978]: I0225 08:30:05.255529 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533470-5cs55" Feb 25 08:30:05 crc kubenswrapper[4978]: I0225 08:30:05.349775 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a" path="/var/lib/kubelet/pods/c0295eb9-4a00-4bb7-9eb0-b8c35e9a506a/volumes" Feb 25 08:30:05 crc kubenswrapper[4978]: I0225 08:30:05.371847 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjcs8\" (UniqueName: \"kubernetes.io/projected/a285f988-f118-4b73-aaeb-d1f36a843d85-kube-api-access-hjcs8\") pod \"a285f988-f118-4b73-aaeb-d1f36a843d85\" (UID: \"a285f988-f118-4b73-aaeb-d1f36a843d85\") " Feb 25 08:30:05 crc kubenswrapper[4978]: I0225 08:30:05.380743 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a285f988-f118-4b73-aaeb-d1f36a843d85-kube-api-access-hjcs8" (OuterVolumeSpecName: "kube-api-access-hjcs8") pod "a285f988-f118-4b73-aaeb-d1f36a843d85" (UID: "a285f988-f118-4b73-aaeb-d1f36a843d85"). InnerVolumeSpecName "kube-api-access-hjcs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:30:05 crc kubenswrapper[4978]: I0225 08:30:05.473158 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjcs8\" (UniqueName: \"kubernetes.io/projected/a285f988-f118-4b73-aaeb-d1f36a843d85-kube-api-access-hjcs8\") on node \"crc\" DevicePath \"\"" Feb 25 08:30:05 crc kubenswrapper[4978]: I0225 08:30:05.866111 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533470-5cs55" event={"ID":"a285f988-f118-4b73-aaeb-d1f36a843d85","Type":"ContainerDied","Data":"d58253107d749f353f9cf14313bf71cccee2d36fd5420c6b2d8c348ed80d09b7"} Feb 25 08:30:05 crc kubenswrapper[4978]: I0225 08:30:05.866175 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d58253107d749f353f9cf14313bf71cccee2d36fd5420c6b2d8c348ed80d09b7" Feb 25 08:30:05 crc kubenswrapper[4978]: I0225 08:30:05.866171 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533470-5cs55" Feb 25 08:30:06 crc kubenswrapper[4978]: I0225 08:30:06.333440 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533464-kpfv7"] Feb 25 08:30:06 crc kubenswrapper[4978]: I0225 08:30:06.342262 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533464-kpfv7"] Feb 25 08:30:07 crc kubenswrapper[4978]: I0225 08:30:07.345759 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f99497f-94dc-4976-96c9-60eba8b5c535" path="/var/lib/kubelet/pods/1f99497f-94dc-4976-96c9-60eba8b5c535/volumes" Feb 25 08:30:26 crc kubenswrapper[4978]: I0225 08:30:26.584161 4978 scope.go:117] "RemoveContainer" containerID="bd179ff7d61dd9b40b85a2b223bc598ac1a749ff1926d9aae5647a98bc03c808" Feb 25 08:30:26 crc kubenswrapper[4978]: I0225 08:30:26.626614 4978 scope.go:117] "RemoveContainer" containerID="6086e2d7bd33ed66ba3c52852b533b074ebe8596f06e09263102928f250cdd62" Feb 25 08:30:26 crc kubenswrapper[4978]: I0225 08:30:26.709933 4978 scope.go:117] "RemoveContainer" containerID="6588bfe6ec137758f126ff6eaaa355f2fc9cf342f3cf6887768a3e31fe853fa1" Feb 25 08:30:46 crc kubenswrapper[4978]: I0225 08:30:46.540635 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:30:46 crc kubenswrapper[4978]: I0225 08:30:46.541238 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:31:16 crc kubenswrapper[4978]: I0225 08:31:16.540739 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:31:16 crc kubenswrapper[4978]: I0225 08:31:16.541350 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:31:46 crc kubenswrapper[4978]: I0225 08:31:46.540925 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:31:46 crc kubenswrapper[4978]: I0225 08:31:46.541685 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:31:46 crc kubenswrapper[4978]: I0225 08:31:46.541750 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 08:31:46 crc kubenswrapper[4978]: I0225 08:31:46.542608 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 08:31:46 crc kubenswrapper[4978]: I0225 08:31:46.542711 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" gracePeriod=600 Feb 25 08:31:46 crc kubenswrapper[4978]: E0225 08:31:46.673619 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:31:46 crc kubenswrapper[4978]: I0225 08:31:46.864816 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" exitCode=0 Feb 25 08:31:46 crc kubenswrapper[4978]: I0225 08:31:46.864882 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e"} Feb 25 08:31:46 crc kubenswrapper[4978]: I0225 08:31:46.864933 4978 scope.go:117] "RemoveContainer" containerID="ae2d008bb6171404f42a195e3e45bb3e9fb1500e39baedfad501f642abfe36c7" Feb 25 08:31:46 crc kubenswrapper[4978]: I0225 08:31:46.865847 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:31:46 crc kubenswrapper[4978]: E0225 08:31:46.866326 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.168995 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533472-8c5nh"] Feb 25 08:32:00 crc kubenswrapper[4978]: E0225 08:32:00.170302 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a285f988-f118-4b73-aaeb-d1f36a843d85" containerName="oc" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.170330 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a285f988-f118-4b73-aaeb-d1f36a843d85" containerName="oc" Feb 25 08:32:00 crc kubenswrapper[4978]: E0225 08:32:00.170460 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59fb3f4d-8672-48ea-946e-5b768b1cb62a" containerName="collect-profiles" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.170488 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="59fb3f4d-8672-48ea-946e-5b768b1cb62a" containerName="collect-profiles" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.170755 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a285f988-f118-4b73-aaeb-d1f36a843d85" containerName="oc" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.170793 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="59fb3f4d-8672-48ea-946e-5b768b1cb62a" containerName="collect-profiles" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.171690 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533472-8c5nh" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.174873 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.175661 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.176581 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.179474 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533472-8c5nh"] Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.249990 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpnqp\" (UniqueName: \"kubernetes.io/projected/1999d0d3-0e65-45ea-8133-f6b40154c125-kube-api-access-tpnqp\") pod \"auto-csr-approver-29533472-8c5nh\" (UID: \"1999d0d3-0e65-45ea-8133-f6b40154c125\") " pod="openshift-infra/auto-csr-approver-29533472-8c5nh" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.351188 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpnqp\" (UniqueName: \"kubernetes.io/projected/1999d0d3-0e65-45ea-8133-f6b40154c125-kube-api-access-tpnqp\") pod \"auto-csr-approver-29533472-8c5nh\" (UID: \"1999d0d3-0e65-45ea-8133-f6b40154c125\") " pod="openshift-infra/auto-csr-approver-29533472-8c5nh" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.368985 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpnqp\" (UniqueName: \"kubernetes.io/projected/1999d0d3-0e65-45ea-8133-f6b40154c125-kube-api-access-tpnqp\") pod \"auto-csr-approver-29533472-8c5nh\" (UID: \"1999d0d3-0e65-45ea-8133-f6b40154c125\") " pod="openshift-infra/auto-csr-approver-29533472-8c5nh" Feb 25 08:32:00 crc kubenswrapper[4978]: I0225 08:32:00.498736 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533472-8c5nh" Feb 25 08:32:01 crc kubenswrapper[4978]: I0225 08:32:01.011229 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533472-8c5nh"] Feb 25 08:32:02 crc kubenswrapper[4978]: I0225 08:32:02.030492 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533472-8c5nh" event={"ID":"1999d0d3-0e65-45ea-8133-f6b40154c125","Type":"ContainerStarted","Data":"a3bbd0c610b35fc80d19c89393a17311552242bb18704a0bd8ee49cdaea09ea1"} Feb 25 08:32:02 crc kubenswrapper[4978]: I0225 08:32:02.328258 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:32:02 crc kubenswrapper[4978]: E0225 08:32:02.328560 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:32:03 crc kubenswrapper[4978]: I0225 08:32:03.041734 4978 generic.go:334] "Generic (PLEG): container finished" podID="1999d0d3-0e65-45ea-8133-f6b40154c125" containerID="b6d2f778106fa1a2a105312237f63ef0c91f2cd87fa098e690c33f0a3335cc52" exitCode=0 Feb 25 08:32:03 crc kubenswrapper[4978]: I0225 08:32:03.041786 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533472-8c5nh" event={"ID":"1999d0d3-0e65-45ea-8133-f6b40154c125","Type":"ContainerDied","Data":"b6d2f778106fa1a2a105312237f63ef0c91f2cd87fa098e690c33f0a3335cc52"} Feb 25 08:32:04 crc kubenswrapper[4978]: I0225 08:32:04.451037 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533472-8c5nh" Feb 25 08:32:04 crc kubenswrapper[4978]: I0225 08:32:04.519547 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpnqp\" (UniqueName: \"kubernetes.io/projected/1999d0d3-0e65-45ea-8133-f6b40154c125-kube-api-access-tpnqp\") pod \"1999d0d3-0e65-45ea-8133-f6b40154c125\" (UID: \"1999d0d3-0e65-45ea-8133-f6b40154c125\") " Feb 25 08:32:04 crc kubenswrapper[4978]: I0225 08:32:04.528800 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1999d0d3-0e65-45ea-8133-f6b40154c125-kube-api-access-tpnqp" (OuterVolumeSpecName: "kube-api-access-tpnqp") pod "1999d0d3-0e65-45ea-8133-f6b40154c125" (UID: "1999d0d3-0e65-45ea-8133-f6b40154c125"). InnerVolumeSpecName "kube-api-access-tpnqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:32:04 crc kubenswrapper[4978]: I0225 08:32:04.621225 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpnqp\" (UniqueName: \"kubernetes.io/projected/1999d0d3-0e65-45ea-8133-f6b40154c125-kube-api-access-tpnqp\") on node \"crc\" DevicePath \"\"" Feb 25 08:32:05 crc kubenswrapper[4978]: I0225 08:32:05.066866 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533472-8c5nh" event={"ID":"1999d0d3-0e65-45ea-8133-f6b40154c125","Type":"ContainerDied","Data":"a3bbd0c610b35fc80d19c89393a17311552242bb18704a0bd8ee49cdaea09ea1"} Feb 25 08:32:05 crc kubenswrapper[4978]: I0225 08:32:05.066936 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3bbd0c610b35fc80d19c89393a17311552242bb18704a0bd8ee49cdaea09ea1" Feb 25 08:32:05 crc kubenswrapper[4978]: I0225 08:32:05.066999 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533472-8c5nh" Feb 25 08:32:05 crc kubenswrapper[4978]: I0225 08:32:05.535807 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533466-mrwn2"] Feb 25 08:32:05 crc kubenswrapper[4978]: I0225 08:32:05.544758 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533466-mrwn2"] Feb 25 08:32:07 crc kubenswrapper[4978]: I0225 08:32:07.347091 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df3eae9-8600-47fc-906e-3a0adf79ad2d" path="/var/lib/kubelet/pods/7df3eae9-8600-47fc-906e-3a0adf79ad2d/volumes" Feb 25 08:32:13 crc kubenswrapper[4978]: I0225 08:32:13.327884 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:32:13 crc kubenswrapper[4978]: E0225 08:32:13.328956 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:32:24 crc kubenswrapper[4978]: I0225 08:32:24.328043 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:32:24 crc kubenswrapper[4978]: E0225 08:32:24.329514 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:32:26 crc kubenswrapper[4978]: I0225 08:32:26.828061 4978 scope.go:117] "RemoveContainer" containerID="ffb5b65fe7df9dd275efa4a8e89dfe64e8a904a82ce5e1ac3ecef7ecf13320da" Feb 25 08:32:35 crc kubenswrapper[4978]: I0225 08:32:35.328837 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:32:35 crc kubenswrapper[4978]: E0225 08:32:35.329532 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:32:49 crc kubenswrapper[4978]: I0225 08:32:49.327990 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:32:49 crc kubenswrapper[4978]: E0225 08:32:49.330791 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:33:03 crc kubenswrapper[4978]: I0225 08:33:03.328394 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:33:03 crc kubenswrapper[4978]: E0225 08:33:03.329114 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:33:17 crc kubenswrapper[4978]: I0225 08:33:17.334939 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:33:17 crc kubenswrapper[4978]: E0225 08:33:17.336058 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:33:26 crc kubenswrapper[4978]: I0225 08:33:26.898114 4978 scope.go:117] "RemoveContainer" containerID="8eb08550ac619cc24779584a8a60dad408e75a3636fff54ad14f71daa4268556" Feb 25 08:33:26 crc kubenswrapper[4978]: I0225 08:33:26.926336 4978 scope.go:117] "RemoveContainer" containerID="ab593c3ff7696a5ce5d120903b7d0230eb4b7dfd21902abd1ee0048bd253f13f" Feb 25 08:33:31 crc kubenswrapper[4978]: I0225 08:33:31.328882 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:33:31 crc kubenswrapper[4978]: E0225 08:33:31.329702 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:33:42 crc kubenswrapper[4978]: I0225 08:33:42.327807 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:33:42 crc kubenswrapper[4978]: E0225 08:33:42.328818 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.723799 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wcgq6"] Feb 25 08:33:48 crc kubenswrapper[4978]: E0225 08:33:48.725051 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1999d0d3-0e65-45ea-8133-f6b40154c125" containerName="oc" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.725079 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1999d0d3-0e65-45ea-8133-f6b40154c125" containerName="oc" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.725436 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1999d0d3-0e65-45ea-8133-f6b40154c125" containerName="oc" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.727837 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.733660 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wcgq6"] Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.846448 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhh68\" (UniqueName: \"kubernetes.io/projected/be6af1fa-c93e-48c5-a738-6db5dcdbec40-kube-api-access-nhh68\") pod \"redhat-operators-wcgq6\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.846538 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-catalog-content\") pod \"redhat-operators-wcgq6\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.846598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-utilities\") pod \"redhat-operators-wcgq6\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.947714 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-utilities\") pod \"redhat-operators-wcgq6\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.947870 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhh68\" (UniqueName: \"kubernetes.io/projected/be6af1fa-c93e-48c5-a738-6db5dcdbec40-kube-api-access-nhh68\") pod \"redhat-operators-wcgq6\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.947933 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-catalog-content\") pod \"redhat-operators-wcgq6\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.948421 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-catalog-content\") pod \"redhat-operators-wcgq6\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.948421 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-utilities\") pod \"redhat-operators-wcgq6\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:48 crc kubenswrapper[4978]: I0225 08:33:48.968892 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhh68\" (UniqueName: \"kubernetes.io/projected/be6af1fa-c93e-48c5-a738-6db5dcdbec40-kube-api-access-nhh68\") pod \"redhat-operators-wcgq6\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:49 crc kubenswrapper[4978]: I0225 08:33:49.088974 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:49 crc kubenswrapper[4978]: I0225 08:33:49.527907 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wcgq6"] Feb 25 08:33:50 crc kubenswrapper[4978]: I0225 08:33:50.104123 4978 generic.go:334] "Generic (PLEG): container finished" podID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerID="520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f" exitCode=0 Feb 25 08:33:50 crc kubenswrapper[4978]: I0225 08:33:50.104166 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgq6" event={"ID":"be6af1fa-c93e-48c5-a738-6db5dcdbec40","Type":"ContainerDied","Data":"520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f"} Feb 25 08:33:50 crc kubenswrapper[4978]: I0225 08:33:50.104387 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgq6" event={"ID":"be6af1fa-c93e-48c5-a738-6db5dcdbec40","Type":"ContainerStarted","Data":"4fd3fa6c3da548b7302d29adb88c4161f113ca62f1a5d9de1483fee77ab70ab6"} Feb 25 08:33:50 crc kubenswrapper[4978]: I0225 08:33:50.105775 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 08:33:52 crc kubenswrapper[4978]: I0225 08:33:52.132681 4978 generic.go:334] "Generic (PLEG): container finished" podID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerID="93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b" exitCode=0 Feb 25 08:33:52 crc kubenswrapper[4978]: I0225 08:33:52.132729 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgq6" event={"ID":"be6af1fa-c93e-48c5-a738-6db5dcdbec40","Type":"ContainerDied","Data":"93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b"} Feb 25 08:33:53 crc kubenswrapper[4978]: I0225 08:33:53.169880 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgq6" event={"ID":"be6af1fa-c93e-48c5-a738-6db5dcdbec40","Type":"ContainerStarted","Data":"c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149"} Feb 25 08:33:53 crc kubenswrapper[4978]: I0225 08:33:53.206009 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wcgq6" podStartSLOduration=2.731764083 podStartE2EDuration="5.20592395s" podCreationTimestamp="2026-02-25 08:33:48 +0000 UTC" firstStartedPulling="2026-02-25 08:33:50.105568692 +0000 UTC m=+6523.544825151" lastFinishedPulling="2026-02-25 08:33:52.579728519 +0000 UTC m=+6526.018985018" observedRunningTime="2026-02-25 08:33:53.201125921 +0000 UTC m=+6526.640382440" watchObservedRunningTime="2026-02-25 08:33:53.20592395 +0000 UTC m=+6526.645180439" Feb 25 08:33:53 crc kubenswrapper[4978]: I0225 08:33:53.329146 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:33:53 crc kubenswrapper[4978]: E0225 08:33:53.329983 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:33:59 crc kubenswrapper[4978]: I0225 08:33:59.090315 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:33:59 crc kubenswrapper[4978]: I0225 08:33:59.091153 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.151711 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533474-plz6z"] Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.153474 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533474-plz6z" Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.156059 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wcgq6" podUID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerName="registry-server" probeResult="failure" output=< Feb 25 08:34:00 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 08:34:00 crc kubenswrapper[4978]: > Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.160663 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.161119 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.164904 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.182898 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533474-plz6z"] Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.274055 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z46lj\" (UniqueName: \"kubernetes.io/projected/f910236e-834f-4eb4-97c5-ca2b8d77546c-kube-api-access-z46lj\") pod \"auto-csr-approver-29533474-plz6z\" (UID: \"f910236e-834f-4eb4-97c5-ca2b8d77546c\") " pod="openshift-infra/auto-csr-approver-29533474-plz6z" Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.375117 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z46lj\" (UniqueName: \"kubernetes.io/projected/f910236e-834f-4eb4-97c5-ca2b8d77546c-kube-api-access-z46lj\") pod \"auto-csr-approver-29533474-plz6z\" (UID: \"f910236e-834f-4eb4-97c5-ca2b8d77546c\") " pod="openshift-infra/auto-csr-approver-29533474-plz6z" Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.396153 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z46lj\" (UniqueName: \"kubernetes.io/projected/f910236e-834f-4eb4-97c5-ca2b8d77546c-kube-api-access-z46lj\") pod \"auto-csr-approver-29533474-plz6z\" (UID: \"f910236e-834f-4eb4-97c5-ca2b8d77546c\") " pod="openshift-infra/auto-csr-approver-29533474-plz6z" Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.477578 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533474-plz6z" Feb 25 08:34:00 crc kubenswrapper[4978]: I0225 08:34:00.747133 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533474-plz6z"] Feb 25 08:34:01 crc kubenswrapper[4978]: I0225 08:34:01.260172 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533474-plz6z" event={"ID":"f910236e-834f-4eb4-97c5-ca2b8d77546c","Type":"ContainerStarted","Data":"a7e6233ad27bb4c638e071ec4066ed91147cb4350bfff64bacc194b6a60bccff"} Feb 25 08:34:02 crc kubenswrapper[4978]: I0225 08:34:02.273453 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533474-plz6z" event={"ID":"f910236e-834f-4eb4-97c5-ca2b8d77546c","Type":"ContainerStarted","Data":"8eeb1a257fd9a2146d07c4ed3d994ce14a1fe84a4bdff4887eaebc530651c80a"} Feb 25 08:34:02 crc kubenswrapper[4978]: I0225 08:34:02.296873 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533474-plz6z" podStartSLOduration=1.178208392 podStartE2EDuration="2.296833537s" podCreationTimestamp="2026-02-25 08:34:00 +0000 UTC" firstStartedPulling="2026-02-25 08:34:00.753252348 +0000 UTC m=+6534.192508817" lastFinishedPulling="2026-02-25 08:34:01.871877463 +0000 UTC m=+6535.311133962" observedRunningTime="2026-02-25 08:34:02.290830281 +0000 UTC m=+6535.730086790" watchObservedRunningTime="2026-02-25 08:34:02.296833537 +0000 UTC m=+6535.736090036" Feb 25 08:34:03 crc kubenswrapper[4978]: I0225 08:34:03.305324 4978 generic.go:334] "Generic (PLEG): container finished" podID="f910236e-834f-4eb4-97c5-ca2b8d77546c" containerID="8eeb1a257fd9a2146d07c4ed3d994ce14a1fe84a4bdff4887eaebc530651c80a" exitCode=0 Feb 25 08:34:03 crc kubenswrapper[4978]: I0225 08:34:03.305487 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533474-plz6z" event={"ID":"f910236e-834f-4eb4-97c5-ca2b8d77546c","Type":"ContainerDied","Data":"8eeb1a257fd9a2146d07c4ed3d994ce14a1fe84a4bdff4887eaebc530651c80a"} Feb 25 08:34:04 crc kubenswrapper[4978]: I0225 08:34:04.738469 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533474-plz6z" Feb 25 08:34:04 crc kubenswrapper[4978]: I0225 08:34:04.854411 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z46lj\" (UniqueName: \"kubernetes.io/projected/f910236e-834f-4eb4-97c5-ca2b8d77546c-kube-api-access-z46lj\") pod \"f910236e-834f-4eb4-97c5-ca2b8d77546c\" (UID: \"f910236e-834f-4eb4-97c5-ca2b8d77546c\") " Feb 25 08:34:04 crc kubenswrapper[4978]: I0225 08:34:04.860734 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f910236e-834f-4eb4-97c5-ca2b8d77546c-kube-api-access-z46lj" (OuterVolumeSpecName: "kube-api-access-z46lj") pod "f910236e-834f-4eb4-97c5-ca2b8d77546c" (UID: "f910236e-834f-4eb4-97c5-ca2b8d77546c"). InnerVolumeSpecName "kube-api-access-z46lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:34:04 crc kubenswrapper[4978]: I0225 08:34:04.956283 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z46lj\" (UniqueName: \"kubernetes.io/projected/f910236e-834f-4eb4-97c5-ca2b8d77546c-kube-api-access-z46lj\") on node \"crc\" DevicePath \"\"" Feb 25 08:34:05 crc kubenswrapper[4978]: I0225 08:34:05.336593 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533474-plz6z" Feb 25 08:34:05 crc kubenswrapper[4978]: I0225 08:34:05.360574 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533474-plz6z" event={"ID":"f910236e-834f-4eb4-97c5-ca2b8d77546c","Type":"ContainerDied","Data":"a7e6233ad27bb4c638e071ec4066ed91147cb4350bfff64bacc194b6a60bccff"} Feb 25 08:34:05 crc kubenswrapper[4978]: I0225 08:34:05.360650 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7e6233ad27bb4c638e071ec4066ed91147cb4350bfff64bacc194b6a60bccff" Feb 25 08:34:05 crc kubenswrapper[4978]: I0225 08:34:05.407081 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533468-z8znk"] Feb 25 08:34:05 crc kubenswrapper[4978]: I0225 08:34:05.412435 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533468-z8znk"] Feb 25 08:34:07 crc kubenswrapper[4978]: I0225 08:34:07.359469 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95b01cb1-795a-4462-b98c-669366afaa8a" path="/var/lib/kubelet/pods/95b01cb1-795a-4462-b98c-669366afaa8a/volumes" Feb 25 08:34:08 crc kubenswrapper[4978]: I0225 08:34:08.328665 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:34:08 crc kubenswrapper[4978]: E0225 08:34:08.329218 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:34:09 crc kubenswrapper[4978]: I0225 08:34:09.164651 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:34:09 crc kubenswrapper[4978]: I0225 08:34:09.243125 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:34:09 crc kubenswrapper[4978]: I0225 08:34:09.423390 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wcgq6"] Feb 25 08:34:10 crc kubenswrapper[4978]: I0225 08:34:10.395324 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wcgq6" podUID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerName="registry-server" containerID="cri-o://c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149" gracePeriod=2 Feb 25 08:34:10 crc kubenswrapper[4978]: I0225 08:34:10.921204 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:34:10 crc kubenswrapper[4978]: I0225 08:34:10.965852 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhh68\" (UniqueName: \"kubernetes.io/projected/be6af1fa-c93e-48c5-a738-6db5dcdbec40-kube-api-access-nhh68\") pod \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " Feb 25 08:34:10 crc kubenswrapper[4978]: I0225 08:34:10.965929 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-utilities\") pod \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " Feb 25 08:34:10 crc kubenswrapper[4978]: I0225 08:34:10.965990 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-catalog-content\") pod \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\" (UID: \"be6af1fa-c93e-48c5-a738-6db5dcdbec40\") " Feb 25 08:34:10 crc kubenswrapper[4978]: I0225 08:34:10.967277 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-utilities" (OuterVolumeSpecName: "utilities") pod "be6af1fa-c93e-48c5-a738-6db5dcdbec40" (UID: "be6af1fa-c93e-48c5-a738-6db5dcdbec40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:34:10 crc kubenswrapper[4978]: I0225 08:34:10.973622 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be6af1fa-c93e-48c5-a738-6db5dcdbec40-kube-api-access-nhh68" (OuterVolumeSpecName: "kube-api-access-nhh68") pod "be6af1fa-c93e-48c5-a738-6db5dcdbec40" (UID: "be6af1fa-c93e-48c5-a738-6db5dcdbec40"). InnerVolumeSpecName "kube-api-access-nhh68". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.067789 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhh68\" (UniqueName: \"kubernetes.io/projected/be6af1fa-c93e-48c5-a738-6db5dcdbec40-kube-api-access-nhh68\") on node \"crc\" DevicePath \"\"" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.067823 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.120215 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be6af1fa-c93e-48c5-a738-6db5dcdbec40" (UID: "be6af1fa-c93e-48c5-a738-6db5dcdbec40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.169634 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be6af1fa-c93e-48c5-a738-6db5dcdbec40-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.406520 4978 generic.go:334] "Generic (PLEG): container finished" podID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerID="c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149" exitCode=0 Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.406599 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgq6" event={"ID":"be6af1fa-c93e-48c5-a738-6db5dcdbec40","Type":"ContainerDied","Data":"c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149"} Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.406643 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgq6" event={"ID":"be6af1fa-c93e-48c5-a738-6db5dcdbec40","Type":"ContainerDied","Data":"4fd3fa6c3da548b7302d29adb88c4161f113ca62f1a5d9de1483fee77ab70ab6"} Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.406657 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wcgq6" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.406680 4978 scope.go:117] "RemoveContainer" containerID="c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.454117 4978 scope.go:117] "RemoveContainer" containerID="93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.465070 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wcgq6"] Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.476830 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wcgq6"] Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.488749 4978 scope.go:117] "RemoveContainer" containerID="520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.539496 4978 scope.go:117] "RemoveContainer" containerID="c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149" Feb 25 08:34:11 crc kubenswrapper[4978]: E0225 08:34:11.540133 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149\": container with ID starting with c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149 not found: ID does not exist" containerID="c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.540178 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149"} err="failed to get container status \"c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149\": rpc error: code = NotFound desc = could not find container \"c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149\": container with ID starting with c79b515ec35e2d67e7ba30658c56d9d7dbee2af3133fd197f41044f4e50f3149 not found: ID does not exist" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.540212 4978 scope.go:117] "RemoveContainer" containerID="93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b" Feb 25 08:34:11 crc kubenswrapper[4978]: E0225 08:34:11.540889 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b\": container with ID starting with 93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b not found: ID does not exist" containerID="93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.540970 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b"} err="failed to get container status \"93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b\": rpc error: code = NotFound desc = could not find container \"93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b\": container with ID starting with 93c397d28fba671d4f427444588d6656ec2757ae839e86e5df5783b97f8a1a4b not found: ID does not exist" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.541000 4978 scope.go:117] "RemoveContainer" containerID="520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f" Feb 25 08:34:11 crc kubenswrapper[4978]: E0225 08:34:11.541540 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f\": container with ID starting with 520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f not found: ID does not exist" containerID="520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f" Feb 25 08:34:11 crc kubenswrapper[4978]: I0225 08:34:11.541625 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f"} err="failed to get container status \"520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f\": rpc error: code = NotFound desc = could not find container \"520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f\": container with ID starting with 520a265412b79105ec7fae6382e10b91aa41a62e3de01a92dcc602ea8a3e109f not found: ID does not exist" Feb 25 08:34:13 crc kubenswrapper[4978]: I0225 08:34:13.342061 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" path="/var/lib/kubelet/pods/be6af1fa-c93e-48c5-a738-6db5dcdbec40/volumes" Feb 25 08:34:23 crc kubenswrapper[4978]: I0225 08:34:23.327874 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:34:23 crc kubenswrapper[4978]: E0225 08:34:23.328896 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:34:27 crc kubenswrapper[4978]: I0225 08:34:27.011872 4978 scope.go:117] "RemoveContainer" containerID="910915931f4de4652e60446aa0c6e462a3c51a2f57e5567145aa0df216734ca0" Feb 25 08:34:27 crc kubenswrapper[4978]: I0225 08:34:27.048491 4978 scope.go:117] "RemoveContainer" containerID="2836e002724c7861d4a21769b1fe8f1005982b144fb7e7154401b54a1b8f113b" Feb 25 08:34:34 crc kubenswrapper[4978]: I0225 08:34:34.328071 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:34:34 crc kubenswrapper[4978]: E0225 08:34:34.330249 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:34:49 crc kubenswrapper[4978]: I0225 08:34:49.327981 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:34:49 crc kubenswrapper[4978]: E0225 08:34:49.328979 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:35:04 crc kubenswrapper[4978]: I0225 08:35:04.328637 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:35:04 crc kubenswrapper[4978]: E0225 08:35:04.329799 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:35:19 crc kubenswrapper[4978]: I0225 08:35:19.328228 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:35:19 crc kubenswrapper[4978]: E0225 08:35:19.329196 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:35:31 crc kubenswrapper[4978]: I0225 08:35:31.327763 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:35:31 crc kubenswrapper[4978]: E0225 08:35:31.329083 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:35:44 crc kubenswrapper[4978]: I0225 08:35:44.328922 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:35:44 crc kubenswrapper[4978]: E0225 08:35:44.330076 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:35:55 crc kubenswrapper[4978]: I0225 08:35:55.328727 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:35:55 crc kubenswrapper[4978]: E0225 08:35:55.329532 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.170609 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533476-wz5xc"] Feb 25 08:36:00 crc kubenswrapper[4978]: E0225 08:36:00.171640 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerName="extract-content" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.171670 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerName="extract-content" Feb 25 08:36:00 crc kubenswrapper[4978]: E0225 08:36:00.171704 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerName="extract-utilities" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.171720 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerName="extract-utilities" Feb 25 08:36:00 crc kubenswrapper[4978]: E0225 08:36:00.171764 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerName="registry-server" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.171782 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerName="registry-server" Feb 25 08:36:00 crc kubenswrapper[4978]: E0225 08:36:00.171803 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f910236e-834f-4eb4-97c5-ca2b8d77546c" containerName="oc" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.171818 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f910236e-834f-4eb4-97c5-ca2b8d77546c" containerName="oc" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.172127 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f910236e-834f-4eb4-97c5-ca2b8d77546c" containerName="oc" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.172149 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="be6af1fa-c93e-48c5-a738-6db5dcdbec40" containerName="registry-server" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.173024 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533476-wz5xc" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.175566 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.176224 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.177453 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.199585 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533476-wz5xc"] Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.293441 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx62f\" (UniqueName: \"kubernetes.io/projected/55f0901c-b44d-47e6-9dc4-245e937029a6-kube-api-access-cx62f\") pod \"auto-csr-approver-29533476-wz5xc\" (UID: \"55f0901c-b44d-47e6-9dc4-245e937029a6\") " pod="openshift-infra/auto-csr-approver-29533476-wz5xc" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.394576 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx62f\" (UniqueName: \"kubernetes.io/projected/55f0901c-b44d-47e6-9dc4-245e937029a6-kube-api-access-cx62f\") pod \"auto-csr-approver-29533476-wz5xc\" (UID: \"55f0901c-b44d-47e6-9dc4-245e937029a6\") " pod="openshift-infra/auto-csr-approver-29533476-wz5xc" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.413959 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx62f\" (UniqueName: \"kubernetes.io/projected/55f0901c-b44d-47e6-9dc4-245e937029a6-kube-api-access-cx62f\") pod \"auto-csr-approver-29533476-wz5xc\" (UID: \"55f0901c-b44d-47e6-9dc4-245e937029a6\") " pod="openshift-infra/auto-csr-approver-29533476-wz5xc" Feb 25 08:36:00 crc kubenswrapper[4978]: I0225 08:36:00.512581 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533476-wz5xc" Feb 25 08:36:01 crc kubenswrapper[4978]: I0225 08:36:01.011747 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533476-wz5xc"] Feb 25 08:36:01 crc kubenswrapper[4978]: W0225 08:36:01.011776 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55f0901c_b44d_47e6_9dc4_245e937029a6.slice/crio-dbfa70ad416ceccff76d2b740fc82d04ada5041f7dff2d86d0a026d3cf87ef51 WatchSource:0}: Error finding container dbfa70ad416ceccff76d2b740fc82d04ada5041f7dff2d86d0a026d3cf87ef51: Status 404 returned error can't find the container with id dbfa70ad416ceccff76d2b740fc82d04ada5041f7dff2d86d0a026d3cf87ef51 Feb 25 08:36:01 crc kubenswrapper[4978]: I0225 08:36:01.474445 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533476-wz5xc" event={"ID":"55f0901c-b44d-47e6-9dc4-245e937029a6","Type":"ContainerStarted","Data":"dbfa70ad416ceccff76d2b740fc82d04ada5041f7dff2d86d0a026d3cf87ef51"} Feb 25 08:36:02 crc kubenswrapper[4978]: I0225 08:36:02.488066 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533476-wz5xc" event={"ID":"55f0901c-b44d-47e6-9dc4-245e937029a6","Type":"ContainerStarted","Data":"9e3f906d24a1336c48c4c76cb0fec7f366bdf78252141d85a4a99800e7d37ce1"} Feb 25 08:36:02 crc kubenswrapper[4978]: I0225 08:36:02.511645 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533476-wz5xc" podStartSLOduration=1.378729337 podStartE2EDuration="2.511614095s" podCreationTimestamp="2026-02-25 08:36:00 +0000 UTC" firstStartedPulling="2026-02-25 08:36:01.014704043 +0000 UTC m=+6654.453960502" lastFinishedPulling="2026-02-25 08:36:02.147588761 +0000 UTC m=+6655.586845260" observedRunningTime="2026-02-25 08:36:02.50822577 +0000 UTC m=+6655.947482269" watchObservedRunningTime="2026-02-25 08:36:02.511614095 +0000 UTC m=+6655.950870584" Feb 25 08:36:03 crc kubenswrapper[4978]: I0225 08:36:03.501451 4978 generic.go:334] "Generic (PLEG): container finished" podID="55f0901c-b44d-47e6-9dc4-245e937029a6" containerID="9e3f906d24a1336c48c4c76cb0fec7f366bdf78252141d85a4a99800e7d37ce1" exitCode=0 Feb 25 08:36:03 crc kubenswrapper[4978]: I0225 08:36:03.501513 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533476-wz5xc" event={"ID":"55f0901c-b44d-47e6-9dc4-245e937029a6","Type":"ContainerDied","Data":"9e3f906d24a1336c48c4c76cb0fec7f366bdf78252141d85a4a99800e7d37ce1"} Feb 25 08:36:04 crc kubenswrapper[4978]: I0225 08:36:04.955578 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533476-wz5xc" Feb 25 08:36:05 crc kubenswrapper[4978]: I0225 08:36:05.087283 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx62f\" (UniqueName: \"kubernetes.io/projected/55f0901c-b44d-47e6-9dc4-245e937029a6-kube-api-access-cx62f\") pod \"55f0901c-b44d-47e6-9dc4-245e937029a6\" (UID: \"55f0901c-b44d-47e6-9dc4-245e937029a6\") " Feb 25 08:36:05 crc kubenswrapper[4978]: I0225 08:36:05.093322 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55f0901c-b44d-47e6-9dc4-245e937029a6-kube-api-access-cx62f" (OuterVolumeSpecName: "kube-api-access-cx62f") pod "55f0901c-b44d-47e6-9dc4-245e937029a6" (UID: "55f0901c-b44d-47e6-9dc4-245e937029a6"). InnerVolumeSpecName "kube-api-access-cx62f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:36:05 crc kubenswrapper[4978]: I0225 08:36:05.189108 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx62f\" (UniqueName: \"kubernetes.io/projected/55f0901c-b44d-47e6-9dc4-245e937029a6-kube-api-access-cx62f\") on node \"crc\" DevicePath \"\"" Feb 25 08:36:05 crc kubenswrapper[4978]: I0225 08:36:05.534340 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533476-wz5xc" event={"ID":"55f0901c-b44d-47e6-9dc4-245e937029a6","Type":"ContainerDied","Data":"dbfa70ad416ceccff76d2b740fc82d04ada5041f7dff2d86d0a026d3cf87ef51"} Feb 25 08:36:05 crc kubenswrapper[4978]: I0225 08:36:05.534828 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbfa70ad416ceccff76d2b740fc82d04ada5041f7dff2d86d0a026d3cf87ef51" Feb 25 08:36:05 crc kubenswrapper[4978]: I0225 08:36:05.534507 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533476-wz5xc" Feb 25 08:36:05 crc kubenswrapper[4978]: I0225 08:36:05.599800 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533470-5cs55"] Feb 25 08:36:05 crc kubenswrapper[4978]: I0225 08:36:05.607102 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533470-5cs55"] Feb 25 08:36:07 crc kubenswrapper[4978]: I0225 08:36:07.345640 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a285f988-f118-4b73-aaeb-d1f36a843d85" path="/var/lib/kubelet/pods/a285f988-f118-4b73-aaeb-d1f36a843d85/volumes" Feb 25 08:36:09 crc kubenswrapper[4978]: I0225 08:36:09.327830 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:36:09 crc kubenswrapper[4978]: E0225 08:36:09.328314 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:36:20 crc kubenswrapper[4978]: I0225 08:36:20.327760 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:36:20 crc kubenswrapper[4978]: E0225 08:36:20.328676 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:36:27 crc kubenswrapper[4978]: I0225 08:36:27.142015 4978 scope.go:117] "RemoveContainer" containerID="3eea4b2ca29e6aa79cb0f7100898c21e0efced1fe6f69485bfee3cd9f0e073d9" Feb 25 08:36:31 crc kubenswrapper[4978]: I0225 08:36:31.327835 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:36:31 crc kubenswrapper[4978]: E0225 08:36:31.329177 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.215237 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-slf9q"] Feb 25 08:36:43 crc kubenswrapper[4978]: E0225 08:36:43.216389 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55f0901c-b44d-47e6-9dc4-245e937029a6" containerName="oc" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.216413 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="55f0901c-b44d-47e6-9dc4-245e937029a6" containerName="oc" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.216660 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="55f0901c-b44d-47e6-9dc4-245e937029a6" containerName="oc" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.218396 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.225902 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slf9q"] Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.350338 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-catalog-content\") pod \"certified-operators-slf9q\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.350684 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-utilities\") pod \"certified-operators-slf9q\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.350773 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7jsv\" (UniqueName: \"kubernetes.io/projected/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-kube-api-access-l7jsv\") pod \"certified-operators-slf9q\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.452625 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-catalog-content\") pod \"certified-operators-slf9q\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.452718 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-utilities\") pod \"certified-operators-slf9q\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.452822 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7jsv\" (UniqueName: \"kubernetes.io/projected/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-kube-api-access-l7jsv\") pod \"certified-operators-slf9q\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.453335 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-catalog-content\") pod \"certified-operators-slf9q\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.453463 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-utilities\") pod \"certified-operators-slf9q\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.488187 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7jsv\" (UniqueName: \"kubernetes.io/projected/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-kube-api-access-l7jsv\") pod \"certified-operators-slf9q\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.534235 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.793176 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slf9q"] Feb 25 08:36:43 crc kubenswrapper[4978]: I0225 08:36:43.935915 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slf9q" event={"ID":"99b45ecc-b1e9-4127-be9a-4d3e48d92e67","Type":"ContainerStarted","Data":"1e1bf3c2859e795db3318f1bd5b77f91c2d85865f727782ca51a2997f9cd5e32"} Feb 25 08:36:44 crc kubenswrapper[4978]: I0225 08:36:44.949281 4978 generic.go:334] "Generic (PLEG): container finished" podID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerID="de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836" exitCode=0 Feb 25 08:36:44 crc kubenswrapper[4978]: I0225 08:36:44.949439 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slf9q" event={"ID":"99b45ecc-b1e9-4127-be9a-4d3e48d92e67","Type":"ContainerDied","Data":"de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836"} Feb 25 08:36:46 crc kubenswrapper[4978]: I0225 08:36:46.328599 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:36:46 crc kubenswrapper[4978]: E0225 08:36:46.329228 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:36:46 crc kubenswrapper[4978]: I0225 08:36:46.975258 4978 generic.go:334] "Generic (PLEG): container finished" podID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerID="ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02" exitCode=0 Feb 25 08:36:46 crc kubenswrapper[4978]: I0225 08:36:46.975318 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slf9q" event={"ID":"99b45ecc-b1e9-4127-be9a-4d3e48d92e67","Type":"ContainerDied","Data":"ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02"} Feb 25 08:36:47 crc kubenswrapper[4978]: I0225 08:36:47.986079 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slf9q" event={"ID":"99b45ecc-b1e9-4127-be9a-4d3e48d92e67","Type":"ContainerStarted","Data":"6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa"} Feb 25 08:36:48 crc kubenswrapper[4978]: I0225 08:36:48.014399 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-slf9q" podStartSLOduration=2.600225843 podStartE2EDuration="5.014360968s" podCreationTimestamp="2026-02-25 08:36:43 +0000 UTC" firstStartedPulling="2026-02-25 08:36:44.953365001 +0000 UTC m=+6698.392648891" lastFinishedPulling="2026-02-25 08:36:47.367527537 +0000 UTC m=+6700.806784016" observedRunningTime="2026-02-25 08:36:48.008090474 +0000 UTC m=+6701.447346943" watchObservedRunningTime="2026-02-25 08:36:48.014360968 +0000 UTC m=+6701.453617427" Feb 25 08:36:53 crc kubenswrapper[4978]: I0225 08:36:53.534544 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:53 crc kubenswrapper[4978]: I0225 08:36:53.535131 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:53 crc kubenswrapper[4978]: I0225 08:36:53.610959 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:54 crc kubenswrapper[4978]: I0225 08:36:54.099305 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:54 crc kubenswrapper[4978]: I0225 08:36:54.165245 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-slf9q"] Feb 25 08:36:56 crc kubenswrapper[4978]: I0225 08:36:56.112215 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-slf9q" podUID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerName="registry-server" containerID="cri-o://6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa" gracePeriod=2 Feb 25 08:36:56 crc kubenswrapper[4978]: I0225 08:36:56.610480 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:56 crc kubenswrapper[4978]: I0225 08:36:56.729452 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7jsv\" (UniqueName: \"kubernetes.io/projected/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-kube-api-access-l7jsv\") pod \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " Feb 25 08:36:56 crc kubenswrapper[4978]: I0225 08:36:56.729566 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-catalog-content\") pod \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " Feb 25 08:36:56 crc kubenswrapper[4978]: I0225 08:36:56.729655 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-utilities\") pod \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\" (UID: \"99b45ecc-b1e9-4127-be9a-4d3e48d92e67\") " Feb 25 08:36:56 crc kubenswrapper[4978]: I0225 08:36:56.731063 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-utilities" (OuterVolumeSpecName: "utilities") pod "99b45ecc-b1e9-4127-be9a-4d3e48d92e67" (UID: "99b45ecc-b1e9-4127-be9a-4d3e48d92e67"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:36:56 crc kubenswrapper[4978]: I0225 08:36:56.737314 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-kube-api-access-l7jsv" (OuterVolumeSpecName: "kube-api-access-l7jsv") pod "99b45ecc-b1e9-4127-be9a-4d3e48d92e67" (UID: "99b45ecc-b1e9-4127-be9a-4d3e48d92e67"). InnerVolumeSpecName "kube-api-access-l7jsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:36:56 crc kubenswrapper[4978]: I0225 08:36:56.832191 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7jsv\" (UniqueName: \"kubernetes.io/projected/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-kube-api-access-l7jsv\") on node \"crc\" DevicePath \"\"" Feb 25 08:36:56 crc kubenswrapper[4978]: I0225 08:36:56.832240 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.044237 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99b45ecc-b1e9-4127-be9a-4d3e48d92e67" (UID: "99b45ecc-b1e9-4127-be9a-4d3e48d92e67"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.124986 4978 generic.go:334] "Generic (PLEG): container finished" podID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerID="6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa" exitCode=0 Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.125044 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slf9q" event={"ID":"99b45ecc-b1e9-4127-be9a-4d3e48d92e67","Type":"ContainerDied","Data":"6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa"} Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.125080 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slf9q" event={"ID":"99b45ecc-b1e9-4127-be9a-4d3e48d92e67","Type":"ContainerDied","Data":"1e1bf3c2859e795db3318f1bd5b77f91c2d85865f727782ca51a2997f9cd5e32"} Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.125100 4978 scope.go:117] "RemoveContainer" containerID="6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.125113 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slf9q" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.139301 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99b45ecc-b1e9-4127-be9a-4d3e48d92e67-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.153616 4978 scope.go:117] "RemoveContainer" containerID="ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.177770 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-slf9q"] Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.186270 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-slf9q"] Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.191878 4978 scope.go:117] "RemoveContainer" containerID="de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.216651 4978 scope.go:117] "RemoveContainer" containerID="6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa" Feb 25 08:36:57 crc kubenswrapper[4978]: E0225 08:36:57.217256 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa\": container with ID starting with 6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa not found: ID does not exist" containerID="6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.217312 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa"} err="failed to get container status \"6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa\": rpc error: code = NotFound desc = could not find container \"6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa\": container with ID starting with 6934d5517d09f846651ba55bef5e9c65d7ce8788a13c808819aee600c3f536aa not found: ID does not exist" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.217354 4978 scope.go:117] "RemoveContainer" containerID="ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02" Feb 25 08:36:57 crc kubenswrapper[4978]: E0225 08:36:57.217900 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02\": container with ID starting with ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02 not found: ID does not exist" containerID="ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.217942 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02"} err="failed to get container status \"ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02\": rpc error: code = NotFound desc = could not find container \"ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02\": container with ID starting with ac7ddf47e19c990ee2bcdd264496e9d879a0e2491cfeb40d52cc8afa0f2a5d02 not found: ID does not exist" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.217973 4978 scope.go:117] "RemoveContainer" containerID="de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836" Feb 25 08:36:57 crc kubenswrapper[4978]: E0225 08:36:57.218552 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836\": container with ID starting with de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836 not found: ID does not exist" containerID="de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.218602 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836"} err="failed to get container status \"de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836\": rpc error: code = NotFound desc = could not find container \"de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836\": container with ID starting with de82f8bf0a319905c1f51bc70cafc5bfd7f8a1c1c51435afc4140c6f62c95836 not found: ID does not exist" Feb 25 08:36:57 crc kubenswrapper[4978]: I0225 08:36:57.340848 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" path="/var/lib/kubelet/pods/99b45ecc-b1e9-4127-be9a-4d3e48d92e67/volumes" Feb 25 08:37:00 crc kubenswrapper[4978]: I0225 08:37:00.327790 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:37:01 crc kubenswrapper[4978]: I0225 08:37:01.174950 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"914d9830f2d8d5aaf326a3efa925940acc7734a6c345c2cd844ec55b8a401b2b"} Feb 25 08:37:23 crc kubenswrapper[4978]: I0225 08:37:23.075059 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-5zkv5"] Feb 25 08:37:23 crc kubenswrapper[4978]: I0225 08:37:23.081606 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-5zkv5"] Feb 25 08:37:23 crc kubenswrapper[4978]: I0225 08:37:23.342997 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd608922-a6fb-4b0a-b693-a2409079044e" path="/var/lib/kubelet/pods/bd608922-a6fb-4b0a-b693-a2409079044e/volumes" Feb 25 08:37:27 crc kubenswrapper[4978]: I0225 08:37:27.236776 4978 scope.go:117] "RemoveContainer" containerID="ab6e81ad25d93636134c19ac61fed0a82345386a33ad7e66e753035460048efb" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.478521 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Feb 25 08:37:33 crc kubenswrapper[4978]: E0225 08:37:33.480275 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerName="extract-content" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.480310 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerName="extract-content" Feb 25 08:37:33 crc kubenswrapper[4978]: E0225 08:37:33.480347 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerName="registry-server" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.480398 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerName="registry-server" Feb 25 08:37:33 crc kubenswrapper[4978]: E0225 08:37:33.480432 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerName="extract-utilities" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.480451 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerName="extract-utilities" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.480772 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="99b45ecc-b1e9-4127-be9a-4d3e48d92e67" containerName="registry-server" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.481612 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.509790 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fm969" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.526033 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.582900 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmscv\" (UniqueName: \"kubernetes.io/projected/4c2aab8b-8456-4e4c-b282-6b7074a116cc-kube-api-access-rmscv\") pod \"mariadb-copy-data\" (UID: \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\") " pod="openstack/mariadb-copy-data" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.582961 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a6207369-0c50-4031-8864-29f04e3de609\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6207369-0c50-4031-8864-29f04e3de609\") pod \"mariadb-copy-data\" (UID: \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\") " pod="openstack/mariadb-copy-data" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.684827 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmscv\" (UniqueName: \"kubernetes.io/projected/4c2aab8b-8456-4e4c-b282-6b7074a116cc-kube-api-access-rmscv\") pod \"mariadb-copy-data\" (UID: \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\") " pod="openstack/mariadb-copy-data" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.684894 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a6207369-0c50-4031-8864-29f04e3de609\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6207369-0c50-4031-8864-29f04e3de609\") pod \"mariadb-copy-data\" (UID: \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\") " pod="openstack/mariadb-copy-data" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.688735 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.688811 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a6207369-0c50-4031-8864-29f04e3de609\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6207369-0c50-4031-8864-29f04e3de609\") pod \"mariadb-copy-data\" (UID: \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f83b9d16490610f7f6e6390b610cbd87ad0fedc690884b39e16c758955f582a6/globalmount\"" pod="openstack/mariadb-copy-data" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.722683 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmscv\" (UniqueName: \"kubernetes.io/projected/4c2aab8b-8456-4e4c-b282-6b7074a116cc-kube-api-access-rmscv\") pod \"mariadb-copy-data\" (UID: \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\") " pod="openstack/mariadb-copy-data" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.726688 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a6207369-0c50-4031-8864-29f04e3de609\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6207369-0c50-4031-8864-29f04e3de609\") pod \"mariadb-copy-data\" (UID: \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\") " pod="openstack/mariadb-copy-data" Feb 25 08:37:33 crc kubenswrapper[4978]: I0225 08:37:33.840154 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Feb 25 08:37:34 crc kubenswrapper[4978]: I0225 08:37:34.238548 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Feb 25 08:37:34 crc kubenswrapper[4978]: W0225 08:37:34.244792 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c2aab8b_8456_4e4c_b282_6b7074a116cc.slice/crio-2699cc89291b12f2fee518daf3f8115c24c110ba0bd2753e738bfe8a72b6881b WatchSource:0}: Error finding container 2699cc89291b12f2fee518daf3f8115c24c110ba0bd2753e738bfe8a72b6881b: Status 404 returned error can't find the container with id 2699cc89291b12f2fee518daf3f8115c24c110ba0bd2753e738bfe8a72b6881b Feb 25 08:37:34 crc kubenswrapper[4978]: I0225 08:37:34.529048 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"4c2aab8b-8456-4e4c-b282-6b7074a116cc","Type":"ContainerStarted","Data":"2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a"} Feb 25 08:37:34 crc kubenswrapper[4978]: I0225 08:37:34.529133 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"4c2aab8b-8456-4e4c-b282-6b7074a116cc","Type":"ContainerStarted","Data":"2699cc89291b12f2fee518daf3f8115c24c110ba0bd2753e738bfe8a72b6881b"} Feb 25 08:37:34 crc kubenswrapper[4978]: I0225 08:37:34.556244 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=2.556218183 podStartE2EDuration="2.556218183s" podCreationTimestamp="2026-02-25 08:37:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:37:34.550767934 +0000 UTC m=+6747.990024453" watchObservedRunningTime="2026-02-25 08:37:34.556218183 +0000 UTC m=+6747.995474682" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.434100 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hkvw9"] Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.439221 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.460791 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-catalog-content\") pod \"redhat-marketplace-hkvw9\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.460957 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-utilities\") pod \"redhat-marketplace-hkvw9\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.461169 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm76j\" (UniqueName: \"kubernetes.io/projected/5376316c-4289-4d08-9291-0f1e6a851f88-kube-api-access-lm76j\") pod \"redhat-marketplace-hkvw9\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.470079 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkvw9"] Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.563224 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm76j\" (UniqueName: \"kubernetes.io/projected/5376316c-4289-4d08-9291-0f1e6a851f88-kube-api-access-lm76j\") pod \"redhat-marketplace-hkvw9\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.563309 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-catalog-content\") pod \"redhat-marketplace-hkvw9\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.563385 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-utilities\") pod \"redhat-marketplace-hkvw9\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.563962 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-utilities\") pod \"redhat-marketplace-hkvw9\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.564068 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-catalog-content\") pod \"redhat-marketplace-hkvw9\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.588208 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm76j\" (UniqueName: \"kubernetes.io/projected/5376316c-4289-4d08-9291-0f1e6a851f88-kube-api-access-lm76j\") pod \"redhat-marketplace-hkvw9\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.774845 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.950174 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.951783 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.959003 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:37:37 crc kubenswrapper[4978]: I0225 08:37:37.969920 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h59jd\" (UniqueName: \"kubernetes.io/projected/a54a3194-edc7-409d-99cb-823d6fba0364-kube-api-access-h59jd\") pod \"mariadb-client\" (UID: \"a54a3194-edc7-409d-99cb-823d6fba0364\") " pod="openstack/mariadb-client" Feb 25 08:37:38 crc kubenswrapper[4978]: I0225 08:37:38.064165 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkvw9"] Feb 25 08:37:38 crc kubenswrapper[4978]: W0225 08:37:38.066884 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5376316c_4289_4d08_9291_0f1e6a851f88.slice/crio-87333d6b3a6b3fa4c98d546b3b1a5103905bfc2b626e6acbddadee0fb54a2b63 WatchSource:0}: Error finding container 87333d6b3a6b3fa4c98d546b3b1a5103905bfc2b626e6acbddadee0fb54a2b63: Status 404 returned error can't find the container with id 87333d6b3a6b3fa4c98d546b3b1a5103905bfc2b626e6acbddadee0fb54a2b63 Feb 25 08:37:38 crc kubenswrapper[4978]: I0225 08:37:38.070770 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h59jd\" (UniqueName: \"kubernetes.io/projected/a54a3194-edc7-409d-99cb-823d6fba0364-kube-api-access-h59jd\") pod \"mariadb-client\" (UID: \"a54a3194-edc7-409d-99cb-823d6fba0364\") " pod="openstack/mariadb-client" Feb 25 08:37:38 crc kubenswrapper[4978]: I0225 08:37:38.088744 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h59jd\" (UniqueName: \"kubernetes.io/projected/a54a3194-edc7-409d-99cb-823d6fba0364-kube-api-access-h59jd\") pod \"mariadb-client\" (UID: \"a54a3194-edc7-409d-99cb-823d6fba0364\") " pod="openstack/mariadb-client" Feb 25 08:37:38 crc kubenswrapper[4978]: I0225 08:37:38.277343 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:37:38 crc kubenswrapper[4978]: I0225 08:37:38.533283 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:37:38 crc kubenswrapper[4978]: I0225 08:37:38.572661 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"a54a3194-edc7-409d-99cb-823d6fba0364","Type":"ContainerStarted","Data":"1f166e7647699708d337964a105c8ee921c6454ebaef0449879f2934095a9319"} Feb 25 08:37:38 crc kubenswrapper[4978]: I0225 08:37:38.575674 4978 generic.go:334] "Generic (PLEG): container finished" podID="5376316c-4289-4d08-9291-0f1e6a851f88" containerID="17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202" exitCode=0 Feb 25 08:37:38 crc kubenswrapper[4978]: I0225 08:37:38.575793 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkvw9" event={"ID":"5376316c-4289-4d08-9291-0f1e6a851f88","Type":"ContainerDied","Data":"17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202"} Feb 25 08:37:38 crc kubenswrapper[4978]: I0225 08:37:38.576891 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkvw9" event={"ID":"5376316c-4289-4d08-9291-0f1e6a851f88","Type":"ContainerStarted","Data":"87333d6b3a6b3fa4c98d546b3b1a5103905bfc2b626e6acbddadee0fb54a2b63"} Feb 25 08:37:39 crc kubenswrapper[4978]: I0225 08:37:39.588872 4978 generic.go:334] "Generic (PLEG): container finished" podID="5376316c-4289-4d08-9291-0f1e6a851f88" containerID="6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772" exitCode=0 Feb 25 08:37:39 crc kubenswrapper[4978]: I0225 08:37:39.589000 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkvw9" event={"ID":"5376316c-4289-4d08-9291-0f1e6a851f88","Type":"ContainerDied","Data":"6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772"} Feb 25 08:37:39 crc kubenswrapper[4978]: I0225 08:37:39.591038 4978 generic.go:334] "Generic (PLEG): container finished" podID="a54a3194-edc7-409d-99cb-823d6fba0364" containerID="b662c58f0f7392a0281b50089cf1dc397c5fa13a4e2d8f04351df0c98727d004" exitCode=0 Feb 25 08:37:39 crc kubenswrapper[4978]: I0225 08:37:39.591091 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"a54a3194-edc7-409d-99cb-823d6fba0364","Type":"ContainerDied","Data":"b662c58f0f7392a0281b50089cf1dc397c5fa13a4e2d8f04351df0c98727d004"} Feb 25 08:37:40 crc kubenswrapper[4978]: I0225 08:37:40.601629 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkvw9" event={"ID":"5376316c-4289-4d08-9291-0f1e6a851f88","Type":"ContainerStarted","Data":"83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6"} Feb 25 08:37:40 crc kubenswrapper[4978]: I0225 08:37:40.637112 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hkvw9" podStartSLOduration=2.250218581 podStartE2EDuration="3.637088902s" podCreationTimestamp="2026-02-25 08:37:37 +0000 UTC" firstStartedPulling="2026-02-25 08:37:38.578316473 +0000 UTC m=+6752.017572962" lastFinishedPulling="2026-02-25 08:37:39.965186784 +0000 UTC m=+6753.404443283" observedRunningTime="2026-02-25 08:37:40.628458965 +0000 UTC m=+6754.067715474" watchObservedRunningTime="2026-02-25 08:37:40.637088902 +0000 UTC m=+6754.076345371" Feb 25 08:37:40 crc kubenswrapper[4978]: I0225 08:37:40.921945 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:37:40 crc kubenswrapper[4978]: I0225 08:37:40.947881 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_a54a3194-edc7-409d-99cb-823d6fba0364/mariadb-client/0.log" Feb 25 08:37:40 crc kubenswrapper[4978]: I0225 08:37:40.984675 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:37:40 crc kubenswrapper[4978]: I0225 08:37:40.990924 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.016805 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h59jd\" (UniqueName: \"kubernetes.io/projected/a54a3194-edc7-409d-99cb-823d6fba0364-kube-api-access-h59jd\") pod \"a54a3194-edc7-409d-99cb-823d6fba0364\" (UID: \"a54a3194-edc7-409d-99cb-823d6fba0364\") " Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.022613 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a54a3194-edc7-409d-99cb-823d6fba0364-kube-api-access-h59jd" (OuterVolumeSpecName: "kube-api-access-h59jd") pod "a54a3194-edc7-409d-99cb-823d6fba0364" (UID: "a54a3194-edc7-409d-99cb-823d6fba0364"). InnerVolumeSpecName "kube-api-access-h59jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.116416 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Feb 25 08:37:41 crc kubenswrapper[4978]: E0225 08:37:41.117041 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54a3194-edc7-409d-99cb-823d6fba0364" containerName="mariadb-client" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.117060 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54a3194-edc7-409d-99cb-823d6fba0364" containerName="mariadb-client" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.117190 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a54a3194-edc7-409d-99cb-823d6fba0364" containerName="mariadb-client" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.117716 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.118408 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h59jd\" (UniqueName: \"kubernetes.io/projected/a54a3194-edc7-409d-99cb-823d6fba0364-kube-api-access-h59jd\") on node \"crc\" DevicePath \"\"" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.126151 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.220063 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hprr\" (UniqueName: \"kubernetes.io/projected/849f82d9-5232-4acf-90b4-e24517454a6d-kube-api-access-9hprr\") pod \"mariadb-client\" (UID: \"849f82d9-5232-4acf-90b4-e24517454a6d\") " pod="openstack/mariadb-client" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.321835 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hprr\" (UniqueName: \"kubernetes.io/projected/849f82d9-5232-4acf-90b4-e24517454a6d-kube-api-access-9hprr\") pod \"mariadb-client\" (UID: \"849f82d9-5232-4acf-90b4-e24517454a6d\") " pod="openstack/mariadb-client" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.335320 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a54a3194-edc7-409d-99cb-823d6fba0364" path="/var/lib/kubelet/pods/a54a3194-edc7-409d-99cb-823d6fba0364/volumes" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.350744 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hprr\" (UniqueName: \"kubernetes.io/projected/849f82d9-5232-4acf-90b4-e24517454a6d-kube-api-access-9hprr\") pod \"mariadb-client\" (UID: \"849f82d9-5232-4acf-90b4-e24517454a6d\") " pod="openstack/mariadb-client" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.441596 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.610386 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.610398 4978 scope.go:117] "RemoveContainer" containerID="b662c58f0f7392a0281b50089cf1dc397c5fa13a4e2d8f04351df0c98727d004" Feb 25 08:37:41 crc kubenswrapper[4978]: I0225 08:37:41.935126 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:37:41 crc kubenswrapper[4978]: W0225 08:37:41.941413 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod849f82d9_5232_4acf_90b4_e24517454a6d.slice/crio-42641f59ac9fb48405e6f93ac520b2c5117c6b8cf631162c8ce232f2d770dbce WatchSource:0}: Error finding container 42641f59ac9fb48405e6f93ac520b2c5117c6b8cf631162c8ce232f2d770dbce: Status 404 returned error can't find the container with id 42641f59ac9fb48405e6f93ac520b2c5117c6b8cf631162c8ce232f2d770dbce Feb 25 08:37:42 crc kubenswrapper[4978]: I0225 08:37:42.625004 4978 generic.go:334] "Generic (PLEG): container finished" podID="849f82d9-5232-4acf-90b4-e24517454a6d" containerID="16b71432cdae2d7aa2358d36a154f1cab03d2f076ca387cf97dcd7116d85126d" exitCode=0 Feb 25 08:37:42 crc kubenswrapper[4978]: I0225 08:37:42.625304 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"849f82d9-5232-4acf-90b4-e24517454a6d","Type":"ContainerDied","Data":"16b71432cdae2d7aa2358d36a154f1cab03d2f076ca387cf97dcd7116d85126d"} Feb 25 08:37:42 crc kubenswrapper[4978]: I0225 08:37:42.625438 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"849f82d9-5232-4acf-90b4-e24517454a6d","Type":"ContainerStarted","Data":"42641f59ac9fb48405e6f93ac520b2c5117c6b8cf631162c8ce232f2d770dbce"} Feb 25 08:37:44 crc kubenswrapper[4978]: I0225 08:37:44.109969 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:37:44 crc kubenswrapper[4978]: I0225 08:37:44.134154 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_849f82d9-5232-4acf-90b4-e24517454a6d/mariadb-client/0.log" Feb 25 08:37:44 crc kubenswrapper[4978]: I0225 08:37:44.171658 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:37:44 crc kubenswrapper[4978]: I0225 08:37:44.183443 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Feb 25 08:37:44 crc kubenswrapper[4978]: I0225 08:37:44.277254 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hprr\" (UniqueName: \"kubernetes.io/projected/849f82d9-5232-4acf-90b4-e24517454a6d-kube-api-access-9hprr\") pod \"849f82d9-5232-4acf-90b4-e24517454a6d\" (UID: \"849f82d9-5232-4acf-90b4-e24517454a6d\") " Feb 25 08:37:44 crc kubenswrapper[4978]: I0225 08:37:44.290733 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/849f82d9-5232-4acf-90b4-e24517454a6d-kube-api-access-9hprr" (OuterVolumeSpecName: "kube-api-access-9hprr") pod "849f82d9-5232-4acf-90b4-e24517454a6d" (UID: "849f82d9-5232-4acf-90b4-e24517454a6d"). InnerVolumeSpecName "kube-api-access-9hprr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:37:44 crc kubenswrapper[4978]: I0225 08:37:44.379738 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hprr\" (UniqueName: \"kubernetes.io/projected/849f82d9-5232-4acf-90b4-e24517454a6d-kube-api-access-9hprr\") on node \"crc\" DevicePath \"\"" Feb 25 08:37:44 crc kubenswrapper[4978]: I0225 08:37:44.645669 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42641f59ac9fb48405e6f93ac520b2c5117c6b8cf631162c8ce232f2d770dbce" Feb 25 08:37:44 crc kubenswrapper[4978]: I0225 08:37:44.645754 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Feb 25 08:37:45 crc kubenswrapper[4978]: I0225 08:37:45.351877 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="849f82d9-5232-4acf-90b4-e24517454a6d" path="/var/lib/kubelet/pods/849f82d9-5232-4acf-90b4-e24517454a6d/volumes" Feb 25 08:37:47 crc kubenswrapper[4978]: I0225 08:37:47.775591 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:47 crc kubenswrapper[4978]: I0225 08:37:47.777094 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:47 crc kubenswrapper[4978]: I0225 08:37:47.855922 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:48 crc kubenswrapper[4978]: I0225 08:37:48.766344 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:50 crc kubenswrapper[4978]: I0225 08:37:50.379066 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkvw9"] Feb 25 08:37:51 crc kubenswrapper[4978]: I0225 08:37:51.724688 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hkvw9" podUID="5376316c-4289-4d08-9291-0f1e6a851f88" containerName="registry-server" containerID="cri-o://83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6" gracePeriod=2 Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.190197 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.318355 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm76j\" (UniqueName: \"kubernetes.io/projected/5376316c-4289-4d08-9291-0f1e6a851f88-kube-api-access-lm76j\") pod \"5376316c-4289-4d08-9291-0f1e6a851f88\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.318635 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-utilities\") pod \"5376316c-4289-4d08-9291-0f1e6a851f88\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.318753 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-catalog-content\") pod \"5376316c-4289-4d08-9291-0f1e6a851f88\" (UID: \"5376316c-4289-4d08-9291-0f1e6a851f88\") " Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.320524 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-utilities" (OuterVolumeSpecName: "utilities") pod "5376316c-4289-4d08-9291-0f1e6a851f88" (UID: "5376316c-4289-4d08-9291-0f1e6a851f88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.325815 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5376316c-4289-4d08-9291-0f1e6a851f88-kube-api-access-lm76j" (OuterVolumeSpecName: "kube-api-access-lm76j") pod "5376316c-4289-4d08-9291-0f1e6a851f88" (UID: "5376316c-4289-4d08-9291-0f1e6a851f88"). InnerVolumeSpecName "kube-api-access-lm76j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.353296 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5376316c-4289-4d08-9291-0f1e6a851f88" (UID: "5376316c-4289-4d08-9291-0f1e6a851f88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.420809 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm76j\" (UniqueName: \"kubernetes.io/projected/5376316c-4289-4d08-9291-0f1e6a851f88-kube-api-access-lm76j\") on node \"crc\" DevicePath \"\"" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.420852 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.420863 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5376316c-4289-4d08-9291-0f1e6a851f88-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.738467 4978 generic.go:334] "Generic (PLEG): container finished" podID="5376316c-4289-4d08-9291-0f1e6a851f88" containerID="83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6" exitCode=0 Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.738516 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkvw9" event={"ID":"5376316c-4289-4d08-9291-0f1e6a851f88","Type":"ContainerDied","Data":"83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6"} Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.738595 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hkvw9" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.738635 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkvw9" event={"ID":"5376316c-4289-4d08-9291-0f1e6a851f88","Type":"ContainerDied","Data":"87333d6b3a6b3fa4c98d546b3b1a5103905bfc2b626e6acbddadee0fb54a2b63"} Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.738674 4978 scope.go:117] "RemoveContainer" containerID="83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.768186 4978 scope.go:117] "RemoveContainer" containerID="6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.790594 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkvw9"] Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.797111 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkvw9"] Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.813569 4978 scope.go:117] "RemoveContainer" containerID="17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.834767 4978 scope.go:117] "RemoveContainer" containerID="83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6" Feb 25 08:37:52 crc kubenswrapper[4978]: E0225 08:37:52.835259 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6\": container with ID starting with 83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6 not found: ID does not exist" containerID="83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.835290 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6"} err="failed to get container status \"83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6\": rpc error: code = NotFound desc = could not find container \"83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6\": container with ID starting with 83a89512c26aff5ed647091c304cd52b9e8f3082fa66c6593af8f2e3199b02a6 not found: ID does not exist" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.835314 4978 scope.go:117] "RemoveContainer" containerID="6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772" Feb 25 08:37:52 crc kubenswrapper[4978]: E0225 08:37:52.835641 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772\": container with ID starting with 6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772 not found: ID does not exist" containerID="6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.835702 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772"} err="failed to get container status \"6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772\": rpc error: code = NotFound desc = could not find container \"6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772\": container with ID starting with 6f041cf308632865b9f762ad28a137a7a8f1eb92b6a9997bda3f65b44b81e772 not found: ID does not exist" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.835746 4978 scope.go:117] "RemoveContainer" containerID="17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202" Feb 25 08:37:52 crc kubenswrapper[4978]: E0225 08:37:52.836193 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202\": container with ID starting with 17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202 not found: ID does not exist" containerID="17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202" Feb 25 08:37:52 crc kubenswrapper[4978]: I0225 08:37:52.836365 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202"} err="failed to get container status \"17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202\": rpc error: code = NotFound desc = could not find container \"17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202\": container with ID starting with 17184389a28c385d10d46881c1a40f28b27a7310e96845246140c5cf85a88202 not found: ID does not exist" Feb 25 08:37:53 crc kubenswrapper[4978]: I0225 08:37:53.343640 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5376316c-4289-4d08-9291-0f1e6a851f88" path="/var/lib/kubelet/pods/5376316c-4289-4d08-9291-0f1e6a851f88/volumes" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.165985 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533478-vvt88"] Feb 25 08:38:00 crc kubenswrapper[4978]: E0225 08:38:00.167004 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5376316c-4289-4d08-9291-0f1e6a851f88" containerName="extract-utilities" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.167025 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5376316c-4289-4d08-9291-0f1e6a851f88" containerName="extract-utilities" Feb 25 08:38:00 crc kubenswrapper[4978]: E0225 08:38:00.167057 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5376316c-4289-4d08-9291-0f1e6a851f88" containerName="registry-server" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.167070 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5376316c-4289-4d08-9291-0f1e6a851f88" containerName="registry-server" Feb 25 08:38:00 crc kubenswrapper[4978]: E0225 08:38:00.167093 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="849f82d9-5232-4acf-90b4-e24517454a6d" containerName="mariadb-client" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.167106 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="849f82d9-5232-4acf-90b4-e24517454a6d" containerName="mariadb-client" Feb 25 08:38:00 crc kubenswrapper[4978]: E0225 08:38:00.167140 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5376316c-4289-4d08-9291-0f1e6a851f88" containerName="extract-content" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.167152 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5376316c-4289-4d08-9291-0f1e6a851f88" containerName="extract-content" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.167438 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5376316c-4289-4d08-9291-0f1e6a851f88" containerName="registry-server" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.167460 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="849f82d9-5232-4acf-90b4-e24517454a6d" containerName="mariadb-client" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.168256 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533478-vvt88" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.170549 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.171596 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.171949 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.188843 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533478-vvt88"] Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.361209 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zmf2\" (UniqueName: \"kubernetes.io/projected/7fbf1727-1311-44ae-a207-0e856f22c017-kube-api-access-8zmf2\") pod \"auto-csr-approver-29533478-vvt88\" (UID: \"7fbf1727-1311-44ae-a207-0e856f22c017\") " pod="openshift-infra/auto-csr-approver-29533478-vvt88" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.463243 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zmf2\" (UniqueName: \"kubernetes.io/projected/7fbf1727-1311-44ae-a207-0e856f22c017-kube-api-access-8zmf2\") pod \"auto-csr-approver-29533478-vvt88\" (UID: \"7fbf1727-1311-44ae-a207-0e856f22c017\") " pod="openshift-infra/auto-csr-approver-29533478-vvt88" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.495817 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zmf2\" (UniqueName: \"kubernetes.io/projected/7fbf1727-1311-44ae-a207-0e856f22c017-kube-api-access-8zmf2\") pod \"auto-csr-approver-29533478-vvt88\" (UID: \"7fbf1727-1311-44ae-a207-0e856f22c017\") " pod="openshift-infra/auto-csr-approver-29533478-vvt88" Feb 25 08:38:00 crc kubenswrapper[4978]: I0225 08:38:00.531238 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533478-vvt88" Feb 25 08:38:01 crc kubenswrapper[4978]: I0225 08:38:01.003682 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533478-vvt88"] Feb 25 08:38:01 crc kubenswrapper[4978]: I0225 08:38:01.818989 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533478-vvt88" event={"ID":"7fbf1727-1311-44ae-a207-0e856f22c017","Type":"ContainerStarted","Data":"1bf969854698219a32f1949ca18fcee1b19228243e1797d7c63ce598bca3da0a"} Feb 25 08:38:02 crc kubenswrapper[4978]: I0225 08:38:02.827940 4978 generic.go:334] "Generic (PLEG): container finished" podID="7fbf1727-1311-44ae-a207-0e856f22c017" containerID="f409e122243cfd8dbfca58352923255a229ee32e8796f4f38d263375c901945e" exitCode=0 Feb 25 08:38:02 crc kubenswrapper[4978]: I0225 08:38:02.828058 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533478-vvt88" event={"ID":"7fbf1727-1311-44ae-a207-0e856f22c017","Type":"ContainerDied","Data":"f409e122243cfd8dbfca58352923255a229ee32e8796f4f38d263375c901945e"} Feb 25 08:38:04 crc kubenswrapper[4978]: I0225 08:38:04.227202 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533478-vvt88" Feb 25 08:38:04 crc kubenswrapper[4978]: I0225 08:38:04.375137 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zmf2\" (UniqueName: \"kubernetes.io/projected/7fbf1727-1311-44ae-a207-0e856f22c017-kube-api-access-8zmf2\") pod \"7fbf1727-1311-44ae-a207-0e856f22c017\" (UID: \"7fbf1727-1311-44ae-a207-0e856f22c017\") " Feb 25 08:38:04 crc kubenswrapper[4978]: I0225 08:38:04.384022 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fbf1727-1311-44ae-a207-0e856f22c017-kube-api-access-8zmf2" (OuterVolumeSpecName: "kube-api-access-8zmf2") pod "7fbf1727-1311-44ae-a207-0e856f22c017" (UID: "7fbf1727-1311-44ae-a207-0e856f22c017"). InnerVolumeSpecName "kube-api-access-8zmf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:38:04 crc kubenswrapper[4978]: I0225 08:38:04.476997 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zmf2\" (UniqueName: \"kubernetes.io/projected/7fbf1727-1311-44ae-a207-0e856f22c017-kube-api-access-8zmf2\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:04 crc kubenswrapper[4978]: I0225 08:38:04.850552 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533478-vvt88" event={"ID":"7fbf1727-1311-44ae-a207-0e856f22c017","Type":"ContainerDied","Data":"1bf969854698219a32f1949ca18fcee1b19228243e1797d7c63ce598bca3da0a"} Feb 25 08:38:04 crc kubenswrapper[4978]: I0225 08:38:04.850988 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bf969854698219a32f1949ca18fcee1b19228243e1797d7c63ce598bca3da0a" Feb 25 08:38:04 crc kubenswrapper[4978]: I0225 08:38:04.851049 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533478-vvt88" Feb 25 08:38:05 crc kubenswrapper[4978]: I0225 08:38:05.346517 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533472-8c5nh"] Feb 25 08:38:05 crc kubenswrapper[4978]: I0225 08:38:05.355899 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533472-8c5nh"] Feb 25 08:38:07 crc kubenswrapper[4978]: I0225 08:38:07.336136 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1999d0d3-0e65-45ea-8133-f6b40154c125" path="/var/lib/kubelet/pods/1999d0d3-0e65-45ea-8133-f6b40154c125/volumes" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.473664 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 08:38:15 crc kubenswrapper[4978]: E0225 08:38:15.474612 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fbf1727-1311-44ae-a207-0e856f22c017" containerName="oc" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.474631 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fbf1727-1311-44ae-a207-0e856f22c017" containerName="oc" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.474847 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fbf1727-1311-44ae-a207-0e856f22c017" containerName="oc" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.476261 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.480493 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-t6ck6" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.480787 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.481009 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.481157 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.485924 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.487665 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.488577 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.504432 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.520118 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.521765 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.535947 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564188 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564248 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e21d9172-38ac-4816-b77d-db144cdfd65b-config\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564328 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564361 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mhkn\" (UniqueName: \"kubernetes.io/projected/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-kube-api-access-8mhkn\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564406 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-config\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564461 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e21d9172-38ac-4816-b77d-db144cdfd65b-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564506 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-765346ea-d307-4617-a231-2213e2bbe605\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-765346ea-d307-4617-a231-2213e2bbe605\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564530 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8gwr\" (UniqueName: \"kubernetes.io/projected/e21d9172-38ac-4816-b77d-db144cdfd65b-kube-api-access-k8gwr\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564568 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e21d9172-38ac-4816-b77d-db144cdfd65b-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564602 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564655 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e21d9172-38ac-4816-b77d-db144cdfd65b-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564702 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564734 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564764 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21d9172-38ac-4816-b77d-db144cdfd65b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564831 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21d9172-38ac-4816-b77d-db144cdfd65b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.564872 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f753909c-ec37-443a-a843-a48c68c72093\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f753909c-ec37-443a-a843-a48c68c72093\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.576788 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666397 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666478 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e21d9172-38ac-4816-b77d-db144cdfd65b-config\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666522 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666560 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mhkn\" (UniqueName: \"kubernetes.io/projected/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-kube-api-access-8mhkn\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666591 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-config\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666653 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e21d9172-38ac-4816-b77d-db144cdfd65b-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666716 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8621c881-a82f-4922-9e91-43422305fcea-config\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666749 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvvvq\" (UniqueName: \"kubernetes.io/projected/8621c881-a82f-4922-9e91-43422305fcea-kube-api-access-wvvvq\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666785 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8621c881-a82f-4922-9e91-43422305fcea-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666815 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8621c881-a82f-4922-9e91-43422305fcea-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666847 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8621c881-a82f-4922-9e91-43422305fcea-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666887 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-765346ea-d307-4617-a231-2213e2bbe605\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-765346ea-d307-4617-a231-2213e2bbe605\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666922 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8gwr\" (UniqueName: \"kubernetes.io/projected/e21d9172-38ac-4816-b77d-db144cdfd65b-kube-api-access-k8gwr\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.666953 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8621c881-a82f-4922-9e91-43422305fcea-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667001 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0b74e929-e2c3-4f8d-a76b-f89249d6a43a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b74e929-e2c3-4f8d-a76b-f89249d6a43a\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667044 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e21d9172-38ac-4816-b77d-db144cdfd65b-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667095 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667136 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e21d9172-38ac-4816-b77d-db144cdfd65b-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667179 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667224 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667241 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667292 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21d9172-38ac-4816-b77d-db144cdfd65b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667356 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21d9172-38ac-4816-b77d-db144cdfd65b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667488 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8621c881-a82f-4922-9e91-43422305fcea-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667536 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f753909c-ec37-443a-a843-a48c68c72093\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f753909c-ec37-443a-a843-a48c68c72093\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.667928 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e21d9172-38ac-4816-b77d-db144cdfd65b-config\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.668502 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-config\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.668898 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e21d9172-38ac-4816-b77d-db144cdfd65b-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.670046 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e21d9172-38ac-4816-b77d-db144cdfd65b-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.673614 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.673636 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.673652 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.673644 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-765346ea-d307-4617-a231-2213e2bbe605\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-765346ea-d307-4617-a231-2213e2bbe605\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0646fc19f56e6d753d7deb378299187c0c39d0d00e2b77b719314d1b774b7fb5/globalmount\"" pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.673686 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f753909c-ec37-443a-a843-a48c68c72093\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f753909c-ec37-443a-a843-a48c68c72093\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/04f357bcea7bc71de05a87a32824e8cbe8f897c50be21ade5fc94daeb56d6e5c/globalmount\"" pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.674752 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.675284 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.681650 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e21d9172-38ac-4816-b77d-db144cdfd65b-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.682043 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.687329 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21d9172-38ac-4816-b77d-db144cdfd65b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.687678 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8gwr\" (UniqueName: \"kubernetes.io/projected/e21d9172-38ac-4816-b77d-db144cdfd65b-kube-api-access-k8gwr\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.688966 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mhkn\" (UniqueName: \"kubernetes.io/projected/99c42bcd-7266-4cf3-a7c4-b8db18645a4f-kube-api-access-8mhkn\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.698988 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e21d9172-38ac-4816-b77d-db144cdfd65b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.709411 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-765346ea-d307-4617-a231-2213e2bbe605\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-765346ea-d307-4617-a231-2213e2bbe605\") pod \"ovsdbserver-sb-1\" (UID: \"e21d9172-38ac-4816-b77d-db144cdfd65b\") " pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.709937 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f753909c-ec37-443a-a843-a48c68c72093\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f753909c-ec37-443a-a843-a48c68c72093\") pod \"ovsdbserver-sb-0\" (UID: \"99c42bcd-7266-4cf3-a7c4-b8db18645a4f\") " pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.769041 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8621c881-a82f-4922-9e91-43422305fcea-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.769238 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8621c881-a82f-4922-9e91-43422305fcea-config\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.769292 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvvvq\" (UniqueName: \"kubernetes.io/projected/8621c881-a82f-4922-9e91-43422305fcea-kube-api-access-wvvvq\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.769504 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8621c881-a82f-4922-9e91-43422305fcea-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.770840 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8621c881-a82f-4922-9e91-43422305fcea-config\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.770931 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8621c881-a82f-4922-9e91-43422305fcea-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.770995 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8621c881-a82f-4922-9e91-43422305fcea-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.771505 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8621c881-a82f-4922-9e91-43422305fcea-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.777458 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8621c881-a82f-4922-9e91-43422305fcea-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.777545 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8621c881-a82f-4922-9e91-43422305fcea-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.777627 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8621c881-a82f-4922-9e91-43422305fcea-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.777699 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0b74e929-e2c3-4f8d-a76b-f89249d6a43a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b74e929-e2c3-4f8d-a76b-f89249d6a43a\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.777769 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8621c881-a82f-4922-9e91-43422305fcea-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.781707 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.781742 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0b74e929-e2c3-4f8d-a76b-f89249d6a43a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b74e929-e2c3-4f8d-a76b-f89249d6a43a\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/81f49849315da0d848cb682cefa4d57e9a4fec4ff042ddbb17f6c36205adfaf0/globalmount\"" pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.784125 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8621c881-a82f-4922-9e91-43422305fcea-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.788619 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvvvq\" (UniqueName: \"kubernetes.io/projected/8621c881-a82f-4922-9e91-43422305fcea-kube-api-access-wvvvq\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.808855 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.819962 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.834485 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0b74e929-e2c3-4f8d-a76b-f89249d6a43a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b74e929-e2c3-4f8d-a76b-f89249d6a43a\") pod \"ovsdbserver-sb-2\" (UID: \"8621c881-a82f-4922-9e91-43422305fcea\") " pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:15 crc kubenswrapper[4978]: I0225 08:38:15.847048 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:16 crc kubenswrapper[4978]: I0225 08:38:16.431852 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Feb 25 08:38:16 crc kubenswrapper[4978]: W0225 08:38:16.433246 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8621c881_a82f_4922_9e91_43422305fcea.slice/crio-5bec3fe5282515ac190e68f0ed08513427fc492929a70c04035f1bae3c39c42e WatchSource:0}: Error finding container 5bec3fe5282515ac190e68f0ed08513427fc492929a70c04035f1bae3c39c42e: Status 404 returned error can't find the container with id 5bec3fe5282515ac190e68f0ed08513427fc492929a70c04035f1bae3c39c42e Feb 25 08:38:16 crc kubenswrapper[4978]: I0225 08:38:16.517403 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 25 08:38:16 crc kubenswrapper[4978]: W0225 08:38:16.522126 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99c42bcd_7266_4cf3_a7c4_b8db18645a4f.slice/crio-9524367d1dd84d478ab038ac1ae949a56cdefdb5c796cea2b30d7e65b8cbd4c4 WatchSource:0}: Error finding container 9524367d1dd84d478ab038ac1ae949a56cdefdb5c796cea2b30d7e65b8cbd4c4: Status 404 returned error can't find the container with id 9524367d1dd84d478ab038ac1ae949a56cdefdb5c796cea2b30d7e65b8cbd4c4 Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.137648 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"99c42bcd-7266-4cf3-a7c4-b8db18645a4f","Type":"ContainerStarted","Data":"9524367d1dd84d478ab038ac1ae949a56cdefdb5c796cea2b30d7e65b8cbd4c4"} Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.140856 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"8621c881-a82f-4922-9e91-43422305fcea","Type":"ContainerStarted","Data":"5bec3fe5282515ac190e68f0ed08513427fc492929a70c04035f1bae3c39c42e"} Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.619112 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.682020 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.683224 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.692644 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-4lg22" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.692711 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.692801 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.692837 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.732125 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.733451 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.745436 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.747188 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.753814 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.765323 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.779951 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821183 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f671c5-0151-4f3d-b919-e5c9ac280038-config\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821227 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9119e44-2b2c-479b-a465-405ec797b89d-config\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821246 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f671c5-0151-4f3d-b919-e5c9ac280038-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821272 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9119e44-2b2c-479b-a465-405ec797b89d-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821299 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f671c5-0151-4f3d-b919-e5c9ac280038-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821324 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6b47\" (UniqueName: \"kubernetes.io/projected/aa16e3af-ea25-4a1c-83d1-3300ab386437-kube-api-access-r6b47\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821346 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9119e44-2b2c-479b-a465-405ec797b89d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821381 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa16e3af-ea25-4a1c-83d1-3300ab386437-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821407 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f671c5-0151-4f3d-b919-e5c9ac280038-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821435 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/aa16e3af-ea25-4a1c-83d1-3300ab386437-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821647 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa16e3af-ea25-4a1c-83d1-3300ab386437-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821810 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa16e3af-ea25-4a1c-83d1-3300ab386437-config\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821886 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa16e3af-ea25-4a1c-83d1-3300ab386437-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821913 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjwqm\" (UniqueName: \"kubernetes.io/projected/a9119e44-2b2c-479b-a465-405ec797b89d-kube-api-access-gjwqm\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821950 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9119e44-2b2c-479b-a465-405ec797b89d-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.821971 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c629e70e-2cd6-4ada-82d8-5c18aeda6c7f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c629e70e-2cd6-4ada-82d8-5c18aeda6c7f\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.822016 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5806c7d4-402d-43b2-ab20-b33c3b975162\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5806c7d4-402d-43b2-ab20-b33c3b975162\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.823946 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9119e44-2b2c-479b-a465-405ec797b89d-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.824056 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa16e3af-ea25-4a1c-83d1-3300ab386437-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.824108 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7f671c5-0151-4f3d-b919-e5c9ac280038-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.824130 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a7f671c5-0151-4f3d-b919-e5c9ac280038-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.824154 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srgdb\" (UniqueName: \"kubernetes.io/projected/a7f671c5-0151-4f3d-b919-e5c9ac280038-kube-api-access-srgdb\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.824196 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9119e44-2b2c-479b-a465-405ec797b89d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.824314 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-74687da8-0b1e-4df4-b1d8-447be467489f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74687da8-0b1e-4df4-b1d8-447be467489f\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.926978 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9119e44-2b2c-479b-a465-405ec797b89d-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927024 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f671c5-0151-4f3d-b919-e5c9ac280038-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927041 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6b47\" (UniqueName: \"kubernetes.io/projected/aa16e3af-ea25-4a1c-83d1-3300ab386437-kube-api-access-r6b47\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927061 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9119e44-2b2c-479b-a465-405ec797b89d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927075 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa16e3af-ea25-4a1c-83d1-3300ab386437-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927095 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f671c5-0151-4f3d-b919-e5c9ac280038-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927117 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/aa16e3af-ea25-4a1c-83d1-3300ab386437-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927136 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa16e3af-ea25-4a1c-83d1-3300ab386437-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927152 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa16e3af-ea25-4a1c-83d1-3300ab386437-config\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927169 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa16e3af-ea25-4a1c-83d1-3300ab386437-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927185 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjwqm\" (UniqueName: \"kubernetes.io/projected/a9119e44-2b2c-479b-a465-405ec797b89d-kube-api-access-gjwqm\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927200 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9119e44-2b2c-479b-a465-405ec797b89d-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.927964 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a9119e44-2b2c-479b-a465-405ec797b89d-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928228 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c629e70e-2cd6-4ada-82d8-5c18aeda6c7f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c629e70e-2cd6-4ada-82d8-5c18aeda6c7f\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928271 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5806c7d4-402d-43b2-ab20-b33c3b975162\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5806c7d4-402d-43b2-ab20-b33c3b975162\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928296 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9119e44-2b2c-479b-a465-405ec797b89d-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928326 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa16e3af-ea25-4a1c-83d1-3300ab386437-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928342 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7f671c5-0151-4f3d-b919-e5c9ac280038-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928354 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a7f671c5-0151-4f3d-b919-e5c9ac280038-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928386 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srgdb\" (UniqueName: \"kubernetes.io/projected/a7f671c5-0151-4f3d-b919-e5c9ac280038-kube-api-access-srgdb\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928402 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9119e44-2b2c-479b-a465-405ec797b89d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928438 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-74687da8-0b1e-4df4-b1d8-447be467489f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74687da8-0b1e-4df4-b1d8-447be467489f\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928484 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f671c5-0151-4f3d-b919-e5c9ac280038-config\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928506 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9119e44-2b2c-479b-a465-405ec797b89d-config\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.928524 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f671c5-0151-4f3d-b919-e5c9ac280038-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.933245 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f671c5-0151-4f3d-b919-e5c9ac280038-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.934002 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7f671c5-0151-4f3d-b919-e5c9ac280038-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.935148 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7f671c5-0151-4f3d-b919-e5c9ac280038-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.936654 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa16e3af-ea25-4a1c-83d1-3300ab386437-config\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.936899 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa16e3af-ea25-4a1c-83d1-3300ab386437-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.937126 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a7f671c5-0151-4f3d-b919-e5c9ac280038-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.937838 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7f671c5-0151-4f3d-b919-e5c9ac280038-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.938260 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/aa16e3af-ea25-4a1c-83d1-3300ab386437-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.938598 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9119e44-2b2c-479b-a465-405ec797b89d-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.938918 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f671c5-0151-4f3d-b919-e5c9ac280038-config\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.938990 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa16e3af-ea25-4a1c-83d1-3300ab386437-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.939320 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.939349 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c629e70e-2cd6-4ada-82d8-5c18aeda6c7f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c629e70e-2cd6-4ada-82d8-5c18aeda6c7f\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e4ce742a93da7528ac397d5bb250a5a9fb521ca3121b225e6d132539a4e8b278/globalmount\"" pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.939518 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.939542 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5806c7d4-402d-43b2-ab20-b33c3b975162\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5806c7d4-402d-43b2-ab20-b33c3b975162\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/42c8e6993cee3983c823f6ac416781acaed774bc072a5e7f7bade1eaf4993a48/globalmount\"" pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.939908 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9119e44-2b2c-479b-a465-405ec797b89d-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.940730 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.940890 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-74687da8-0b1e-4df4-b1d8-447be467489f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74687da8-0b1e-4df4-b1d8-447be467489f\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fa79714ec5deb7ade6486bf49b1f3bfa09d66f8681dde54b099299c65efc6fb3/globalmount\"" pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.943540 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa16e3af-ea25-4a1c-83d1-3300ab386437-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.944979 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9119e44-2b2c-479b-a465-405ec797b89d-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.946291 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa16e3af-ea25-4a1c-83d1-3300ab386437-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.946627 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a9119e44-2b2c-479b-a465-405ec797b89d-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.947252 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6b47\" (UniqueName: \"kubernetes.io/projected/aa16e3af-ea25-4a1c-83d1-3300ab386437-kube-api-access-r6b47\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.947446 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9119e44-2b2c-479b-a465-405ec797b89d-config\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.954900 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srgdb\" (UniqueName: \"kubernetes.io/projected/a7f671c5-0151-4f3d-b919-e5c9ac280038-kube-api-access-srgdb\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.958601 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjwqm\" (UniqueName: \"kubernetes.io/projected/a9119e44-2b2c-479b-a465-405ec797b89d-kube-api-access-gjwqm\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.970187 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5806c7d4-402d-43b2-ab20-b33c3b975162\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5806c7d4-402d-43b2-ab20-b33c3b975162\") pod \"ovsdbserver-nb-0\" (UID: \"a7f671c5-0151-4f3d-b919-e5c9ac280038\") " pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.976026 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-74687da8-0b1e-4df4-b1d8-447be467489f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-74687da8-0b1e-4df4-b1d8-447be467489f\") pod \"ovsdbserver-nb-2\" (UID: \"a9119e44-2b2c-479b-a465-405ec797b89d\") " pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:17 crc kubenswrapper[4978]: I0225 08:38:17.986476 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c629e70e-2cd6-4ada-82d8-5c18aeda6c7f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c629e70e-2cd6-4ada-82d8-5c18aeda6c7f\") pod \"ovsdbserver-nb-1\" (UID: \"aa16e3af-ea25-4a1c-83d1-3300ab386437\") " pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:18 crc kubenswrapper[4978]: I0225 08:38:18.026607 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:18 crc kubenswrapper[4978]: I0225 08:38:18.065918 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:18 crc kubenswrapper[4978]: I0225 08:38:18.082239 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:18 crc kubenswrapper[4978]: I0225 08:38:18.170786 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e21d9172-38ac-4816-b77d-db144cdfd65b","Type":"ContainerStarted","Data":"0e80a9e51d746b17013672ab2eb68fd68d5af363027e2f5115eb862ba1f1a1b6"} Feb 25 08:38:18 crc kubenswrapper[4978]: I0225 08:38:18.617499 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 25 08:38:18 crc kubenswrapper[4978]: I0225 08:38:18.720060 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Feb 25 08:38:19 crc kubenswrapper[4978]: I0225 08:38:19.618574 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Feb 25 08:38:20 crc kubenswrapper[4978]: W0225 08:38:20.089734 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa16e3af_ea25_4a1c_83d1_3300ab386437.slice/crio-b1cf5e91bafc2cf3382bf14b5ecfde5e1002c4579a9443f34f1a20f1b02a7d3f WatchSource:0}: Error finding container b1cf5e91bafc2cf3382bf14b5ecfde5e1002c4579a9443f34f1a20f1b02a7d3f: Status 404 returned error can't find the container with id b1cf5e91bafc2cf3382bf14b5ecfde5e1002c4579a9443f34f1a20f1b02a7d3f Feb 25 08:38:20 crc kubenswrapper[4978]: I0225 08:38:20.219518 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"aa16e3af-ea25-4a1c-83d1-3300ab386437","Type":"ContainerStarted","Data":"b1cf5e91bafc2cf3382bf14b5ecfde5e1002c4579a9443f34f1a20f1b02a7d3f"} Feb 25 08:38:20 crc kubenswrapper[4978]: I0225 08:38:20.229529 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"a9119e44-2b2c-479b-a465-405ec797b89d","Type":"ContainerStarted","Data":"d9b1daf88d7e4e7227518affedbcda97288297a27fd8b43778873549af5a61bd"} Feb 25 08:38:20 crc kubenswrapper[4978]: I0225 08:38:20.247494 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a7f671c5-0151-4f3d-b919-e5c9ac280038","Type":"ContainerStarted","Data":"9c2b0fbae7ba219353acbee0bda176c2d0a1e70c171a03307aa4bbe1f57ecf46"} Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.283224 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"99c42bcd-7266-4cf3-a7c4-b8db18645a4f","Type":"ContainerStarted","Data":"18a465ca99ce18cd1833a5ec44e73bb70a97d537baac3d0c4a949ca4b601093f"} Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.283925 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"99c42bcd-7266-4cf3-a7c4-b8db18645a4f","Type":"ContainerStarted","Data":"e45ebb226951217d64d5583263ed81b716e69d3201b41e9d83e43868604f3c30"} Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.289632 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"8621c881-a82f-4922-9e91-43422305fcea","Type":"ContainerStarted","Data":"0feb88d643392f4dd06f2890ba18e669f55ff743f567c73792df590b016ac6b9"} Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.289692 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"8621c881-a82f-4922-9e91-43422305fcea","Type":"ContainerStarted","Data":"fc87f987f051bc5552cc014c59ecb2ff54d165fa19fd4187da741bc628400745"} Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.293478 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e21d9172-38ac-4816-b77d-db144cdfd65b","Type":"ContainerStarted","Data":"4253aae89b97b4c06c6364446837276f0afb665af0fe9dda9710b5cc9fe20890"} Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.293538 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e21d9172-38ac-4816-b77d-db144cdfd65b","Type":"ContainerStarted","Data":"0667797784c4ee62d7af7059c51868a941564674baea72b3f7029985e1806018"} Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.303874 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.5269535960000002 podStartE2EDuration="7.303848465s" podCreationTimestamp="2026-02-25 08:38:14 +0000 UTC" firstStartedPulling="2026-02-25 08:38:16.529272218 +0000 UTC m=+6789.968528677" lastFinishedPulling="2026-02-25 08:38:20.306167087 +0000 UTC m=+6793.745423546" observedRunningTime="2026-02-25 08:38:21.301281065 +0000 UTC m=+6794.740537564" watchObservedRunningTime="2026-02-25 08:38:21.303848465 +0000 UTC m=+6794.743104964" Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.329608 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.6429642300000005 podStartE2EDuration="7.329591993s" podCreationTimestamp="2026-02-25 08:38:14 +0000 UTC" firstStartedPulling="2026-02-25 08:38:17.621426534 +0000 UTC m=+6791.060682993" lastFinishedPulling="2026-02-25 08:38:20.308054297 +0000 UTC m=+6793.747310756" observedRunningTime="2026-02-25 08:38:21.327940851 +0000 UTC m=+6794.767197320" watchObservedRunningTime="2026-02-25 08:38:21.329591993 +0000 UTC m=+6794.768848442" Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.348245 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=3.462021642 podStartE2EDuration="7.34821703s" podCreationTimestamp="2026-02-25 08:38:14 +0000 UTC" firstStartedPulling="2026-02-25 08:38:16.435642746 +0000 UTC m=+6789.874899205" lastFinishedPulling="2026-02-25 08:38:20.321838134 +0000 UTC m=+6793.761094593" observedRunningTime="2026-02-25 08:38:21.3436966 +0000 UTC m=+6794.782953059" watchObservedRunningTime="2026-02-25 08:38:21.34821703 +0000 UTC m=+6794.787473489" Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.810144 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.823233 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:21 crc kubenswrapper[4978]: I0225 08:38:21.847988 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:22 crc kubenswrapper[4978]: I0225 08:38:22.307678 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"a9119e44-2b2c-479b-a465-405ec797b89d","Type":"ContainerStarted","Data":"dbf9c967a2c177977bb51ed94d72b4a9f0c4d3a669e1748f54fc5b316d9951d9"} Feb 25 08:38:22 crc kubenswrapper[4978]: I0225 08:38:22.307751 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"a9119e44-2b2c-479b-a465-405ec797b89d","Type":"ContainerStarted","Data":"ec07a80e55a193f7aa4665b03caaebdd2eaf60f88fa6b6ef6d40c584b27e5b0a"} Feb 25 08:38:22 crc kubenswrapper[4978]: I0225 08:38:22.311793 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a7f671c5-0151-4f3d-b919-e5c9ac280038","Type":"ContainerStarted","Data":"f481463c1807ed38fdb8d7b6e034b4b1994204581afb69dad4ce9d2251404db3"} Feb 25 08:38:22 crc kubenswrapper[4978]: I0225 08:38:22.311878 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a7f671c5-0151-4f3d-b919-e5c9ac280038","Type":"ContainerStarted","Data":"428fbbf931bb90e5388c1b326a915d804b3645cad72fe76089c7a95da6699a76"} Feb 25 08:38:22 crc kubenswrapper[4978]: I0225 08:38:22.314907 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"aa16e3af-ea25-4a1c-83d1-3300ab386437","Type":"ContainerStarted","Data":"79e3f74a29ee50ba74c4daac0a8f21ee31f7de643a95f222287356446b1f50cd"} Feb 25 08:38:22 crc kubenswrapper[4978]: I0225 08:38:22.314974 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"aa16e3af-ea25-4a1c-83d1-3300ab386437","Type":"ContainerStarted","Data":"48db4f85165d630329800ded754724068a481b08312cf046c28b46bdb4bbd690"} Feb 25 08:38:22 crc kubenswrapper[4978]: I0225 08:38:22.341833 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=5.561549443 podStartE2EDuration="6.34181374s" podCreationTimestamp="2026-02-25 08:38:16 +0000 UTC" firstStartedPulling="2026-02-25 08:38:20.229612735 +0000 UTC m=+6793.668869194" lastFinishedPulling="2026-02-25 08:38:21.009877012 +0000 UTC m=+6794.449133491" observedRunningTime="2026-02-25 08:38:22.338383144 +0000 UTC m=+6795.777639613" watchObservedRunningTime="2026-02-25 08:38:22.34181374 +0000 UTC m=+6795.781070209" Feb 25 08:38:22 crc kubenswrapper[4978]: I0225 08:38:22.380159 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=5.443045739 podStartE2EDuration="6.380130568s" podCreationTimestamp="2026-02-25 08:38:16 +0000 UTC" firstStartedPulling="2026-02-25 08:38:20.093957319 +0000 UTC m=+6793.533213818" lastFinishedPulling="2026-02-25 08:38:21.031042188 +0000 UTC m=+6794.470298647" observedRunningTime="2026-02-25 08:38:22.366591228 +0000 UTC m=+6795.805847767" watchObservedRunningTime="2026-02-25 08:38:22.380130568 +0000 UTC m=+6795.819387037" Feb 25 08:38:22 crc kubenswrapper[4978]: I0225 08:38:22.406489 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.431998097 podStartE2EDuration="6.406464804s" podCreationTimestamp="2026-02-25 08:38:16 +0000 UTC" firstStartedPulling="2026-02-25 08:38:20.094576999 +0000 UTC m=+6793.533833458" lastFinishedPulling="2026-02-25 08:38:21.069043706 +0000 UTC m=+6794.508300165" observedRunningTime="2026-02-25 08:38:22.396960129 +0000 UTC m=+6795.836216628" watchObservedRunningTime="2026-02-25 08:38:22.406464804 +0000 UTC m=+6795.845721273" Feb 25 08:38:23 crc kubenswrapper[4978]: I0225 08:38:23.027485 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:23 crc kubenswrapper[4978]: I0225 08:38:23.066988 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:23 crc kubenswrapper[4978]: I0225 08:38:23.083497 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.027924 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.067138 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.071798 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.083554 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.135266 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.138962 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.846511 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.847060 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.860464 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.860749 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.891030 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:24 crc kubenswrapper[4978]: I0225 08:38:24.891448 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.417698 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.438652 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.441287 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.627572 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-697c57b9b5-sf2b5"] Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.628977 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.647349 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.647483 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-697c57b9b5-sf2b5"] Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.773832 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-config\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.773928 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cxc8\" (UniqueName: \"kubernetes.io/projected/a20030e2-d291-44f8-aaa4-6b33d5e11122-kube-api-access-2cxc8\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.774030 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-ovsdbserver-sb\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.774088 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-dns-svc\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.875821 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-config\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.876312 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cxc8\" (UniqueName: \"kubernetes.io/projected/a20030e2-d291-44f8-aaa4-6b33d5e11122-kube-api-access-2cxc8\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.876359 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-ovsdbserver-sb\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.876427 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-dns-svc\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.877441 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-dns-svc\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.877549 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-config\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.877964 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-ovsdbserver-sb\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.899022 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cxc8\" (UniqueName: \"kubernetes.io/projected/a20030e2-d291-44f8-aaa4-6b33d5e11122-kube-api-access-2cxc8\") pod \"dnsmasq-dns-697c57b9b5-sf2b5\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:25 crc kubenswrapper[4978]: I0225 08:38:25.967620 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:26 crc kubenswrapper[4978]: I0225 08:38:26.212110 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-697c57b9b5-sf2b5"] Feb 25 08:38:26 crc kubenswrapper[4978]: I0225 08:38:26.359483 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" event={"ID":"a20030e2-d291-44f8-aaa4-6b33d5e11122","Type":"ContainerStarted","Data":"8262b3075f545ffb8f6c549b40f56cb502783535cfcf58b440f2ea7831470697"} Feb 25 08:38:27 crc kubenswrapper[4978]: I0225 08:38:27.337890 4978 scope.go:117] "RemoveContainer" containerID="b6d2f778106fa1a2a105312237f63ef0c91f2cd87fa098e690c33f0a3335cc52" Feb 25 08:38:27 crc kubenswrapper[4978]: I0225 08:38:27.382725 4978 generic.go:334] "Generic (PLEG): container finished" podID="a20030e2-d291-44f8-aaa4-6b33d5e11122" containerID="dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a" exitCode=0 Feb 25 08:38:27 crc kubenswrapper[4978]: I0225 08:38:27.382804 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" event={"ID":"a20030e2-d291-44f8-aaa4-6b33d5e11122","Type":"ContainerDied","Data":"dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a"} Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.094993 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.138595 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.149737 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.344021 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-697c57b9b5-sf2b5"] Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.388802 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbc4875f7-rfcdw"] Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.390838 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" event={"ID":"a20030e2-d291-44f8-aaa4-6b33d5e11122","Type":"ContainerStarted","Data":"c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e"} Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.390908 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.391391 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.392864 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.406336 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbc4875f7-rfcdw"] Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.454714 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" podStartSLOduration=3.454687173 podStartE2EDuration="3.454687173s" podCreationTimestamp="2026-02-25 08:38:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:38:28.446239281 +0000 UTC m=+6801.885495730" watchObservedRunningTime="2026-02-25 08:38:28.454687173 +0000 UTC m=+6801.893943652" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.525712 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrqg2\" (UniqueName: \"kubernetes.io/projected/13e7f32e-2be0-476f-9043-9446462ae7df-kube-api-access-vrqg2\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.526231 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-sb\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.526359 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-config\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.526521 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-dns-svc\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.526625 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-nb\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.629148 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrqg2\" (UniqueName: \"kubernetes.io/projected/13e7f32e-2be0-476f-9043-9446462ae7df-kube-api-access-vrqg2\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.629235 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-sb\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.629295 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-config\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.629351 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-dns-svc\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.629438 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-nb\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.630255 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-config\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.630282 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-dns-svc\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.630885 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-sb\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.630962 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-nb\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.662424 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrqg2\" (UniqueName: \"kubernetes.io/projected/13e7f32e-2be0-476f-9043-9446462ae7df-kube-api-access-vrqg2\") pod \"dnsmasq-dns-bbc4875f7-rfcdw\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:28 crc kubenswrapper[4978]: I0225 08:38:28.707265 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:29 crc kubenswrapper[4978]: I0225 08:38:29.027162 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbc4875f7-rfcdw"] Feb 25 08:38:29 crc kubenswrapper[4978]: I0225 08:38:29.400808 4978 generic.go:334] "Generic (PLEG): container finished" podID="13e7f32e-2be0-476f-9043-9446462ae7df" containerID="a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d" exitCode=0 Feb 25 08:38:29 crc kubenswrapper[4978]: I0225 08:38:29.400891 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" event={"ID":"13e7f32e-2be0-476f-9043-9446462ae7df","Type":"ContainerDied","Data":"a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d"} Feb 25 08:38:29 crc kubenswrapper[4978]: I0225 08:38:29.401302 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" event={"ID":"13e7f32e-2be0-476f-9043-9446462ae7df","Type":"ContainerStarted","Data":"fa02475ae23657c34cd1652041b244739e89eee827c8aa497498b492e0cedfa8"} Feb 25 08:38:29 crc kubenswrapper[4978]: I0225 08:38:29.401545 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" podUID="a20030e2-d291-44f8-aaa4-6b33d5e11122" containerName="dnsmasq-dns" containerID="cri-o://c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e" gracePeriod=10 Feb 25 08:38:29 crc kubenswrapper[4978]: I0225 08:38:29.951475 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.057159 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-ovsdbserver-sb\") pod \"a20030e2-d291-44f8-aaa4-6b33d5e11122\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.057342 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-config\") pod \"a20030e2-d291-44f8-aaa4-6b33d5e11122\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.058189 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-dns-svc\") pod \"a20030e2-d291-44f8-aaa4-6b33d5e11122\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.058350 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cxc8\" (UniqueName: \"kubernetes.io/projected/a20030e2-d291-44f8-aaa4-6b33d5e11122-kube-api-access-2cxc8\") pod \"a20030e2-d291-44f8-aaa4-6b33d5e11122\" (UID: \"a20030e2-d291-44f8-aaa4-6b33d5e11122\") " Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.065566 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a20030e2-d291-44f8-aaa4-6b33d5e11122-kube-api-access-2cxc8" (OuterVolumeSpecName: "kube-api-access-2cxc8") pod "a20030e2-d291-44f8-aaa4-6b33d5e11122" (UID: "a20030e2-d291-44f8-aaa4-6b33d5e11122"). InnerVolumeSpecName "kube-api-access-2cxc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.118763 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a20030e2-d291-44f8-aaa4-6b33d5e11122" (UID: "a20030e2-d291-44f8-aaa4-6b33d5e11122"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.119520 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-config" (OuterVolumeSpecName: "config") pod "a20030e2-d291-44f8-aaa4-6b33d5e11122" (UID: "a20030e2-d291-44f8-aaa4-6b33d5e11122"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.131581 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a20030e2-d291-44f8-aaa4-6b33d5e11122" (UID: "a20030e2-d291-44f8-aaa4-6b33d5e11122"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.160756 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.160802 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.160821 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a20030e2-d291-44f8-aaa4-6b33d5e11122-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.160840 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cxc8\" (UniqueName: \"kubernetes.io/projected/a20030e2-d291-44f8-aaa4-6b33d5e11122-kube-api-access-2cxc8\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.414244 4978 generic.go:334] "Generic (PLEG): container finished" podID="a20030e2-d291-44f8-aaa4-6b33d5e11122" containerID="c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e" exitCode=0 Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.414314 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" event={"ID":"a20030e2-d291-44f8-aaa4-6b33d5e11122","Type":"ContainerDied","Data":"c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e"} Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.414348 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" event={"ID":"a20030e2-d291-44f8-aaa4-6b33d5e11122","Type":"ContainerDied","Data":"8262b3075f545ffb8f6c549b40f56cb502783535cfcf58b440f2ea7831470697"} Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.414354 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697c57b9b5-sf2b5" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.414396 4978 scope.go:117] "RemoveContainer" containerID="c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.422500 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" event={"ID":"13e7f32e-2be0-476f-9043-9446462ae7df","Type":"ContainerStarted","Data":"5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a"} Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.422787 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.446718 4978 scope.go:117] "RemoveContainer" containerID="dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.452167 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" podStartSLOduration=2.4521518909999998 podStartE2EDuration="2.452151891s" podCreationTimestamp="2026-02-25 08:38:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:38:30.450682806 +0000 UTC m=+6803.889939275" watchObservedRunningTime="2026-02-25 08:38:30.452151891 +0000 UTC m=+6803.891408350" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.481428 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-697c57b9b5-sf2b5"] Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.487993 4978 scope.go:117] "RemoveContainer" containerID="c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e" Feb 25 08:38:30 crc kubenswrapper[4978]: E0225 08:38:30.488507 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e\": container with ID starting with c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e not found: ID does not exist" containerID="c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.488539 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e"} err="failed to get container status \"c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e\": rpc error: code = NotFound desc = could not find container \"c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e\": container with ID starting with c9af04fba8de58c3b00ad8fee8655af56cdd41780b37ece9a9d52a848cd94e9e not found: ID does not exist" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.488560 4978 scope.go:117] "RemoveContainer" containerID="dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a" Feb 25 08:38:30 crc kubenswrapper[4978]: E0225 08:38:30.488906 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a\": container with ID starting with dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a not found: ID does not exist" containerID="dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.488927 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a"} err="failed to get container status \"dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a\": rpc error: code = NotFound desc = could not find container \"dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a\": container with ID starting with dcfda99f3caea42383a4f91f6c7f22a462df3f430f5e70948cd190e8b6252e8a not found: ID does not exist" Feb 25 08:38:30 crc kubenswrapper[4978]: I0225 08:38:30.491325 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-697c57b9b5-sf2b5"] Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.352018 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a20030e2-d291-44f8-aaa4-6b33d5e11122" path="/var/lib/kubelet/pods/a20030e2-d291-44f8-aaa4-6b33d5e11122/volumes" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.379210 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Feb 25 08:38:31 crc kubenswrapper[4978]: E0225 08:38:31.379776 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a20030e2-d291-44f8-aaa4-6b33d5e11122" containerName="init" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.379810 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a20030e2-d291-44f8-aaa4-6b33d5e11122" containerName="init" Feb 25 08:38:31 crc kubenswrapper[4978]: E0225 08:38:31.379843 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a20030e2-d291-44f8-aaa4-6b33d5e11122" containerName="dnsmasq-dns" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.379856 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a20030e2-d291-44f8-aaa4-6b33d5e11122" containerName="dnsmasq-dns" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.380204 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a20030e2-d291-44f8-aaa4-6b33d5e11122" containerName="dnsmasq-dns" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.381619 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.395572 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.417363 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.491286 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52nnr\" (UniqueName: \"kubernetes.io/projected/d8a2072e-806f-4e38-8cee-18b683945ad2-kube-api-access-52nnr\") pod \"ovn-copy-data\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.491337 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\") pod \"ovn-copy-data\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.491432 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/d8a2072e-806f-4e38-8cee-18b683945ad2-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.592996 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\") pod \"ovn-copy-data\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.593092 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/d8a2072e-806f-4e38-8cee-18b683945ad2-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.593502 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52nnr\" (UniqueName: \"kubernetes.io/projected/d8a2072e-806f-4e38-8cee-18b683945ad2-kube-api-access-52nnr\") pod \"ovn-copy-data\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.603548 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.603611 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\") pod \"ovn-copy-data\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7ecdf915d9ac6b1e7346308b06a52b5b0f150ba9e9bc4c815d90f83b42664e11/globalmount\"" pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.603922 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/d8a2072e-806f-4e38-8cee-18b683945ad2-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.614289 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52nnr\" (UniqueName: \"kubernetes.io/projected/d8a2072e-806f-4e38-8cee-18b683945ad2-kube-api-access-52nnr\") pod \"ovn-copy-data\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.657224 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\") pod \"ovn-copy-data\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " pod="openstack/ovn-copy-data" Feb 25 08:38:31 crc kubenswrapper[4978]: I0225 08:38:31.714259 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Feb 25 08:38:32 crc kubenswrapper[4978]: I0225 08:38:32.335763 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Feb 25 08:38:32 crc kubenswrapper[4978]: W0225 08:38:32.337201 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8a2072e_806f_4e38_8cee_18b683945ad2.slice/crio-668da62b3b748145515f268b5e691cab63eaf7693cf3f093494274f54e6aa874 WatchSource:0}: Error finding container 668da62b3b748145515f268b5e691cab63eaf7693cf3f093494274f54e6aa874: Status 404 returned error can't find the container with id 668da62b3b748145515f268b5e691cab63eaf7693cf3f093494274f54e6aa874 Feb 25 08:38:32 crc kubenswrapper[4978]: I0225 08:38:32.448701 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"d8a2072e-806f-4e38-8cee-18b683945ad2","Type":"ContainerStarted","Data":"668da62b3b748145515f268b5e691cab63eaf7693cf3f093494274f54e6aa874"} Feb 25 08:38:33 crc kubenswrapper[4978]: I0225 08:38:33.469523 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"d8a2072e-806f-4e38-8cee-18b683945ad2","Type":"ContainerStarted","Data":"449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9"} Feb 25 08:38:33 crc kubenswrapper[4978]: I0225 08:38:33.509298 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.292296343 podStartE2EDuration="3.509269708s" podCreationTimestamp="2026-02-25 08:38:30 +0000 UTC" firstStartedPulling="2026-02-25 08:38:32.343169811 +0000 UTC m=+6805.782426310" lastFinishedPulling="2026-02-25 08:38:32.560143216 +0000 UTC m=+6805.999399675" observedRunningTime="2026-02-25 08:38:33.492844559 +0000 UTC m=+6806.932101078" watchObservedRunningTime="2026-02-25 08:38:33.509269708 +0000 UTC m=+6806.948526207" Feb 25 08:38:38 crc kubenswrapper[4978]: I0225 08:38:38.709679 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:38:38 crc kubenswrapper[4978]: I0225 08:38:38.887780 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76d5b4c5b5-crsmw"] Feb 25 08:38:38 crc kubenswrapper[4978]: I0225 08:38:38.888576 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" podUID="f887ca94-f583-4674-b0ed-8b45710a1804" containerName="dnsmasq-dns" containerID="cri-o://4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7" gracePeriod=10 Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.332326 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.480150 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-config\") pod \"f887ca94-f583-4674-b0ed-8b45710a1804\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.480316 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb9h2\" (UniqueName: \"kubernetes.io/projected/f887ca94-f583-4674-b0ed-8b45710a1804-kube-api-access-wb9h2\") pod \"f887ca94-f583-4674-b0ed-8b45710a1804\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.480415 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-dns-svc\") pod \"f887ca94-f583-4674-b0ed-8b45710a1804\" (UID: \"f887ca94-f583-4674-b0ed-8b45710a1804\") " Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.489553 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f887ca94-f583-4674-b0ed-8b45710a1804-kube-api-access-wb9h2" (OuterVolumeSpecName: "kube-api-access-wb9h2") pod "f887ca94-f583-4674-b0ed-8b45710a1804" (UID: "f887ca94-f583-4674-b0ed-8b45710a1804"). InnerVolumeSpecName "kube-api-access-wb9h2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.531700 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-config" (OuterVolumeSpecName: "config") pod "f887ca94-f583-4674-b0ed-8b45710a1804" (UID: "f887ca94-f583-4674-b0ed-8b45710a1804"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.536507 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f887ca94-f583-4674-b0ed-8b45710a1804" (UID: "f887ca94-f583-4674-b0ed-8b45710a1804"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.549847 4978 generic.go:334] "Generic (PLEG): container finished" podID="f887ca94-f583-4674-b0ed-8b45710a1804" containerID="4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7" exitCode=0 Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.549886 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" event={"ID":"f887ca94-f583-4674-b0ed-8b45710a1804","Type":"ContainerDied","Data":"4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7"} Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.549946 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" event={"ID":"f887ca94-f583-4674-b0ed-8b45710a1804","Type":"ContainerDied","Data":"f28f115feb721f8bc99ada00ef097dd49b06b27a2ea8b802c1a40a43fb2c85db"} Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.549953 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76d5b4c5b5-crsmw" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.549967 4978 scope.go:117] "RemoveContainer" containerID="4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.582492 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb9h2\" (UniqueName: \"kubernetes.io/projected/f887ca94-f583-4674-b0ed-8b45710a1804-kube-api-access-wb9h2\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.582530 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.582542 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f887ca94-f583-4674-b0ed-8b45710a1804-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.587231 4978 scope.go:117] "RemoveContainer" containerID="c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.590135 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76d5b4c5b5-crsmw"] Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.595993 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76d5b4c5b5-crsmw"] Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.615777 4978 scope.go:117] "RemoveContainer" containerID="4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7" Feb 25 08:38:39 crc kubenswrapper[4978]: E0225 08:38:39.616067 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7\": container with ID starting with 4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7 not found: ID does not exist" containerID="4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.616099 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7"} err="failed to get container status \"4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7\": rpc error: code = NotFound desc = could not find container \"4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7\": container with ID starting with 4c903f0e77eba38cedaf3b0b594ecde902deb11572ebab06543634f9a126d5d7 not found: ID does not exist" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.616120 4978 scope.go:117] "RemoveContainer" containerID="c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e" Feb 25 08:38:39 crc kubenswrapper[4978]: E0225 08:38:39.616883 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e\": container with ID starting with c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e not found: ID does not exist" containerID="c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e" Feb 25 08:38:39 crc kubenswrapper[4978]: I0225 08:38:39.616903 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e"} err="failed to get container status \"c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e\": rpc error: code = NotFound desc = could not find container \"c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e\": container with ID starting with c649ac0264cb93b8b4d387f7cea753f3e9e9eb36a0fe56d47cb93964e6f8c91e not found: ID does not exist" Feb 25 08:38:41 crc kubenswrapper[4978]: I0225 08:38:41.338952 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f887ca94-f583-4674-b0ed-8b45710a1804" path="/var/lib/kubelet/pods/f887ca94-f583-4674-b0ed-8b45710a1804/volumes" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.405249 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 25 08:38:42 crc kubenswrapper[4978]: E0225 08:38:42.406256 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f887ca94-f583-4674-b0ed-8b45710a1804" containerName="dnsmasq-dns" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.406274 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f887ca94-f583-4674-b0ed-8b45710a1804" containerName="dnsmasq-dns" Feb 25 08:38:42 crc kubenswrapper[4978]: E0225 08:38:42.406304 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f887ca94-f583-4674-b0ed-8b45710a1804" containerName="init" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.406313 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f887ca94-f583-4674-b0ed-8b45710a1804" containerName="init" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.406731 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f887ca94-f583-4674-b0ed-8b45710a1804" containerName="dnsmasq-dns" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.408518 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.411350 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.411598 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-jb8mt" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.411813 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.411816 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.433331 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.531879 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c197b2cc-cc65-4113-ab1a-517539270d88-scripts\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.532016 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c197b2cc-cc65-4113-ab1a-517539270d88-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.532080 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c197b2cc-cc65-4113-ab1a-517539270d88-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.532109 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c197b2cc-cc65-4113-ab1a-517539270d88-config\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.532133 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c197b2cc-cc65-4113-ab1a-517539270d88-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.532159 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c197b2cc-cc65-4113-ab1a-517539270d88-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.532191 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvdxm\" (UniqueName: \"kubernetes.io/projected/c197b2cc-cc65-4113-ab1a-517539270d88-kube-api-access-jvdxm\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.633478 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c197b2cc-cc65-4113-ab1a-517539270d88-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.633544 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c197b2cc-cc65-4113-ab1a-517539270d88-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.633569 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c197b2cc-cc65-4113-ab1a-517539270d88-config\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.633589 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c197b2cc-cc65-4113-ab1a-517539270d88-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.633607 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c197b2cc-cc65-4113-ab1a-517539270d88-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.633632 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvdxm\" (UniqueName: \"kubernetes.io/projected/c197b2cc-cc65-4113-ab1a-517539270d88-kube-api-access-jvdxm\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.633654 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c197b2cc-cc65-4113-ab1a-517539270d88-scripts\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.634531 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c197b2cc-cc65-4113-ab1a-517539270d88-scripts\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.635674 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c197b2cc-cc65-4113-ab1a-517539270d88-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.635998 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c197b2cc-cc65-4113-ab1a-517539270d88-config\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.639970 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c197b2cc-cc65-4113-ab1a-517539270d88-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.640970 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c197b2cc-cc65-4113-ab1a-517539270d88-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.652802 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c197b2cc-cc65-4113-ab1a-517539270d88-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.668398 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvdxm\" (UniqueName: \"kubernetes.io/projected/c197b2cc-cc65-4113-ab1a-517539270d88-kube-api-access-jvdxm\") pod \"ovn-northd-0\" (UID: \"c197b2cc-cc65-4113-ab1a-517539270d88\") " pod="openstack/ovn-northd-0" Feb 25 08:38:42 crc kubenswrapper[4978]: I0225 08:38:42.734232 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 25 08:38:43 crc kubenswrapper[4978]: I0225 08:38:43.179234 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 25 08:38:43 crc kubenswrapper[4978]: W0225 08:38:43.193806 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc197b2cc_cc65_4113_ab1a_517539270d88.slice/crio-3c78ba89fca260123d463cbbdccdd845cd2fe1847f7c17c2b022a4098dfe10d6 WatchSource:0}: Error finding container 3c78ba89fca260123d463cbbdccdd845cd2fe1847f7c17c2b022a4098dfe10d6: Status 404 returned error can't find the container with id 3c78ba89fca260123d463cbbdccdd845cd2fe1847f7c17c2b022a4098dfe10d6 Feb 25 08:38:43 crc kubenswrapper[4978]: I0225 08:38:43.588296 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c197b2cc-cc65-4113-ab1a-517539270d88","Type":"ContainerStarted","Data":"3c78ba89fca260123d463cbbdccdd845cd2fe1847f7c17c2b022a4098dfe10d6"} Feb 25 08:38:44 crc kubenswrapper[4978]: I0225 08:38:44.598558 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c197b2cc-cc65-4113-ab1a-517539270d88","Type":"ContainerStarted","Data":"29a0a9085bb36e4fc467a6d5f69dfb217aa8ef31a72e4063d97dd78be7ec6376"} Feb 25 08:38:44 crc kubenswrapper[4978]: I0225 08:38:44.598930 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c197b2cc-cc65-4113-ab1a-517539270d88","Type":"ContainerStarted","Data":"970dc234d25bdb44738d0cdca75f5bc9d9315ed62abd4685486665f9a00329aa"} Feb 25 08:38:44 crc kubenswrapper[4978]: I0225 08:38:44.598953 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 25 08:38:44 crc kubenswrapper[4978]: I0225 08:38:44.620104 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.952670221 podStartE2EDuration="2.6200852s" podCreationTimestamp="2026-02-25 08:38:42 +0000 UTC" firstStartedPulling="2026-02-25 08:38:43.197066268 +0000 UTC m=+6816.636322727" lastFinishedPulling="2026-02-25 08:38:43.864481247 +0000 UTC m=+6817.303737706" observedRunningTime="2026-02-25 08:38:44.6171895 +0000 UTC m=+6818.056445979" watchObservedRunningTime="2026-02-25 08:38:44.6200852 +0000 UTC m=+6818.059341659" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.410902 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-znvxz"] Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.413515 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-znvxz" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.452278 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-znvxz"] Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.503470 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rd6l\" (UniqueName: \"kubernetes.io/projected/cdab92b0-d524-4536-bf2d-a474a8cd2787-kube-api-access-2rd6l\") pod \"keystone-db-create-znvxz\" (UID: \"cdab92b0-d524-4536-bf2d-a474a8cd2787\") " pod="openstack/keystone-db-create-znvxz" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.503527 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdab92b0-d524-4536-bf2d-a474a8cd2787-operator-scripts\") pod \"keystone-db-create-znvxz\" (UID: \"cdab92b0-d524-4536-bf2d-a474a8cd2787\") " pod="openstack/keystone-db-create-znvxz" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.510790 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-608f-account-create-update-cwl9c"] Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.511955 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-608f-account-create-update-cwl9c" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.513968 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.517535 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-608f-account-create-update-cwl9c"] Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.605505 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg9rd\" (UniqueName: \"kubernetes.io/projected/47319f12-b693-4085-84d3-8d0d1942b36c-kube-api-access-hg9rd\") pod \"keystone-608f-account-create-update-cwl9c\" (UID: \"47319f12-b693-4085-84d3-8d0d1942b36c\") " pod="openstack/keystone-608f-account-create-update-cwl9c" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.605594 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rd6l\" (UniqueName: \"kubernetes.io/projected/cdab92b0-d524-4536-bf2d-a474a8cd2787-kube-api-access-2rd6l\") pod \"keystone-db-create-znvxz\" (UID: \"cdab92b0-d524-4536-bf2d-a474a8cd2787\") " pod="openstack/keystone-db-create-znvxz" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.605629 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdab92b0-d524-4536-bf2d-a474a8cd2787-operator-scripts\") pod \"keystone-db-create-znvxz\" (UID: \"cdab92b0-d524-4536-bf2d-a474a8cd2787\") " pod="openstack/keystone-db-create-znvxz" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.605691 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47319f12-b693-4085-84d3-8d0d1942b36c-operator-scripts\") pod \"keystone-608f-account-create-update-cwl9c\" (UID: \"47319f12-b693-4085-84d3-8d0d1942b36c\") " pod="openstack/keystone-608f-account-create-update-cwl9c" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.606884 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdab92b0-d524-4536-bf2d-a474a8cd2787-operator-scripts\") pod \"keystone-db-create-znvxz\" (UID: \"cdab92b0-d524-4536-bf2d-a474a8cd2787\") " pod="openstack/keystone-db-create-znvxz" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.633219 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rd6l\" (UniqueName: \"kubernetes.io/projected/cdab92b0-d524-4536-bf2d-a474a8cd2787-kube-api-access-2rd6l\") pod \"keystone-db-create-znvxz\" (UID: \"cdab92b0-d524-4536-bf2d-a474a8cd2787\") " pod="openstack/keystone-db-create-znvxz" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.707600 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47319f12-b693-4085-84d3-8d0d1942b36c-operator-scripts\") pod \"keystone-608f-account-create-update-cwl9c\" (UID: \"47319f12-b693-4085-84d3-8d0d1942b36c\") " pod="openstack/keystone-608f-account-create-update-cwl9c" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.707688 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg9rd\" (UniqueName: \"kubernetes.io/projected/47319f12-b693-4085-84d3-8d0d1942b36c-kube-api-access-hg9rd\") pod \"keystone-608f-account-create-update-cwl9c\" (UID: \"47319f12-b693-4085-84d3-8d0d1942b36c\") " pod="openstack/keystone-608f-account-create-update-cwl9c" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.709289 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47319f12-b693-4085-84d3-8d0d1942b36c-operator-scripts\") pod \"keystone-608f-account-create-update-cwl9c\" (UID: \"47319f12-b693-4085-84d3-8d0d1942b36c\") " pod="openstack/keystone-608f-account-create-update-cwl9c" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.727707 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg9rd\" (UniqueName: \"kubernetes.io/projected/47319f12-b693-4085-84d3-8d0d1942b36c-kube-api-access-hg9rd\") pod \"keystone-608f-account-create-update-cwl9c\" (UID: \"47319f12-b693-4085-84d3-8d0d1942b36c\") " pod="openstack/keystone-608f-account-create-update-cwl9c" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.772634 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-znvxz" Feb 25 08:38:50 crc kubenswrapper[4978]: I0225 08:38:50.828494 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-608f-account-create-update-cwl9c" Feb 25 08:38:51 crc kubenswrapper[4978]: I0225 08:38:51.203313 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-znvxz"] Feb 25 08:38:51 crc kubenswrapper[4978]: W0225 08:38:51.216008 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdab92b0_d524_4536_bf2d_a474a8cd2787.slice/crio-86a4f934d1d6c36878a9996d37fb9136314e967affa105610500814d4dde4110 WatchSource:0}: Error finding container 86a4f934d1d6c36878a9996d37fb9136314e967affa105610500814d4dde4110: Status 404 returned error can't find the container with id 86a4f934d1d6c36878a9996d37fb9136314e967affa105610500814d4dde4110 Feb 25 08:38:51 crc kubenswrapper[4978]: I0225 08:38:51.288596 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-608f-account-create-update-cwl9c"] Feb 25 08:38:51 crc kubenswrapper[4978]: W0225 08:38:51.293482 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47319f12_b693_4085_84d3_8d0d1942b36c.slice/crio-61deb2fbc24fb4a7e8c349d9f4022b79fd7e32d5c11f9f7a3d01868a04a5d98a WatchSource:0}: Error finding container 61deb2fbc24fb4a7e8c349d9f4022b79fd7e32d5c11f9f7a3d01868a04a5d98a: Status 404 returned error can't find the container with id 61deb2fbc24fb4a7e8c349d9f4022b79fd7e32d5c11f9f7a3d01868a04a5d98a Feb 25 08:38:51 crc kubenswrapper[4978]: I0225 08:38:51.670653 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-608f-account-create-update-cwl9c" event={"ID":"47319f12-b693-4085-84d3-8d0d1942b36c","Type":"ContainerStarted","Data":"e484ad914e91a8f456a4df58a588f86fd699f9f1247438b959006b04c1a1ddd7"} Feb 25 08:38:51 crc kubenswrapper[4978]: I0225 08:38:51.670709 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-608f-account-create-update-cwl9c" event={"ID":"47319f12-b693-4085-84d3-8d0d1942b36c","Type":"ContainerStarted","Data":"61deb2fbc24fb4a7e8c349d9f4022b79fd7e32d5c11f9f7a3d01868a04a5d98a"} Feb 25 08:38:51 crc kubenswrapper[4978]: I0225 08:38:51.673676 4978 generic.go:334] "Generic (PLEG): container finished" podID="cdab92b0-d524-4536-bf2d-a474a8cd2787" containerID="2e3aadcf94ca6bd2f03f7793def413ca633a70114ffae9b6d4e415755c398d25" exitCode=0 Feb 25 08:38:51 crc kubenswrapper[4978]: I0225 08:38:51.673716 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-znvxz" event={"ID":"cdab92b0-d524-4536-bf2d-a474a8cd2787","Type":"ContainerDied","Data":"2e3aadcf94ca6bd2f03f7793def413ca633a70114ffae9b6d4e415755c398d25"} Feb 25 08:38:51 crc kubenswrapper[4978]: I0225 08:38:51.673735 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-znvxz" event={"ID":"cdab92b0-d524-4536-bf2d-a474a8cd2787","Type":"ContainerStarted","Data":"86a4f934d1d6c36878a9996d37fb9136314e967affa105610500814d4dde4110"} Feb 25 08:38:51 crc kubenswrapper[4978]: I0225 08:38:51.692596 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-608f-account-create-update-cwl9c" podStartSLOduration=1.6925719689999998 podStartE2EDuration="1.692571969s" podCreationTimestamp="2026-02-25 08:38:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:38:51.684941073 +0000 UTC m=+6825.124197562" watchObservedRunningTime="2026-02-25 08:38:51.692571969 +0000 UTC m=+6825.131828448" Feb 25 08:38:52 crc kubenswrapper[4978]: I0225 08:38:52.686612 4978 generic.go:334] "Generic (PLEG): container finished" podID="47319f12-b693-4085-84d3-8d0d1942b36c" containerID="e484ad914e91a8f456a4df58a588f86fd699f9f1247438b959006b04c1a1ddd7" exitCode=0 Feb 25 08:38:52 crc kubenswrapper[4978]: I0225 08:38:52.686754 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-608f-account-create-update-cwl9c" event={"ID":"47319f12-b693-4085-84d3-8d0d1942b36c","Type":"ContainerDied","Data":"e484ad914e91a8f456a4df58a588f86fd699f9f1247438b959006b04c1a1ddd7"} Feb 25 08:38:53 crc kubenswrapper[4978]: I0225 08:38:53.061569 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-znvxz" Feb 25 08:38:53 crc kubenswrapper[4978]: I0225 08:38:53.155223 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdab92b0-d524-4536-bf2d-a474a8cd2787-operator-scripts\") pod \"cdab92b0-d524-4536-bf2d-a474a8cd2787\" (UID: \"cdab92b0-d524-4536-bf2d-a474a8cd2787\") " Feb 25 08:38:53 crc kubenswrapper[4978]: I0225 08:38:53.155358 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rd6l\" (UniqueName: \"kubernetes.io/projected/cdab92b0-d524-4536-bf2d-a474a8cd2787-kube-api-access-2rd6l\") pod \"cdab92b0-d524-4536-bf2d-a474a8cd2787\" (UID: \"cdab92b0-d524-4536-bf2d-a474a8cd2787\") " Feb 25 08:38:53 crc kubenswrapper[4978]: I0225 08:38:53.156056 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdab92b0-d524-4536-bf2d-a474a8cd2787-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cdab92b0-d524-4536-bf2d-a474a8cd2787" (UID: "cdab92b0-d524-4536-bf2d-a474a8cd2787"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:38:53 crc kubenswrapper[4978]: I0225 08:38:53.164510 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdab92b0-d524-4536-bf2d-a474a8cd2787-kube-api-access-2rd6l" (OuterVolumeSpecName: "kube-api-access-2rd6l") pod "cdab92b0-d524-4536-bf2d-a474a8cd2787" (UID: "cdab92b0-d524-4536-bf2d-a474a8cd2787"). InnerVolumeSpecName "kube-api-access-2rd6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:38:53 crc kubenswrapper[4978]: I0225 08:38:53.257452 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cdab92b0-d524-4536-bf2d-a474a8cd2787-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:53 crc kubenswrapper[4978]: I0225 08:38:53.257490 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rd6l\" (UniqueName: \"kubernetes.io/projected/cdab92b0-d524-4536-bf2d-a474a8cd2787-kube-api-access-2rd6l\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:53 crc kubenswrapper[4978]: I0225 08:38:53.703362 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-znvxz" Feb 25 08:38:53 crc kubenswrapper[4978]: I0225 08:38:53.704323 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-znvxz" event={"ID":"cdab92b0-d524-4536-bf2d-a474a8cd2787","Type":"ContainerDied","Data":"86a4f934d1d6c36878a9996d37fb9136314e967affa105610500814d4dde4110"} Feb 25 08:38:53 crc kubenswrapper[4978]: I0225 08:38:53.704358 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86a4f934d1d6c36878a9996d37fb9136314e967affa105610500814d4dde4110" Feb 25 08:38:54 crc kubenswrapper[4978]: I0225 08:38:54.119021 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-608f-account-create-update-cwl9c" Feb 25 08:38:54 crc kubenswrapper[4978]: I0225 08:38:54.274139 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47319f12-b693-4085-84d3-8d0d1942b36c-operator-scripts\") pod \"47319f12-b693-4085-84d3-8d0d1942b36c\" (UID: \"47319f12-b693-4085-84d3-8d0d1942b36c\") " Feb 25 08:38:54 crc kubenswrapper[4978]: I0225 08:38:54.274705 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg9rd\" (UniqueName: \"kubernetes.io/projected/47319f12-b693-4085-84d3-8d0d1942b36c-kube-api-access-hg9rd\") pod \"47319f12-b693-4085-84d3-8d0d1942b36c\" (UID: \"47319f12-b693-4085-84d3-8d0d1942b36c\") " Feb 25 08:38:54 crc kubenswrapper[4978]: I0225 08:38:54.275098 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47319f12-b693-4085-84d3-8d0d1942b36c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "47319f12-b693-4085-84d3-8d0d1942b36c" (UID: "47319f12-b693-4085-84d3-8d0d1942b36c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:38:54 crc kubenswrapper[4978]: I0225 08:38:54.280314 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47319f12-b693-4085-84d3-8d0d1942b36c-kube-api-access-hg9rd" (OuterVolumeSpecName: "kube-api-access-hg9rd") pod "47319f12-b693-4085-84d3-8d0d1942b36c" (UID: "47319f12-b693-4085-84d3-8d0d1942b36c"). InnerVolumeSpecName "kube-api-access-hg9rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:38:54 crc kubenswrapper[4978]: I0225 08:38:54.377402 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47319f12-b693-4085-84d3-8d0d1942b36c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:54 crc kubenswrapper[4978]: I0225 08:38:54.377455 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg9rd\" (UniqueName: \"kubernetes.io/projected/47319f12-b693-4085-84d3-8d0d1942b36c-kube-api-access-hg9rd\") on node \"crc\" DevicePath \"\"" Feb 25 08:38:54 crc kubenswrapper[4978]: I0225 08:38:54.736457 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-608f-account-create-update-cwl9c" event={"ID":"47319f12-b693-4085-84d3-8d0d1942b36c","Type":"ContainerDied","Data":"61deb2fbc24fb4a7e8c349d9f4022b79fd7e32d5c11f9f7a3d01868a04a5d98a"} Feb 25 08:38:54 crc kubenswrapper[4978]: I0225 08:38:54.736521 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-608f-account-create-update-cwl9c" Feb 25 08:38:54 crc kubenswrapper[4978]: I0225 08:38:54.736538 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61deb2fbc24fb4a7e8c349d9f4022b79fd7e32d5c11f9f7a3d01868a04a5d98a" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.967724 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-l9hnq"] Feb 25 08:38:55 crc kubenswrapper[4978]: E0225 08:38:55.968607 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdab92b0-d524-4536-bf2d-a474a8cd2787" containerName="mariadb-database-create" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.968632 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdab92b0-d524-4536-bf2d-a474a8cd2787" containerName="mariadb-database-create" Feb 25 08:38:55 crc kubenswrapper[4978]: E0225 08:38:55.968679 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47319f12-b693-4085-84d3-8d0d1942b36c" containerName="mariadb-account-create-update" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.968694 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="47319f12-b693-4085-84d3-8d0d1942b36c" containerName="mariadb-account-create-update" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.968962 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdab92b0-d524-4536-bf2d-a474a8cd2787" containerName="mariadb-database-create" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.968990 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="47319f12-b693-4085-84d3-8d0d1942b36c" containerName="mariadb-account-create-update" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.969882 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.976451 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.976528 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fnhrp" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.976540 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.978588 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 08:38:55 crc kubenswrapper[4978]: I0225 08:38:55.981125 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-l9hnq"] Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.044345 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-config-data\") pod \"keystone-db-sync-l9hnq\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.044438 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fttbl\" (UniqueName: \"kubernetes.io/projected/d2e7af06-5b02-45cf-bdad-685aee4a49f4-kube-api-access-fttbl\") pod \"keystone-db-sync-l9hnq\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.044475 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-combined-ca-bundle\") pod \"keystone-db-sync-l9hnq\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.146174 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-config-data\") pod \"keystone-db-sync-l9hnq\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.146234 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fttbl\" (UniqueName: \"kubernetes.io/projected/d2e7af06-5b02-45cf-bdad-685aee4a49f4-kube-api-access-fttbl\") pod \"keystone-db-sync-l9hnq\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.146270 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-combined-ca-bundle\") pod \"keystone-db-sync-l9hnq\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.152560 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-config-data\") pod \"keystone-db-sync-l9hnq\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.153222 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-combined-ca-bundle\") pod \"keystone-db-sync-l9hnq\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.169915 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fttbl\" (UniqueName: \"kubernetes.io/projected/d2e7af06-5b02-45cf-bdad-685aee4a49f4-kube-api-access-fttbl\") pod \"keystone-db-sync-l9hnq\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.302072 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.858450 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-l9hnq"] Feb 25 08:38:56 crc kubenswrapper[4978]: I0225 08:38:56.876169 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 08:38:57 crc kubenswrapper[4978]: I0225 08:38:57.765640 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l9hnq" event={"ID":"d2e7af06-5b02-45cf-bdad-685aee4a49f4","Type":"ContainerStarted","Data":"d1248d5dd3579f98919f207166d630cd008fb4f9a02660338fb596ef5ef07ae9"} Feb 25 08:39:01 crc kubenswrapper[4978]: I0225 08:39:01.806568 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l9hnq" event={"ID":"d2e7af06-5b02-45cf-bdad-685aee4a49f4","Type":"ContainerStarted","Data":"b502eadcbed72640c53debfbc5666578bfc0a91bdf42f7286b86218df0a76f9c"} Feb 25 08:39:01 crc kubenswrapper[4978]: I0225 08:39:01.829886 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-l9hnq" podStartSLOduration=2.260815519 podStartE2EDuration="6.829868754s" podCreationTimestamp="2026-02-25 08:38:55 +0000 UTC" firstStartedPulling="2026-02-25 08:38:56.875558446 +0000 UTC m=+6830.314814955" lastFinishedPulling="2026-02-25 08:39:01.444611691 +0000 UTC m=+6834.883868190" observedRunningTime="2026-02-25 08:39:01.829682068 +0000 UTC m=+6835.268938607" watchObservedRunningTime="2026-02-25 08:39:01.829868754 +0000 UTC m=+6835.269125213" Feb 25 08:39:02 crc kubenswrapper[4978]: I0225 08:39:02.800326 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 25 08:39:03 crc kubenswrapper[4978]: I0225 08:39:03.848254 4978 generic.go:334] "Generic (PLEG): container finished" podID="d2e7af06-5b02-45cf-bdad-685aee4a49f4" containerID="b502eadcbed72640c53debfbc5666578bfc0a91bdf42f7286b86218df0a76f9c" exitCode=0 Feb 25 08:39:03 crc kubenswrapper[4978]: I0225 08:39:03.848332 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l9hnq" event={"ID":"d2e7af06-5b02-45cf-bdad-685aee4a49f4","Type":"ContainerDied","Data":"b502eadcbed72640c53debfbc5666578bfc0a91bdf42f7286b86218df0a76f9c"} Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.242048 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.317644 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-config-data\") pod \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.317734 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-combined-ca-bundle\") pod \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.317894 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fttbl\" (UniqueName: \"kubernetes.io/projected/d2e7af06-5b02-45cf-bdad-685aee4a49f4-kube-api-access-fttbl\") pod \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\" (UID: \"d2e7af06-5b02-45cf-bdad-685aee4a49f4\") " Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.324623 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2e7af06-5b02-45cf-bdad-685aee4a49f4-kube-api-access-fttbl" (OuterVolumeSpecName: "kube-api-access-fttbl") pod "d2e7af06-5b02-45cf-bdad-685aee4a49f4" (UID: "d2e7af06-5b02-45cf-bdad-685aee4a49f4"). InnerVolumeSpecName "kube-api-access-fttbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.346810 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2e7af06-5b02-45cf-bdad-685aee4a49f4" (UID: "d2e7af06-5b02-45cf-bdad-685aee4a49f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.382928 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-config-data" (OuterVolumeSpecName: "config-data") pod "d2e7af06-5b02-45cf-bdad-685aee4a49f4" (UID: "d2e7af06-5b02-45cf-bdad-685aee4a49f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.420912 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.420954 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2e7af06-5b02-45cf-bdad-685aee4a49f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.420968 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fttbl\" (UniqueName: \"kubernetes.io/projected/d2e7af06-5b02-45cf-bdad-685aee4a49f4-kube-api-access-fttbl\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.871393 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-l9hnq" event={"ID":"d2e7af06-5b02-45cf-bdad-685aee4a49f4","Type":"ContainerDied","Data":"d1248d5dd3579f98919f207166d630cd008fb4f9a02660338fb596ef5ef07ae9"} Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.871434 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1248d5dd3579f98919f207166d630cd008fb4f9a02660338fb596ef5ef07ae9" Feb 25 08:39:05 crc kubenswrapper[4978]: I0225 08:39:05.871511 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-l9hnq" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.126350 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869c9ccd85-7hn7z"] Feb 25 08:39:06 crc kubenswrapper[4978]: E0225 08:39:06.126927 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2e7af06-5b02-45cf-bdad-685aee4a49f4" containerName="keystone-db-sync" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.126942 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2e7af06-5b02-45cf-bdad-685aee4a49f4" containerName="keystone-db-sync" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.127080 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2e7af06-5b02-45cf-bdad-685aee4a49f4" containerName="keystone-db-sync" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.127864 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.147991 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869c9ccd85-7hn7z"] Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.196126 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-zrzfd"] Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.197135 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.199644 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.199698 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.199736 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fnhrp" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.199847 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.199930 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.213727 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zrzfd"] Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.234801 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-dns-svc\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.234888 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-sb\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.234915 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-config\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.234990 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwnmn\" (UniqueName: \"kubernetes.io/projected/9b375187-4921-42b6-a5e3-00fc2cb0c881-kube-api-access-hwnmn\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.235011 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-nb\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336615 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-sb\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336659 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-config\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336695 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ckws\" (UniqueName: \"kubernetes.io/projected/bf25ea07-97cc-4be0-9d77-72e66864940c-kube-api-access-8ckws\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336719 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-scripts\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336766 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwnmn\" (UniqueName: \"kubernetes.io/projected/9b375187-4921-42b6-a5e3-00fc2cb0c881-kube-api-access-hwnmn\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336790 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-nb\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336807 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-combined-ca-bundle\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336836 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-credential-keys\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336866 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-dns-svc\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336879 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-fernet-keys\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.336898 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-config-data\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.337494 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-sb\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.337736 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-config\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.338086 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-nb\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.338293 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-dns-svc\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.360408 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwnmn\" (UniqueName: \"kubernetes.io/projected/9b375187-4921-42b6-a5e3-00fc2cb0c881-kube-api-access-hwnmn\") pod \"dnsmasq-dns-869c9ccd85-7hn7z\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.438105 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-credential-keys\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.438198 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-fernet-keys\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.438231 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-config-data\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.438317 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ckws\" (UniqueName: \"kubernetes.io/projected/bf25ea07-97cc-4be0-9d77-72e66864940c-kube-api-access-8ckws\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.438357 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-scripts\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.438438 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-combined-ca-bundle\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.440977 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-credential-keys\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.442054 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-config-data\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.443728 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-scripts\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.444009 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-fernet-keys\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.444099 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-combined-ca-bundle\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.461547 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ckws\" (UniqueName: \"kubernetes.io/projected/bf25ea07-97cc-4be0-9d77-72e66864940c-kube-api-access-8ckws\") pod \"keystone-bootstrap-zrzfd\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.464322 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:06 crc kubenswrapper[4978]: I0225 08:39:06.519936 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:07 crc kubenswrapper[4978]: I0225 08:39:07.016145 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869c9ccd85-7hn7z"] Feb 25 08:39:07 crc kubenswrapper[4978]: I0225 08:39:07.036033 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-zrzfd"] Feb 25 08:39:07 crc kubenswrapper[4978]: I0225 08:39:07.903416 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zrzfd" event={"ID":"bf25ea07-97cc-4be0-9d77-72e66864940c","Type":"ContainerStarted","Data":"21c86d0f01079552721b78ea7c677a42c0056cb0501c68e66ecab2e8deabee6f"} Feb 25 08:39:07 crc kubenswrapper[4978]: I0225 08:39:07.905194 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zrzfd" event={"ID":"bf25ea07-97cc-4be0-9d77-72e66864940c","Type":"ContainerStarted","Data":"9a07af5848e5f06bd216a319323b8bda41ed2dbfd17a7a978c36786baf8a36de"} Feb 25 08:39:07 crc kubenswrapper[4978]: I0225 08:39:07.912598 4978 generic.go:334] "Generic (PLEG): container finished" podID="9b375187-4921-42b6-a5e3-00fc2cb0c881" containerID="42de0ce6b0ed4214ac56d8732f38bcb69f51d2dcfbad218c53ff0c5e10431866" exitCode=0 Feb 25 08:39:07 crc kubenswrapper[4978]: I0225 08:39:07.912676 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" event={"ID":"9b375187-4921-42b6-a5e3-00fc2cb0c881","Type":"ContainerDied","Data":"42de0ce6b0ed4214ac56d8732f38bcb69f51d2dcfbad218c53ff0c5e10431866"} Feb 25 08:39:07 crc kubenswrapper[4978]: I0225 08:39:07.912716 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" event={"ID":"9b375187-4921-42b6-a5e3-00fc2cb0c881","Type":"ContainerStarted","Data":"9ee140e2a0ab6e00f47ad91c3b13336f5397f6a4264e2a0982d05e04ad50b978"} Feb 25 08:39:07 crc kubenswrapper[4978]: I0225 08:39:07.939188 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-zrzfd" podStartSLOduration=1.939168435 podStartE2EDuration="1.939168435s" podCreationTimestamp="2026-02-25 08:39:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:39:07.927187583 +0000 UTC m=+6841.366444082" watchObservedRunningTime="2026-02-25 08:39:07.939168435 +0000 UTC m=+6841.378424904" Feb 25 08:39:08 crc kubenswrapper[4978]: I0225 08:39:08.925110 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" event={"ID":"9b375187-4921-42b6-a5e3-00fc2cb0c881","Type":"ContainerStarted","Data":"c48acaea28da976ead56105c2ebbc4573da9737059f67fbb13863a0241a1db5f"} Feb 25 08:39:08 crc kubenswrapper[4978]: I0225 08:39:08.925745 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:08 crc kubenswrapper[4978]: I0225 08:39:08.948545 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" podStartSLOduration=2.948532714 podStartE2EDuration="2.948532714s" podCreationTimestamp="2026-02-25 08:39:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:39:08.94809021 +0000 UTC m=+6842.387346709" watchObservedRunningTime="2026-02-25 08:39:08.948532714 +0000 UTC m=+6842.387789173" Feb 25 08:39:10 crc kubenswrapper[4978]: I0225 08:39:10.953882 4978 generic.go:334] "Generic (PLEG): container finished" podID="bf25ea07-97cc-4be0-9d77-72e66864940c" containerID="21c86d0f01079552721b78ea7c677a42c0056cb0501c68e66ecab2e8deabee6f" exitCode=0 Feb 25 08:39:10 crc kubenswrapper[4978]: I0225 08:39:10.954408 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zrzfd" event={"ID":"bf25ea07-97cc-4be0-9d77-72e66864940c","Type":"ContainerDied","Data":"21c86d0f01079552721b78ea7c677a42c0056cb0501c68e66ecab2e8deabee6f"} Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.412390 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.565733 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-fernet-keys\") pod \"bf25ea07-97cc-4be0-9d77-72e66864940c\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.566317 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-combined-ca-bundle\") pod \"bf25ea07-97cc-4be0-9d77-72e66864940c\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.566451 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ckws\" (UniqueName: \"kubernetes.io/projected/bf25ea07-97cc-4be0-9d77-72e66864940c-kube-api-access-8ckws\") pod \"bf25ea07-97cc-4be0-9d77-72e66864940c\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.566560 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-scripts\") pod \"bf25ea07-97cc-4be0-9d77-72e66864940c\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.566659 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-credential-keys\") pod \"bf25ea07-97cc-4be0-9d77-72e66864940c\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.566714 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-config-data\") pod \"bf25ea07-97cc-4be0-9d77-72e66864940c\" (UID: \"bf25ea07-97cc-4be0-9d77-72e66864940c\") " Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.575592 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bf25ea07-97cc-4be0-9d77-72e66864940c" (UID: "bf25ea07-97cc-4be0-9d77-72e66864940c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.579593 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf25ea07-97cc-4be0-9d77-72e66864940c-kube-api-access-8ckws" (OuterVolumeSpecName: "kube-api-access-8ckws") pod "bf25ea07-97cc-4be0-9d77-72e66864940c" (UID: "bf25ea07-97cc-4be0-9d77-72e66864940c"). InnerVolumeSpecName "kube-api-access-8ckws". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.580247 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-scripts" (OuterVolumeSpecName: "scripts") pod "bf25ea07-97cc-4be0-9d77-72e66864940c" (UID: "bf25ea07-97cc-4be0-9d77-72e66864940c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.589630 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "bf25ea07-97cc-4be0-9d77-72e66864940c" (UID: "bf25ea07-97cc-4be0-9d77-72e66864940c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.605728 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-config-data" (OuterVolumeSpecName: "config-data") pod "bf25ea07-97cc-4be0-9d77-72e66864940c" (UID: "bf25ea07-97cc-4be0-9d77-72e66864940c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.610159 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf25ea07-97cc-4be0-9d77-72e66864940c" (UID: "bf25ea07-97cc-4be0-9d77-72e66864940c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.669256 4978 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.669293 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.669307 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ckws\" (UniqueName: \"kubernetes.io/projected/bf25ea07-97cc-4be0-9d77-72e66864940c-kube-api-access-8ckws\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.669556 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.669573 4978 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:12 crc kubenswrapper[4978]: I0225 08:39:12.669584 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf25ea07-97cc-4be0-9d77-72e66864940c-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.115712 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-zrzfd" event={"ID":"bf25ea07-97cc-4be0-9d77-72e66864940c","Type":"ContainerDied","Data":"9a07af5848e5f06bd216a319323b8bda41ed2dbfd17a7a978c36786baf8a36de"} Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.115767 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a07af5848e5f06bd216a319323b8bda41ed2dbfd17a7a978c36786baf8a36de" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.115851 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-zrzfd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.118442 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-zrzfd"] Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.132066 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-zrzfd"] Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.188234 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5nqzd"] Feb 25 08:39:13 crc kubenswrapper[4978]: E0225 08:39:13.188819 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf25ea07-97cc-4be0-9d77-72e66864940c" containerName="keystone-bootstrap" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.188856 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf25ea07-97cc-4be0-9d77-72e66864940c" containerName="keystone-bootstrap" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.189084 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf25ea07-97cc-4be0-9d77-72e66864940c" containerName="keystone-bootstrap" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.189782 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.193936 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.194413 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.194686 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.196126 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.196170 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fnhrp" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.200890 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5nqzd"] Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.288062 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-fernet-keys\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.288725 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-credential-keys\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.288761 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-scripts\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.288805 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpkxf\" (UniqueName: \"kubernetes.io/projected/e73d580e-bb87-4270-aec3-ad2c35da47ad-kube-api-access-bpkxf\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.288936 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-combined-ca-bundle\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.289035 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-config-data\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.343443 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf25ea07-97cc-4be0-9d77-72e66864940c" path="/var/lib/kubelet/pods/bf25ea07-97cc-4be0-9d77-72e66864940c/volumes" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.392103 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-combined-ca-bundle\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.392163 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-config-data\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.392325 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-fernet-keys\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.394513 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-credential-keys\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.395737 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-scripts\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.395808 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpkxf\" (UniqueName: \"kubernetes.io/projected/e73d580e-bb87-4270-aec3-ad2c35da47ad-kube-api-access-bpkxf\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.398644 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-credential-keys\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.398913 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-config-data\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.399767 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-scripts\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.399822 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-fernet-keys\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.401893 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-combined-ca-bundle\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.413944 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpkxf\" (UniqueName: \"kubernetes.io/projected/e73d580e-bb87-4270-aec3-ad2c35da47ad-kube-api-access-bpkxf\") pod \"keystone-bootstrap-5nqzd\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.508188 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:13 crc kubenswrapper[4978]: I0225 08:39:13.989846 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5nqzd"] Feb 25 08:39:14 crc kubenswrapper[4978]: I0225 08:39:14.127663 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5nqzd" event={"ID":"e73d580e-bb87-4270-aec3-ad2c35da47ad","Type":"ContainerStarted","Data":"51d46eef1e9bab11b3c14e96036706b0ca721871ebad6e467f3b4272d8b28b64"} Feb 25 08:39:15 crc kubenswrapper[4978]: I0225 08:39:15.140869 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5nqzd" event={"ID":"e73d580e-bb87-4270-aec3-ad2c35da47ad","Type":"ContainerStarted","Data":"e802925c232e3151d8073c8b7c34dc5a6827acd379157082050a59043f6fda37"} Feb 25 08:39:15 crc kubenswrapper[4978]: I0225 08:39:15.175734 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5nqzd" podStartSLOduration=2.175707679 podStartE2EDuration="2.175707679s" podCreationTimestamp="2026-02-25 08:39:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:39:15.167888487 +0000 UTC m=+6848.607144966" watchObservedRunningTime="2026-02-25 08:39:15.175707679 +0000 UTC m=+6848.614964148" Feb 25 08:39:16 crc kubenswrapper[4978]: I0225 08:39:16.466693 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:39:16 crc kubenswrapper[4978]: I0225 08:39:16.540859 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:39:16 crc kubenswrapper[4978]: I0225 08:39:16.541274 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:39:16 crc kubenswrapper[4978]: I0225 08:39:16.558704 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbc4875f7-rfcdw"] Feb 25 08:39:16 crc kubenswrapper[4978]: I0225 08:39:16.559053 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" podUID="13e7f32e-2be0-476f-9043-9446462ae7df" containerName="dnsmasq-dns" containerID="cri-o://5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a" gracePeriod=10 Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.145167 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.167989 4978 generic.go:334] "Generic (PLEG): container finished" podID="e73d580e-bb87-4270-aec3-ad2c35da47ad" containerID="e802925c232e3151d8073c8b7c34dc5a6827acd379157082050a59043f6fda37" exitCode=0 Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.168097 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5nqzd" event={"ID":"e73d580e-bb87-4270-aec3-ad2c35da47ad","Type":"ContainerDied","Data":"e802925c232e3151d8073c8b7c34dc5a6827acd379157082050a59043f6fda37"} Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.172501 4978 generic.go:334] "Generic (PLEG): container finished" podID="13e7f32e-2be0-476f-9043-9446462ae7df" containerID="5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a" exitCode=0 Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.172546 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" event={"ID":"13e7f32e-2be0-476f-9043-9446462ae7df","Type":"ContainerDied","Data":"5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a"} Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.172573 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" event={"ID":"13e7f32e-2be0-476f-9043-9446462ae7df","Type":"ContainerDied","Data":"fa02475ae23657c34cd1652041b244739e89eee827c8aa497498b492e0cedfa8"} Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.172593 4978 scope.go:117] "RemoveContainer" containerID="5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.172604 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbc4875f7-rfcdw" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.214239 4978 scope.go:117] "RemoveContainer" containerID="a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.242363 4978 scope.go:117] "RemoveContainer" containerID="5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a" Feb 25 08:39:17 crc kubenswrapper[4978]: E0225 08:39:17.243017 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a\": container with ID starting with 5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a not found: ID does not exist" containerID="5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.243083 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a"} err="failed to get container status \"5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a\": rpc error: code = NotFound desc = could not find container \"5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a\": container with ID starting with 5f67ffcb2bd7b4a1e56978bd1c6cecf7479b04009c5c1fb4a43b4fbe12d2944a not found: ID does not exist" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.243128 4978 scope.go:117] "RemoveContainer" containerID="a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d" Feb 25 08:39:17 crc kubenswrapper[4978]: E0225 08:39:17.243564 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d\": container with ID starting with a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d not found: ID does not exist" containerID="a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.243597 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d"} err="failed to get container status \"a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d\": rpc error: code = NotFound desc = could not find container \"a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d\": container with ID starting with a7cba9015181a8704171a41ed9af86a4982e50da032e293e4ce5e1222a066f6d not found: ID does not exist" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.277057 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrqg2\" (UniqueName: \"kubernetes.io/projected/13e7f32e-2be0-476f-9043-9446462ae7df-kube-api-access-vrqg2\") pod \"13e7f32e-2be0-476f-9043-9446462ae7df\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.277179 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-config\") pod \"13e7f32e-2be0-476f-9043-9446462ae7df\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.277236 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-dns-svc\") pod \"13e7f32e-2be0-476f-9043-9446462ae7df\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.277948 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-sb\") pod \"13e7f32e-2be0-476f-9043-9446462ae7df\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.278066 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-nb\") pod \"13e7f32e-2be0-476f-9043-9446462ae7df\" (UID: \"13e7f32e-2be0-476f-9043-9446462ae7df\") " Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.283241 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13e7f32e-2be0-476f-9043-9446462ae7df-kube-api-access-vrqg2" (OuterVolumeSpecName: "kube-api-access-vrqg2") pod "13e7f32e-2be0-476f-9043-9446462ae7df" (UID: "13e7f32e-2be0-476f-9043-9446462ae7df"). InnerVolumeSpecName "kube-api-access-vrqg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.311262 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "13e7f32e-2be0-476f-9043-9446462ae7df" (UID: "13e7f32e-2be0-476f-9043-9446462ae7df"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.347810 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "13e7f32e-2be0-476f-9043-9446462ae7df" (UID: "13e7f32e-2be0-476f-9043-9446462ae7df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.348940 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "13e7f32e-2be0-476f-9043-9446462ae7df" (UID: "13e7f32e-2be0-476f-9043-9446462ae7df"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.350331 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-config" (OuterVolumeSpecName: "config") pod "13e7f32e-2be0-476f-9043-9446462ae7df" (UID: "13e7f32e-2be0-476f-9043-9446462ae7df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.380510 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.380547 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrqg2\" (UniqueName: \"kubernetes.io/projected/13e7f32e-2be0-476f-9043-9446462ae7df-kube-api-access-vrqg2\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.380565 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.380576 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.380588 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/13e7f32e-2be0-476f-9043-9446462ae7df-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.498699 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbc4875f7-rfcdw"] Feb 25 08:39:17 crc kubenswrapper[4978]: I0225 08:39:17.504678 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbc4875f7-rfcdw"] Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.547980 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.601947 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-scripts\") pod \"e73d580e-bb87-4270-aec3-ad2c35da47ad\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.602081 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-credential-keys\") pod \"e73d580e-bb87-4270-aec3-ad2c35da47ad\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.602173 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpkxf\" (UniqueName: \"kubernetes.io/projected/e73d580e-bb87-4270-aec3-ad2c35da47ad-kube-api-access-bpkxf\") pod \"e73d580e-bb87-4270-aec3-ad2c35da47ad\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.602213 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-combined-ca-bundle\") pod \"e73d580e-bb87-4270-aec3-ad2c35da47ad\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.602244 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-config-data\") pod \"e73d580e-bb87-4270-aec3-ad2c35da47ad\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.602269 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-fernet-keys\") pod \"e73d580e-bb87-4270-aec3-ad2c35da47ad\" (UID: \"e73d580e-bb87-4270-aec3-ad2c35da47ad\") " Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.608162 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e73d580e-bb87-4270-aec3-ad2c35da47ad" (UID: "e73d580e-bb87-4270-aec3-ad2c35da47ad"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.609621 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-scripts" (OuterVolumeSpecName: "scripts") pod "e73d580e-bb87-4270-aec3-ad2c35da47ad" (UID: "e73d580e-bb87-4270-aec3-ad2c35da47ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.609784 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e73d580e-bb87-4270-aec3-ad2c35da47ad-kube-api-access-bpkxf" (OuterVolumeSpecName: "kube-api-access-bpkxf") pod "e73d580e-bb87-4270-aec3-ad2c35da47ad" (UID: "e73d580e-bb87-4270-aec3-ad2c35da47ad"). InnerVolumeSpecName "kube-api-access-bpkxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.612322 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e73d580e-bb87-4270-aec3-ad2c35da47ad" (UID: "e73d580e-bb87-4270-aec3-ad2c35da47ad"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.646704 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e73d580e-bb87-4270-aec3-ad2c35da47ad" (UID: "e73d580e-bb87-4270-aec3-ad2c35da47ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.649975 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-config-data" (OuterVolumeSpecName: "config-data") pod "e73d580e-bb87-4270-aec3-ad2c35da47ad" (UID: "e73d580e-bb87-4270-aec3-ad2c35da47ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.704341 4978 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.704420 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpkxf\" (UniqueName: \"kubernetes.io/projected/e73d580e-bb87-4270-aec3-ad2c35da47ad-kube-api-access-bpkxf\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.704448 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.704465 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.704479 4978 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:18 crc kubenswrapper[4978]: I0225 08:39:18.704497 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e73d580e-bb87-4270-aec3-ad2c35da47ad-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.200032 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5nqzd" event={"ID":"e73d580e-bb87-4270-aec3-ad2c35da47ad","Type":"ContainerDied","Data":"51d46eef1e9bab11b3c14e96036706b0ca721871ebad6e467f3b4272d8b28b64"} Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.200486 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51d46eef1e9bab11b3c14e96036706b0ca721871ebad6e467f3b4272d8b28b64" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.200226 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5nqzd" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.347756 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13e7f32e-2be0-476f-9043-9446462ae7df" path="/var/lib/kubelet/pods/13e7f32e-2be0-476f-9043-9446462ae7df/volumes" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.402306 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8d576976-9bm5w"] Feb 25 08:39:19 crc kubenswrapper[4978]: E0225 08:39:19.402664 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e7f32e-2be0-476f-9043-9446462ae7df" containerName="dnsmasq-dns" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.402683 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e7f32e-2be0-476f-9043-9446462ae7df" containerName="dnsmasq-dns" Feb 25 08:39:19 crc kubenswrapper[4978]: E0225 08:39:19.402708 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13e7f32e-2be0-476f-9043-9446462ae7df" containerName="init" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.402717 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="13e7f32e-2be0-476f-9043-9446462ae7df" containerName="init" Feb 25 08:39:19 crc kubenswrapper[4978]: E0225 08:39:19.402733 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e73d580e-bb87-4270-aec3-ad2c35da47ad" containerName="keystone-bootstrap" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.402741 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e73d580e-bb87-4270-aec3-ad2c35da47ad" containerName="keystone-bootstrap" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.402936 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e73d580e-bb87-4270-aec3-ad2c35da47ad" containerName="keystone-bootstrap" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.402952 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="13e7f32e-2be0-476f-9043-9446462ae7df" containerName="dnsmasq-dns" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.403760 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.410736 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.410882 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.411346 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.411497 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.411672 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fnhrp" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.411930 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.419342 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8d576976-9bm5w"] Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.528877 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-internal-tls-certs\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.528932 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-config-data\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.528982 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-combined-ca-bundle\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.529012 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw29c\" (UniqueName: \"kubernetes.io/projected/25443aff-625b-4e6a-b03c-1697c5e2199e-kube-api-access-fw29c\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.529093 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-public-tls-certs\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.529122 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-scripts\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.529146 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-credential-keys\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.529444 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-fernet-keys\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.631723 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-internal-tls-certs\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.632084 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-config-data\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.632167 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-combined-ca-bundle\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.632212 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw29c\" (UniqueName: \"kubernetes.io/projected/25443aff-625b-4e6a-b03c-1697c5e2199e-kube-api-access-fw29c\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.632276 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-public-tls-certs\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.632325 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-scripts\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.632362 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-credential-keys\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.632477 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-fernet-keys\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.636585 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-internal-tls-certs\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.637238 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-public-tls-certs\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.637755 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-config-data\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.638149 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-scripts\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.638998 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-credential-keys\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.639949 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-fernet-keys\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.649343 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25443aff-625b-4e6a-b03c-1697c5e2199e-combined-ca-bundle\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.662186 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw29c\" (UniqueName: \"kubernetes.io/projected/25443aff-625b-4e6a-b03c-1697c5e2199e-kube-api-access-fw29c\") pod \"keystone-8d576976-9bm5w\" (UID: \"25443aff-625b-4e6a-b03c-1697c5e2199e\") " pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:19 crc kubenswrapper[4978]: I0225 08:39:19.722108 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:20 crc kubenswrapper[4978]: I0225 08:39:20.077741 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8d576976-9bm5w"] Feb 25 08:39:20 crc kubenswrapper[4978]: I0225 08:39:20.210446 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8d576976-9bm5w" event={"ID":"25443aff-625b-4e6a-b03c-1697c5e2199e","Type":"ContainerStarted","Data":"2c2dd7408cd658b2aee8c6b2aa43ab753b77767e4cafeda65f06d7a679a7ee9b"} Feb 25 08:39:21 crc kubenswrapper[4978]: I0225 08:39:21.227591 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8d576976-9bm5w" event={"ID":"25443aff-625b-4e6a-b03c-1697c5e2199e","Type":"ContainerStarted","Data":"1ef8f2abce7db4c97a10739e887e6b4b8e60f6ecc0a3d6a0ca8e7cbd2d739490"} Feb 25 08:39:21 crc kubenswrapper[4978]: I0225 08:39:21.228076 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:21 crc kubenswrapper[4978]: I0225 08:39:21.262514 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-8d576976-9bm5w" podStartSLOduration=2.262451861 podStartE2EDuration="2.262451861s" podCreationTimestamp="2026-02-25 08:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:39:21.252828552 +0000 UTC m=+6854.692085051" watchObservedRunningTime="2026-02-25 08:39:21.262451861 +0000 UTC m=+6854.701708320" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.189979 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9g2h5"] Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.192980 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.214342 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9g2h5"] Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.314722 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-catalog-content\") pod \"community-operators-9g2h5\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.314827 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-utilities\") pod \"community-operators-9g2h5\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.314909 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5ppm\" (UniqueName: \"kubernetes.io/projected/27a5109a-8a95-4830-8b4e-b27118090f36-kube-api-access-t5ppm\") pod \"community-operators-9g2h5\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.416709 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-catalog-content\") pod \"community-operators-9g2h5\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.416920 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-utilities\") pod \"community-operators-9g2h5\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.417127 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5ppm\" (UniqueName: \"kubernetes.io/projected/27a5109a-8a95-4830-8b4e-b27118090f36-kube-api-access-t5ppm\") pod \"community-operators-9g2h5\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.417219 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-catalog-content\") pod \"community-operators-9g2h5\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.417687 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-utilities\") pod \"community-operators-9g2h5\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.441684 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5ppm\" (UniqueName: \"kubernetes.io/projected/27a5109a-8a95-4830-8b4e-b27118090f36-kube-api-access-t5ppm\") pod \"community-operators-9g2h5\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:43 crc kubenswrapper[4978]: I0225 08:39:43.558589 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:44 crc kubenswrapper[4978]: I0225 08:39:44.091232 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9g2h5"] Feb 25 08:39:44 crc kubenswrapper[4978]: I0225 08:39:44.462208 4978 generic.go:334] "Generic (PLEG): container finished" podID="27a5109a-8a95-4830-8b4e-b27118090f36" containerID="fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a" exitCode=0 Feb 25 08:39:44 crc kubenswrapper[4978]: I0225 08:39:44.462351 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g2h5" event={"ID":"27a5109a-8a95-4830-8b4e-b27118090f36","Type":"ContainerDied","Data":"fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a"} Feb 25 08:39:44 crc kubenswrapper[4978]: I0225 08:39:44.462761 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g2h5" event={"ID":"27a5109a-8a95-4830-8b4e-b27118090f36","Type":"ContainerStarted","Data":"fe434f909e1dec5e373f6112b958373f32fd6723e178dd1f18fa1c56609e7879"} Feb 25 08:39:45 crc kubenswrapper[4978]: I0225 08:39:45.507257 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g2h5" event={"ID":"27a5109a-8a95-4830-8b4e-b27118090f36","Type":"ContainerStarted","Data":"9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673"} Feb 25 08:39:46 crc kubenswrapper[4978]: I0225 08:39:46.523539 4978 generic.go:334] "Generic (PLEG): container finished" podID="27a5109a-8a95-4830-8b4e-b27118090f36" containerID="9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673" exitCode=0 Feb 25 08:39:46 crc kubenswrapper[4978]: I0225 08:39:46.523644 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g2h5" event={"ID":"27a5109a-8a95-4830-8b4e-b27118090f36","Type":"ContainerDied","Data":"9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673"} Feb 25 08:39:46 crc kubenswrapper[4978]: I0225 08:39:46.540668 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:39:46 crc kubenswrapper[4978]: I0225 08:39:46.540727 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:39:47 crc kubenswrapper[4978]: I0225 08:39:47.534646 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g2h5" event={"ID":"27a5109a-8a95-4830-8b4e-b27118090f36","Type":"ContainerStarted","Data":"37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c"} Feb 25 08:39:47 crc kubenswrapper[4978]: I0225 08:39:47.575023 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9g2h5" podStartSLOduration=2.109294276 podStartE2EDuration="4.574995249s" podCreationTimestamp="2026-02-25 08:39:43 +0000 UTC" firstStartedPulling="2026-02-25 08:39:44.465639413 +0000 UTC m=+6877.904895902" lastFinishedPulling="2026-02-25 08:39:46.931340376 +0000 UTC m=+6880.370596875" observedRunningTime="2026-02-25 08:39:47.564175793 +0000 UTC m=+6881.003432322" watchObservedRunningTime="2026-02-25 08:39:47.574995249 +0000 UTC m=+6881.014251738" Feb 25 08:39:51 crc kubenswrapper[4978]: I0225 08:39:51.214801 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-8d576976-9bm5w" Feb 25 08:39:53 crc kubenswrapper[4978]: I0225 08:39:53.559237 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:53 crc kubenswrapper[4978]: I0225 08:39:53.559573 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:53 crc kubenswrapper[4978]: I0225 08:39:53.648205 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:53 crc kubenswrapper[4978]: I0225 08:39:53.720586 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:53 crc kubenswrapper[4978]: I0225 08:39:53.906565 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9g2h5"] Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.403740 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.405049 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.407507 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-m6bhn" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.407613 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.410324 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.410673 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.423508 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.431622 4978 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e916549b-fbf7-475d-bb6f-c4995c8fceab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T08:39:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T08:39:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T08:39:55Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-25T08:39:55Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.rdoproject.org/podified-antelope-centos9/openstack-openstackclient:4198fe0138d2c46ae51eb6e53fed5006\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem\\\",\\\"name\\\":\\\"combined-ca-bundle\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xgm7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-25T08:39:55Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Feb 25 08:39:55 crc kubenswrapper[4978]: E0225 08:39:55.433126 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-4xgm7 openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[combined-ca-bundle kube-api-access-4xgm7 openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="e916549b-fbf7-475d-bb6f-c4995c8fceab" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.435740 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.452442 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.453554 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.457995 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.467877 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e916549b-fbf7-475d-bb6f-c4995c8fceab" podUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.579307 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hf22\" (UniqueName: \"kubernetes.io/projected/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-kube-api-access-7hf22\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.579609 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.579724 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config-secret\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.579830 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.599525 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.599714 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9g2h5" podUID="27a5109a-8a95-4830-8b4e-b27118090f36" containerName="registry-server" containerID="cri-o://37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c" gracePeriod=2 Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.605998 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e916549b-fbf7-475d-bb6f-c4995c8fceab" podUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.608483 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.613021 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e916549b-fbf7-475d-bb6f-c4995c8fceab" podUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.681396 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hf22\" (UniqueName: \"kubernetes.io/projected/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-kube-api-access-7hf22\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.681684 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.681806 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config-secret\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.681921 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.682788 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.689158 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.689308 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config-secret\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.699693 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hf22\" (UniqueName: \"kubernetes.io/projected/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-kube-api-access-7hf22\") pod \"openstackclient\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " pod="openstack/openstackclient" Feb 25 08:39:55 crc kubenswrapper[4978]: I0225 08:39:55.767940 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.103964 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.105421 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-utilities\") pod \"27a5109a-8a95-4830-8b4e-b27118090f36\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.106355 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-utilities" (OuterVolumeSpecName: "utilities") pod "27a5109a-8a95-4830-8b4e-b27118090f36" (UID: "27a5109a-8a95-4830-8b4e-b27118090f36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.206583 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5ppm\" (UniqueName: \"kubernetes.io/projected/27a5109a-8a95-4830-8b4e-b27118090f36-kube-api-access-t5ppm\") pod \"27a5109a-8a95-4830-8b4e-b27118090f36\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.206734 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-catalog-content\") pod \"27a5109a-8a95-4830-8b4e-b27118090f36\" (UID: \"27a5109a-8a95-4830-8b4e-b27118090f36\") " Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.207146 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.212719 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27a5109a-8a95-4830-8b4e-b27118090f36-kube-api-access-t5ppm" (OuterVolumeSpecName: "kube-api-access-t5ppm") pod "27a5109a-8a95-4830-8b4e-b27118090f36" (UID: "27a5109a-8a95-4830-8b4e-b27118090f36"). InnerVolumeSpecName "kube-api-access-t5ppm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.287970 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.308736 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5ppm\" (UniqueName: \"kubernetes.io/projected/27a5109a-8a95-4830-8b4e-b27118090f36-kube-api-access-t5ppm\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.499070 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27a5109a-8a95-4830-8b4e-b27118090f36" (UID: "27a5109a-8a95-4830-8b4e-b27118090f36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.513638 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27a5109a-8a95-4830-8b4e-b27118090f36-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.612506 4978 generic.go:334] "Generic (PLEG): container finished" podID="27a5109a-8a95-4830-8b4e-b27118090f36" containerID="37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c" exitCode=0 Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.612587 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g2h5" event={"ID":"27a5109a-8a95-4830-8b4e-b27118090f36","Type":"ContainerDied","Data":"37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c"} Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.612622 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9g2h5" event={"ID":"27a5109a-8a95-4830-8b4e-b27118090f36","Type":"ContainerDied","Data":"fe434f909e1dec5e373f6112b958373f32fd6723e178dd1f18fa1c56609e7879"} Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.612647 4978 scope.go:117] "RemoveContainer" containerID="37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.612810 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9g2h5" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.617571 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.617816 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"71c9961c-3b9f-4d2a-b392-abd5eb920a3b","Type":"ContainerStarted","Data":"d88b58e4a6d911aa2f4a86359719e541fd406da4ca858168c9a0413a81b23dbb"} Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.621627 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e916549b-fbf7-475d-bb6f-c4995c8fceab" podUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.642339 4978 scope.go:117] "RemoveContainer" containerID="9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.652547 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e916549b-fbf7-475d-bb6f-c4995c8fceab" podUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.660244 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9g2h5"] Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.665800 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9g2h5"] Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.681511 4978 scope.go:117] "RemoveContainer" containerID="fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.701379 4978 scope.go:117] "RemoveContainer" containerID="37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c" Feb 25 08:39:56 crc kubenswrapper[4978]: E0225 08:39:56.702403 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c\": container with ID starting with 37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c not found: ID does not exist" containerID="37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.702463 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c"} err="failed to get container status \"37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c\": rpc error: code = NotFound desc = could not find container \"37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c\": container with ID starting with 37887a98a265cc4d768213c99431d832a67c8600c2ca7361bad05a0ff394ee8c not found: ID does not exist" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.702495 4978 scope.go:117] "RemoveContainer" containerID="9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673" Feb 25 08:39:56 crc kubenswrapper[4978]: E0225 08:39:56.702900 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673\": container with ID starting with 9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673 not found: ID does not exist" containerID="9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.702937 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673"} err="failed to get container status \"9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673\": rpc error: code = NotFound desc = could not find container \"9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673\": container with ID starting with 9943db18ccd1cb919b652512120fcb783f7d6812017ef8c1af014af8190d0673 not found: ID does not exist" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.702966 4978 scope.go:117] "RemoveContainer" containerID="fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a" Feb 25 08:39:56 crc kubenswrapper[4978]: E0225 08:39:56.703587 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a\": container with ID starting with fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a not found: ID does not exist" containerID="fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a" Feb 25 08:39:56 crc kubenswrapper[4978]: I0225 08:39:56.703775 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a"} err="failed to get container status \"fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a\": rpc error: code = NotFound desc = could not find container \"fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a\": container with ID starting with fe5d31bc0175518cc26296eca7d94702cfbd729f193cd41a63702753c3c2209a not found: ID does not exist" Feb 25 08:39:57 crc kubenswrapper[4978]: I0225 08:39:57.358075 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27a5109a-8a95-4830-8b4e-b27118090f36" path="/var/lib/kubelet/pods/27a5109a-8a95-4830-8b4e-b27118090f36/volumes" Feb 25 08:39:57 crc kubenswrapper[4978]: I0225 08:39:57.358777 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e916549b-fbf7-475d-bb6f-c4995c8fceab" path="/var/lib/kubelet/pods/e916549b-fbf7-475d-bb6f-c4995c8fceab/volumes" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.130259 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533480-txzm8"] Feb 25 08:40:00 crc kubenswrapper[4978]: E0225 08:40:00.132806 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27a5109a-8a95-4830-8b4e-b27118090f36" containerName="extract-content" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.132828 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="27a5109a-8a95-4830-8b4e-b27118090f36" containerName="extract-content" Feb 25 08:40:00 crc kubenswrapper[4978]: E0225 08:40:00.132843 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27a5109a-8a95-4830-8b4e-b27118090f36" containerName="registry-server" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.132850 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="27a5109a-8a95-4830-8b4e-b27118090f36" containerName="registry-server" Feb 25 08:40:00 crc kubenswrapper[4978]: E0225 08:40:00.132867 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27a5109a-8a95-4830-8b4e-b27118090f36" containerName="extract-utilities" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.132873 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="27a5109a-8a95-4830-8b4e-b27118090f36" containerName="extract-utilities" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.133018 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="27a5109a-8a95-4830-8b4e-b27118090f36" containerName="registry-server" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.133548 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533480-txzm8" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.135024 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.135277 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.136013 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.144385 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533480-txzm8"] Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.203147 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqhd5\" (UniqueName: \"kubernetes.io/projected/0498e889-eade-48ed-8b3f-7f7b2dcb9b00-kube-api-access-sqhd5\") pod \"auto-csr-approver-29533480-txzm8\" (UID: \"0498e889-eade-48ed-8b3f-7f7b2dcb9b00\") " pod="openshift-infra/auto-csr-approver-29533480-txzm8" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.304754 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqhd5\" (UniqueName: \"kubernetes.io/projected/0498e889-eade-48ed-8b3f-7f7b2dcb9b00-kube-api-access-sqhd5\") pod \"auto-csr-approver-29533480-txzm8\" (UID: \"0498e889-eade-48ed-8b3f-7f7b2dcb9b00\") " pod="openshift-infra/auto-csr-approver-29533480-txzm8" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.325699 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqhd5\" (UniqueName: \"kubernetes.io/projected/0498e889-eade-48ed-8b3f-7f7b2dcb9b00-kube-api-access-sqhd5\") pod \"auto-csr-approver-29533480-txzm8\" (UID: \"0498e889-eade-48ed-8b3f-7f7b2dcb9b00\") " pod="openshift-infra/auto-csr-approver-29533480-txzm8" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.453609 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533480-txzm8" Feb 25 08:40:00 crc kubenswrapper[4978]: I0225 08:40:00.901746 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533480-txzm8"] Feb 25 08:40:01 crc kubenswrapper[4978]: I0225 08:40:01.657664 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533480-txzm8" event={"ID":"0498e889-eade-48ed-8b3f-7f7b2dcb9b00","Type":"ContainerStarted","Data":"9783c81b9449952b4dfbe980f3b95d51109f0a712cffc986a7f0307600d5d792"} Feb 25 08:40:07 crc kubenswrapper[4978]: I0225 08:40:07.715077 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"71c9961c-3b9f-4d2a-b392-abd5eb920a3b","Type":"ContainerStarted","Data":"51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0"} Feb 25 08:40:07 crc kubenswrapper[4978]: I0225 08:40:07.719945 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533480-txzm8" event={"ID":"0498e889-eade-48ed-8b3f-7f7b2dcb9b00","Type":"ContainerStarted","Data":"e89fc75a2be6143a30a68d5f6e274dc12ab6555b340c3ef7e9f74887319d1034"} Feb 25 08:40:07 crc kubenswrapper[4978]: I0225 08:40:07.743052 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.634014708 podStartE2EDuration="12.743031204s" podCreationTimestamp="2026-02-25 08:39:55 +0000 UTC" firstStartedPulling="2026-02-25 08:39:56.29988162 +0000 UTC m=+6889.739138079" lastFinishedPulling="2026-02-25 08:40:07.408898106 +0000 UTC m=+6900.848154575" observedRunningTime="2026-02-25 08:40:07.740303369 +0000 UTC m=+6901.179559848" watchObservedRunningTime="2026-02-25 08:40:07.743031204 +0000 UTC m=+6901.182287673" Feb 25 08:40:07 crc kubenswrapper[4978]: I0225 08:40:07.760444 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533480-txzm8" podStartSLOduration=1.2782568539999999 podStartE2EDuration="7.760429143s" podCreationTimestamp="2026-02-25 08:40:00 +0000 UTC" firstStartedPulling="2026-02-25 08:40:00.912748654 +0000 UTC m=+6894.352005113" lastFinishedPulling="2026-02-25 08:40:07.394920953 +0000 UTC m=+6900.834177402" observedRunningTime="2026-02-25 08:40:07.757626477 +0000 UTC m=+6901.196882946" watchObservedRunningTime="2026-02-25 08:40:07.760429143 +0000 UTC m=+6901.199685602" Feb 25 08:40:08 crc kubenswrapper[4978]: I0225 08:40:08.731602 4978 generic.go:334] "Generic (PLEG): container finished" podID="0498e889-eade-48ed-8b3f-7f7b2dcb9b00" containerID="e89fc75a2be6143a30a68d5f6e274dc12ab6555b340c3ef7e9f74887319d1034" exitCode=0 Feb 25 08:40:08 crc kubenswrapper[4978]: I0225 08:40:08.731739 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533480-txzm8" event={"ID":"0498e889-eade-48ed-8b3f-7f7b2dcb9b00","Type":"ContainerDied","Data":"e89fc75a2be6143a30a68d5f6e274dc12ab6555b340c3ef7e9f74887319d1034"} Feb 25 08:40:10 crc kubenswrapper[4978]: I0225 08:40:10.132650 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533480-txzm8" Feb 25 08:40:10 crc kubenswrapper[4978]: I0225 08:40:10.184824 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqhd5\" (UniqueName: \"kubernetes.io/projected/0498e889-eade-48ed-8b3f-7f7b2dcb9b00-kube-api-access-sqhd5\") pod \"0498e889-eade-48ed-8b3f-7f7b2dcb9b00\" (UID: \"0498e889-eade-48ed-8b3f-7f7b2dcb9b00\") " Feb 25 08:40:10 crc kubenswrapper[4978]: I0225 08:40:10.190120 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0498e889-eade-48ed-8b3f-7f7b2dcb9b00-kube-api-access-sqhd5" (OuterVolumeSpecName: "kube-api-access-sqhd5") pod "0498e889-eade-48ed-8b3f-7f7b2dcb9b00" (UID: "0498e889-eade-48ed-8b3f-7f7b2dcb9b00"). InnerVolumeSpecName "kube-api-access-sqhd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:40:10 crc kubenswrapper[4978]: I0225 08:40:10.287115 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqhd5\" (UniqueName: \"kubernetes.io/projected/0498e889-eade-48ed-8b3f-7f7b2dcb9b00-kube-api-access-sqhd5\") on node \"crc\" DevicePath \"\"" Feb 25 08:40:10 crc kubenswrapper[4978]: I0225 08:40:10.466273 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533474-plz6z"] Feb 25 08:40:10 crc kubenswrapper[4978]: I0225 08:40:10.473868 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533474-plz6z"] Feb 25 08:40:10 crc kubenswrapper[4978]: I0225 08:40:10.751823 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533480-txzm8" event={"ID":"0498e889-eade-48ed-8b3f-7f7b2dcb9b00","Type":"ContainerDied","Data":"9783c81b9449952b4dfbe980f3b95d51109f0a712cffc986a7f0307600d5d792"} Feb 25 08:40:10 crc kubenswrapper[4978]: I0225 08:40:10.751883 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9783c81b9449952b4dfbe980f3b95d51109f0a712cffc986a7f0307600d5d792" Feb 25 08:40:10 crc kubenswrapper[4978]: I0225 08:40:10.751891 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533480-txzm8" Feb 25 08:40:11 crc kubenswrapper[4978]: I0225 08:40:11.344883 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f910236e-834f-4eb4-97c5-ca2b8d77546c" path="/var/lib/kubelet/pods/f910236e-834f-4eb4-97c5-ca2b8d77546c/volumes" Feb 25 08:40:16 crc kubenswrapper[4978]: I0225 08:40:16.540167 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:40:16 crc kubenswrapper[4978]: I0225 08:40:16.540870 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:40:16 crc kubenswrapper[4978]: I0225 08:40:16.540934 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 08:40:16 crc kubenswrapper[4978]: I0225 08:40:16.541892 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"914d9830f2d8d5aaf326a3efa925940acc7734a6c345c2cd844ec55b8a401b2b"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 08:40:16 crc kubenswrapper[4978]: I0225 08:40:16.541993 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://914d9830f2d8d5aaf326a3efa925940acc7734a6c345c2cd844ec55b8a401b2b" gracePeriod=600 Feb 25 08:40:16 crc kubenswrapper[4978]: I0225 08:40:16.843679 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="914d9830f2d8d5aaf326a3efa925940acc7734a6c345c2cd844ec55b8a401b2b" exitCode=0 Feb 25 08:40:16 crc kubenswrapper[4978]: I0225 08:40:16.843810 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"914d9830f2d8d5aaf326a3efa925940acc7734a6c345c2cd844ec55b8a401b2b"} Feb 25 08:40:16 crc kubenswrapper[4978]: I0225 08:40:16.844165 4978 scope.go:117] "RemoveContainer" containerID="4465d4982cbe67d278bf6acd486fa2f282355ac75ea29f1281b4cd45cfe6559e" Feb 25 08:40:17 crc kubenswrapper[4978]: I0225 08:40:17.855913 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b"} Feb 25 08:40:27 crc kubenswrapper[4978]: I0225 08:40:27.588908 4978 scope.go:117] "RemoveContainer" containerID="8eeb1a257fd9a2146d07c4ed3d994ce14a1fe84a4bdff4887eaebc530651c80a" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.480024 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-62mng"] Feb 25 08:41:31 crc kubenswrapper[4978]: E0225 08:41:31.484729 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0498e889-eade-48ed-8b3f-7f7b2dcb9b00" containerName="oc" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.484836 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0498e889-eade-48ed-8b3f-7f7b2dcb9b00" containerName="oc" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.485098 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0498e889-eade-48ed-8b3f-7f7b2dcb9b00" containerName="oc" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.485884 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-62mng" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.494466 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-62mng"] Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.572222 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-fe52-account-create-update-fbhjz"] Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.573810 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fe52-account-create-update-fbhjz" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.576546 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.581051 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-fe52-account-create-update-fbhjz"] Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.624823 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvswb\" (UniqueName: \"kubernetes.io/projected/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-kube-api-access-jvswb\") pod \"barbican-fe52-account-create-update-fbhjz\" (UID: \"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4\") " pod="openstack/barbican-fe52-account-create-update-fbhjz" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.624878 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-operator-scripts\") pod \"barbican-fe52-account-create-update-fbhjz\" (UID: \"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4\") " pod="openstack/barbican-fe52-account-create-update-fbhjz" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.625036 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62fca817-3596-45b2-9649-aec49cee89be-operator-scripts\") pod \"barbican-db-create-62mng\" (UID: \"62fca817-3596-45b2-9649-aec49cee89be\") " pod="openstack/barbican-db-create-62mng" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.625179 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whzz9\" (UniqueName: \"kubernetes.io/projected/62fca817-3596-45b2-9649-aec49cee89be-kube-api-access-whzz9\") pod \"barbican-db-create-62mng\" (UID: \"62fca817-3596-45b2-9649-aec49cee89be\") " pod="openstack/barbican-db-create-62mng" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.727022 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvswb\" (UniqueName: \"kubernetes.io/projected/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-kube-api-access-jvswb\") pod \"barbican-fe52-account-create-update-fbhjz\" (UID: \"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4\") " pod="openstack/barbican-fe52-account-create-update-fbhjz" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.727717 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-operator-scripts\") pod \"barbican-fe52-account-create-update-fbhjz\" (UID: \"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4\") " pod="openstack/barbican-fe52-account-create-update-fbhjz" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.727952 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62fca817-3596-45b2-9649-aec49cee89be-operator-scripts\") pod \"barbican-db-create-62mng\" (UID: \"62fca817-3596-45b2-9649-aec49cee89be\") " pod="openstack/barbican-db-create-62mng" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.728174 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whzz9\" (UniqueName: \"kubernetes.io/projected/62fca817-3596-45b2-9649-aec49cee89be-kube-api-access-whzz9\") pod \"barbican-db-create-62mng\" (UID: \"62fca817-3596-45b2-9649-aec49cee89be\") " pod="openstack/barbican-db-create-62mng" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.728872 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-operator-scripts\") pod \"barbican-fe52-account-create-update-fbhjz\" (UID: \"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4\") " pod="openstack/barbican-fe52-account-create-update-fbhjz" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.728894 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62fca817-3596-45b2-9649-aec49cee89be-operator-scripts\") pod \"barbican-db-create-62mng\" (UID: \"62fca817-3596-45b2-9649-aec49cee89be\") " pod="openstack/barbican-db-create-62mng" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.749946 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvswb\" (UniqueName: \"kubernetes.io/projected/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-kube-api-access-jvswb\") pod \"barbican-fe52-account-create-update-fbhjz\" (UID: \"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4\") " pod="openstack/barbican-fe52-account-create-update-fbhjz" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.750208 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whzz9\" (UniqueName: \"kubernetes.io/projected/62fca817-3596-45b2-9649-aec49cee89be-kube-api-access-whzz9\") pod \"barbican-db-create-62mng\" (UID: \"62fca817-3596-45b2-9649-aec49cee89be\") " pod="openstack/barbican-db-create-62mng" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.806056 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-62mng" Feb 25 08:41:31 crc kubenswrapper[4978]: I0225 08:41:31.891624 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fe52-account-create-update-fbhjz" Feb 25 08:41:32 crc kubenswrapper[4978]: I0225 08:41:32.240501 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-62mng"] Feb 25 08:41:32 crc kubenswrapper[4978]: I0225 08:41:32.383309 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-fe52-account-create-update-fbhjz"] Feb 25 08:41:32 crc kubenswrapper[4978]: W0225 08:41:32.383324 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f1472a5_fa2d_42d9_8a56_7ea61b2190f4.slice/crio-4d3477df783c300a8913e48b420ee3eb3dfb50a3bfd025bff2adb8cd42e2acc8 WatchSource:0}: Error finding container 4d3477df783c300a8913e48b420ee3eb3dfb50a3bfd025bff2adb8cd42e2acc8: Status 404 returned error can't find the container with id 4d3477df783c300a8913e48b420ee3eb3dfb50a3bfd025bff2adb8cd42e2acc8 Feb 25 08:41:32 crc kubenswrapper[4978]: I0225 08:41:32.642320 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fe52-account-create-update-fbhjz" event={"ID":"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4","Type":"ContainerStarted","Data":"bafeca1f67cde8e70de199fac502d5aa183a599e2c8d87dfe80da5a94b37a495"} Feb 25 08:41:32 crc kubenswrapper[4978]: I0225 08:41:32.643173 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fe52-account-create-update-fbhjz" event={"ID":"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4","Type":"ContainerStarted","Data":"4d3477df783c300a8913e48b420ee3eb3dfb50a3bfd025bff2adb8cd42e2acc8"} Feb 25 08:41:32 crc kubenswrapper[4978]: I0225 08:41:32.643778 4978 generic.go:334] "Generic (PLEG): container finished" podID="62fca817-3596-45b2-9649-aec49cee89be" containerID="ac931080148ca4131204bdaad50dd5c4b10a2fc070727d947b5bf249a6baefce" exitCode=0 Feb 25 08:41:32 crc kubenswrapper[4978]: I0225 08:41:32.643833 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-62mng" event={"ID":"62fca817-3596-45b2-9649-aec49cee89be","Type":"ContainerDied","Data":"ac931080148ca4131204bdaad50dd5c4b10a2fc070727d947b5bf249a6baefce"} Feb 25 08:41:32 crc kubenswrapper[4978]: I0225 08:41:32.643958 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-62mng" event={"ID":"62fca817-3596-45b2-9649-aec49cee89be","Type":"ContainerStarted","Data":"6bfb0d03fec171988663572272c6db6db34cff997187b3fdfef36879f7aa251a"} Feb 25 08:41:32 crc kubenswrapper[4978]: I0225 08:41:32.686050 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-fe52-account-create-update-fbhjz" podStartSLOduration=1.686015667 podStartE2EDuration="1.686015667s" podCreationTimestamp="2026-02-25 08:41:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:41:32.664551251 +0000 UTC m=+6986.103807700" watchObservedRunningTime="2026-02-25 08:41:32.686015667 +0000 UTC m=+6986.125272136" Feb 25 08:41:33 crc kubenswrapper[4978]: I0225 08:41:33.659132 4978 generic.go:334] "Generic (PLEG): container finished" podID="2f1472a5-fa2d-42d9-8a56-7ea61b2190f4" containerID="bafeca1f67cde8e70de199fac502d5aa183a599e2c8d87dfe80da5a94b37a495" exitCode=0 Feb 25 08:41:33 crc kubenswrapper[4978]: I0225 08:41:33.659248 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fe52-account-create-update-fbhjz" event={"ID":"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4","Type":"ContainerDied","Data":"bafeca1f67cde8e70de199fac502d5aa183a599e2c8d87dfe80da5a94b37a495"} Feb 25 08:41:34 crc kubenswrapper[4978]: I0225 08:41:34.108953 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-62mng" Feb 25 08:41:34 crc kubenswrapper[4978]: I0225 08:41:34.177304 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62fca817-3596-45b2-9649-aec49cee89be-operator-scripts\") pod \"62fca817-3596-45b2-9649-aec49cee89be\" (UID: \"62fca817-3596-45b2-9649-aec49cee89be\") " Feb 25 08:41:34 crc kubenswrapper[4978]: I0225 08:41:34.178325 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62fca817-3596-45b2-9649-aec49cee89be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "62fca817-3596-45b2-9649-aec49cee89be" (UID: "62fca817-3596-45b2-9649-aec49cee89be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:41:34 crc kubenswrapper[4978]: I0225 08:41:34.278589 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whzz9\" (UniqueName: \"kubernetes.io/projected/62fca817-3596-45b2-9649-aec49cee89be-kube-api-access-whzz9\") pod \"62fca817-3596-45b2-9649-aec49cee89be\" (UID: \"62fca817-3596-45b2-9649-aec49cee89be\") " Feb 25 08:41:34 crc kubenswrapper[4978]: I0225 08:41:34.278824 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62fca817-3596-45b2-9649-aec49cee89be-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:34 crc kubenswrapper[4978]: I0225 08:41:34.283807 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62fca817-3596-45b2-9649-aec49cee89be-kube-api-access-whzz9" (OuterVolumeSpecName: "kube-api-access-whzz9") pod "62fca817-3596-45b2-9649-aec49cee89be" (UID: "62fca817-3596-45b2-9649-aec49cee89be"). InnerVolumeSpecName "kube-api-access-whzz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:41:34 crc kubenswrapper[4978]: I0225 08:41:34.380962 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whzz9\" (UniqueName: \"kubernetes.io/projected/62fca817-3596-45b2-9649-aec49cee89be-kube-api-access-whzz9\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:34 crc kubenswrapper[4978]: I0225 08:41:34.673144 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-62mng" event={"ID":"62fca817-3596-45b2-9649-aec49cee89be","Type":"ContainerDied","Data":"6bfb0d03fec171988663572272c6db6db34cff997187b3fdfef36879f7aa251a"} Feb 25 08:41:34 crc kubenswrapper[4978]: I0225 08:41:34.673567 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bfb0d03fec171988663572272c6db6db34cff997187b3fdfef36879f7aa251a" Feb 25 08:41:34 crc kubenswrapper[4978]: I0225 08:41:34.673246 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-62mng" Feb 25 08:41:35 crc kubenswrapper[4978]: I0225 08:41:35.035161 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fe52-account-create-update-fbhjz" Feb 25 08:41:35 crc kubenswrapper[4978]: I0225 08:41:35.195139 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvswb\" (UniqueName: \"kubernetes.io/projected/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-kube-api-access-jvswb\") pod \"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4\" (UID: \"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4\") " Feb 25 08:41:35 crc kubenswrapper[4978]: I0225 08:41:35.195463 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-operator-scripts\") pod \"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4\" (UID: \"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4\") " Feb 25 08:41:35 crc kubenswrapper[4978]: I0225 08:41:35.196812 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2f1472a5-fa2d-42d9-8a56-7ea61b2190f4" (UID: "2f1472a5-fa2d-42d9-8a56-7ea61b2190f4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:41:35 crc kubenswrapper[4978]: I0225 08:41:35.203722 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-kube-api-access-jvswb" (OuterVolumeSpecName: "kube-api-access-jvswb") pod "2f1472a5-fa2d-42d9-8a56-7ea61b2190f4" (UID: "2f1472a5-fa2d-42d9-8a56-7ea61b2190f4"). InnerVolumeSpecName "kube-api-access-jvswb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:41:35 crc kubenswrapper[4978]: I0225 08:41:35.298984 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:35 crc kubenswrapper[4978]: I0225 08:41:35.299044 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvswb\" (UniqueName: \"kubernetes.io/projected/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4-kube-api-access-jvswb\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:35 crc kubenswrapper[4978]: I0225 08:41:35.686305 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fe52-account-create-update-fbhjz" event={"ID":"2f1472a5-fa2d-42d9-8a56-7ea61b2190f4","Type":"ContainerDied","Data":"4d3477df783c300a8913e48b420ee3eb3dfb50a3bfd025bff2adb8cd42e2acc8"} Feb 25 08:41:35 crc kubenswrapper[4978]: I0225 08:41:35.686359 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d3477df783c300a8913e48b420ee3eb3dfb50a3bfd025bff2adb8cd42e2acc8" Feb 25 08:41:35 crc kubenswrapper[4978]: I0225 08:41:35.686436 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fe52-account-create-update-fbhjz" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.880387 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-8dpnv"] Feb 25 08:41:36 crc kubenswrapper[4978]: E0225 08:41:36.881384 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f1472a5-fa2d-42d9-8a56-7ea61b2190f4" containerName="mariadb-account-create-update" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.881402 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f1472a5-fa2d-42d9-8a56-7ea61b2190f4" containerName="mariadb-account-create-update" Feb 25 08:41:36 crc kubenswrapper[4978]: E0225 08:41:36.881440 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fca817-3596-45b2-9649-aec49cee89be" containerName="mariadb-database-create" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.881448 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fca817-3596-45b2-9649-aec49cee89be" containerName="mariadb-database-create" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.881941 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fca817-3596-45b2-9649-aec49cee89be" containerName="mariadb-database-create" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.881977 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f1472a5-fa2d-42d9-8a56-7ea61b2190f4" containerName="mariadb-account-create-update" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.883255 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.887715 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.888293 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rvj76" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.916083 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-8dpnv"] Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.932672 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-db-sync-config-data\") pod \"barbican-db-sync-8dpnv\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.932831 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-combined-ca-bundle\") pod \"barbican-db-sync-8dpnv\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:36 crc kubenswrapper[4978]: I0225 08:41:36.932900 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxghx\" (UniqueName: \"kubernetes.io/projected/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-kube-api-access-vxghx\") pod \"barbican-db-sync-8dpnv\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:37 crc kubenswrapper[4978]: I0225 08:41:37.035843 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-combined-ca-bundle\") pod \"barbican-db-sync-8dpnv\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:37 crc kubenswrapper[4978]: I0225 08:41:37.035970 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxghx\" (UniqueName: \"kubernetes.io/projected/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-kube-api-access-vxghx\") pod \"barbican-db-sync-8dpnv\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:37 crc kubenswrapper[4978]: I0225 08:41:37.036229 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-db-sync-config-data\") pod \"barbican-db-sync-8dpnv\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:37 crc kubenswrapper[4978]: I0225 08:41:37.041573 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-db-sync-config-data\") pod \"barbican-db-sync-8dpnv\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:37 crc kubenswrapper[4978]: I0225 08:41:37.050945 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-combined-ca-bundle\") pod \"barbican-db-sync-8dpnv\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:37 crc kubenswrapper[4978]: I0225 08:41:37.053633 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxghx\" (UniqueName: \"kubernetes.io/projected/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-kube-api-access-vxghx\") pod \"barbican-db-sync-8dpnv\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:37 crc kubenswrapper[4978]: I0225 08:41:37.227741 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:37 crc kubenswrapper[4978]: I0225 08:41:37.754551 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-8dpnv"] Feb 25 08:41:38 crc kubenswrapper[4978]: I0225 08:41:38.717255 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8dpnv" event={"ID":"eb15c2e7-425b-40cd-8cb8-27733c9fe83f","Type":"ContainerStarted","Data":"196de483aaada4618afda8f016013be97d3e43ff06fd97daaf0ecd765aa455e3"} Feb 25 08:41:42 crc kubenswrapper[4978]: I0225 08:41:42.758623 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8dpnv" event={"ID":"eb15c2e7-425b-40cd-8cb8-27733c9fe83f","Type":"ContainerStarted","Data":"f2aafde9a69057765fe9dbf66c856cc1f4937c9d15ca0dadbddb7ea6385f0b22"} Feb 25 08:41:42 crc kubenswrapper[4978]: I0225 08:41:42.799499 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-8dpnv" podStartSLOduration=2.759871869 podStartE2EDuration="6.799476812s" podCreationTimestamp="2026-02-25 08:41:36 +0000 UTC" firstStartedPulling="2026-02-25 08:41:37.755269158 +0000 UTC m=+6991.194525657" lastFinishedPulling="2026-02-25 08:41:41.794874141 +0000 UTC m=+6995.234130600" observedRunningTime="2026-02-25 08:41:42.787651315 +0000 UTC m=+6996.226907834" watchObservedRunningTime="2026-02-25 08:41:42.799476812 +0000 UTC m=+6996.238733291" Feb 25 08:41:44 crc kubenswrapper[4978]: I0225 08:41:44.776825 4978 generic.go:334] "Generic (PLEG): container finished" podID="eb15c2e7-425b-40cd-8cb8-27733c9fe83f" containerID="f2aafde9a69057765fe9dbf66c856cc1f4937c9d15ca0dadbddb7ea6385f0b22" exitCode=0 Feb 25 08:41:44 crc kubenswrapper[4978]: I0225 08:41:44.776925 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8dpnv" event={"ID":"eb15c2e7-425b-40cd-8cb8-27733c9fe83f","Type":"ContainerDied","Data":"f2aafde9a69057765fe9dbf66c856cc1f4937c9d15ca0dadbddb7ea6385f0b22"} Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.166767 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.318571 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-db-sync-config-data\") pod \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.318770 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxghx\" (UniqueName: \"kubernetes.io/projected/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-kube-api-access-vxghx\") pod \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.318828 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-combined-ca-bundle\") pod \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\" (UID: \"eb15c2e7-425b-40cd-8cb8-27733c9fe83f\") " Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.331770 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-kube-api-access-vxghx" (OuterVolumeSpecName: "kube-api-access-vxghx") pod "eb15c2e7-425b-40cd-8cb8-27733c9fe83f" (UID: "eb15c2e7-425b-40cd-8cb8-27733c9fe83f"). InnerVolumeSpecName "kube-api-access-vxghx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.331970 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "eb15c2e7-425b-40cd-8cb8-27733c9fe83f" (UID: "eb15c2e7-425b-40cd-8cb8-27733c9fe83f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.348081 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb15c2e7-425b-40cd-8cb8-27733c9fe83f" (UID: "eb15c2e7-425b-40cd-8cb8-27733c9fe83f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.421602 4978 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.421663 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxghx\" (UniqueName: \"kubernetes.io/projected/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-kube-api-access-vxghx\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.421683 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb15c2e7-425b-40cd-8cb8-27733c9fe83f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.800229 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-8dpnv" event={"ID":"eb15c2e7-425b-40cd-8cb8-27733c9fe83f","Type":"ContainerDied","Data":"196de483aaada4618afda8f016013be97d3e43ff06fd97daaf0ecd765aa455e3"} Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.800664 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="196de483aaada4618afda8f016013be97d3e43ff06fd97daaf0ecd765aa455e3" Feb 25 08:41:46 crc kubenswrapper[4978]: I0225 08:41:46.800348 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-8dpnv" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.072534 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-f99df5645-r62ts"] Feb 25 08:41:47 crc kubenswrapper[4978]: E0225 08:41:47.072903 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb15c2e7-425b-40cd-8cb8-27733c9fe83f" containerName="barbican-db-sync" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.072922 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb15c2e7-425b-40cd-8cb8-27733c9fe83f" containerName="barbican-db-sync" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.073125 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb15c2e7-425b-40cd-8cb8-27733c9fe83f" containerName="barbican-db-sync" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.074166 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.080819 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.081020 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rvj76" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.081230 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.104558 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f99df5645-r62ts"] Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.111238 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5684bb8b46-xpzqh"] Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.112539 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.117655 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.124755 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5684bb8b46-xpzqh"] Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.180176 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-959668997-qwvf2"] Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.181400 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.218070 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-959668997-qwvf2"] Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.238534 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7128dcbe-8297-4c05-ae36-ec9d12364871-logs\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.238588 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0648e0c9-8537-4093-9f70-f91e62f1ff62-config-data-custom\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.238613 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7128dcbe-8297-4c05-ae36-ec9d12364871-config-data\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.238647 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7128dcbe-8297-4c05-ae36-ec9d12364871-combined-ca-bundle\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.238673 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhg8d\" (UniqueName: \"kubernetes.io/projected/0648e0c9-8537-4093-9f70-f91e62f1ff62-kube-api-access-xhg8d\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.238691 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtfdv\" (UniqueName: \"kubernetes.io/projected/7128dcbe-8297-4c05-ae36-ec9d12364871-kube-api-access-gtfdv\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.238725 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0648e0c9-8537-4093-9f70-f91e62f1ff62-combined-ca-bundle\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.238741 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7128dcbe-8297-4c05-ae36-ec9d12364871-config-data-custom\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.238777 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0648e0c9-8537-4093-9f70-f91e62f1ff62-config-data\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.238803 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0648e0c9-8537-4093-9f70-f91e62f1ff62-logs\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.340412 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0648e0c9-8537-4093-9f70-f91e62f1ff62-config-data\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341396 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-dns-svc\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341423 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0648e0c9-8537-4093-9f70-f91e62f1ff62-logs\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341480 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-sb\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341509 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7128dcbe-8297-4c05-ae36-ec9d12364871-logs\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341559 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0648e0c9-8537-4093-9f70-f91e62f1ff62-config-data-custom\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341578 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7128dcbe-8297-4c05-ae36-ec9d12364871-config-data\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341593 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-config\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341633 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-nb\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341662 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7128dcbe-8297-4c05-ae36-ec9d12364871-combined-ca-bundle\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341694 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzglq\" (UniqueName: \"kubernetes.io/projected/7c9fc22b-74e3-40f1-b51d-77b380048ab7-kube-api-access-lzglq\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341722 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhg8d\" (UniqueName: \"kubernetes.io/projected/0648e0c9-8537-4093-9f70-f91e62f1ff62-kube-api-access-xhg8d\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341743 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtfdv\" (UniqueName: \"kubernetes.io/projected/7128dcbe-8297-4c05-ae36-ec9d12364871-kube-api-access-gtfdv\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341797 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0648e0c9-8537-4093-9f70-f91e62f1ff62-combined-ca-bundle\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341818 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7128dcbe-8297-4c05-ae36-ec9d12364871-config-data-custom\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.341891 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0648e0c9-8537-4093-9f70-f91e62f1ff62-logs\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.342176 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7128dcbe-8297-4c05-ae36-ec9d12364871-logs\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.352611 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7128dcbe-8297-4c05-ae36-ec9d12364871-config-data\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.352845 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-66b759cf9d-cxrn2"] Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.354268 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.356992 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.358250 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0648e0c9-8537-4093-9f70-f91e62f1ff62-config-data\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.366276 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhg8d\" (UniqueName: \"kubernetes.io/projected/0648e0c9-8537-4093-9f70-f91e62f1ff62-kube-api-access-xhg8d\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.366297 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0648e0c9-8537-4093-9f70-f91e62f1ff62-combined-ca-bundle\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.367939 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7128dcbe-8297-4c05-ae36-ec9d12364871-combined-ca-bundle\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.367940 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0648e0c9-8537-4093-9f70-f91e62f1ff62-config-data-custom\") pod \"barbican-keystone-listener-5684bb8b46-xpzqh\" (UID: \"0648e0c9-8537-4093-9f70-f91e62f1ff62\") " pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.368552 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7128dcbe-8297-4c05-ae36-ec9d12364871-config-data-custom\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.368609 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtfdv\" (UniqueName: \"kubernetes.io/projected/7128dcbe-8297-4c05-ae36-ec9d12364871-kube-api-access-gtfdv\") pod \"barbican-worker-f99df5645-r62ts\" (UID: \"7128dcbe-8297-4c05-ae36-ec9d12364871\") " pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.374564 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-66b759cf9d-cxrn2"] Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.401066 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f99df5645-r62ts" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.437013 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.443101 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzglq\" (UniqueName: \"kubernetes.io/projected/7c9fc22b-74e3-40f1-b51d-77b380048ab7-kube-api-access-lzglq\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.443168 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2thm\" (UniqueName: \"kubernetes.io/projected/02717b6a-29d5-461d-9167-25e5a27daacc-kube-api-access-c2thm\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.443197 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-combined-ca-bundle\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.443246 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-dns-svc\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.443263 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02717b6a-29d5-461d-9167-25e5a27daacc-logs\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.443286 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data-custom\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.443333 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-sb\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.443390 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.443411 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-config\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.443433 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-nb\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.444087 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-nb\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.445727 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-dns-svc\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.446283 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-sb\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.447341 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-config\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.468272 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzglq\" (UniqueName: \"kubernetes.io/projected/7c9fc22b-74e3-40f1-b51d-77b380048ab7-kube-api-access-lzglq\") pod \"dnsmasq-dns-959668997-qwvf2\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.500904 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.546813 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.546907 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2thm\" (UniqueName: \"kubernetes.io/projected/02717b6a-29d5-461d-9167-25e5a27daacc-kube-api-access-c2thm\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.546937 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-combined-ca-bundle\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.547009 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02717b6a-29d5-461d-9167-25e5a27daacc-logs\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.547039 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data-custom\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.548984 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02717b6a-29d5-461d-9167-25e5a27daacc-logs\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.554861 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data-custom\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.555449 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-combined-ca-bundle\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.557855 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.566785 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2thm\" (UniqueName: \"kubernetes.io/projected/02717b6a-29d5-461d-9167-25e5a27daacc-kube-api-access-c2thm\") pod \"barbican-api-66b759cf9d-cxrn2\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.572007 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.892363 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f99df5645-r62ts"] Feb 25 08:41:47 crc kubenswrapper[4978]: I0225 08:41:47.978506 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5684bb8b46-xpzqh"] Feb 25 08:41:47 crc kubenswrapper[4978]: W0225 08:41:47.986606 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0648e0c9_8537_4093_9f70_f91e62f1ff62.slice/crio-3143c8c8733bc8ccd0ce9cea7012ad9dd17a7e2d395b38d4187f1a38633880ac WatchSource:0}: Error finding container 3143c8c8733bc8ccd0ce9cea7012ad9dd17a7e2d395b38d4187f1a38633880ac: Status 404 returned error can't find the container with id 3143c8c8733bc8ccd0ce9cea7012ad9dd17a7e2d395b38d4187f1a38633880ac Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.052749 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-66b759cf9d-cxrn2"] Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.072311 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-959668997-qwvf2"] Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.830905 4978 generic.go:334] "Generic (PLEG): container finished" podID="7c9fc22b-74e3-40f1-b51d-77b380048ab7" containerID="da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc" exitCode=0 Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.831037 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-959668997-qwvf2" event={"ID":"7c9fc22b-74e3-40f1-b51d-77b380048ab7","Type":"ContainerDied","Data":"da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc"} Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.832648 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-959668997-qwvf2" event={"ID":"7c9fc22b-74e3-40f1-b51d-77b380048ab7","Type":"ContainerStarted","Data":"5241b6cf0ca781a761bdf19233e1576a5ff1e4da7daca4c357f2f770e109e43a"} Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.854583 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66b759cf9d-cxrn2" event={"ID":"02717b6a-29d5-461d-9167-25e5a27daacc","Type":"ContainerStarted","Data":"9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268"} Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.854643 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66b759cf9d-cxrn2" event={"ID":"02717b6a-29d5-461d-9167-25e5a27daacc","Type":"ContainerStarted","Data":"09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8"} Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.854654 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66b759cf9d-cxrn2" event={"ID":"02717b6a-29d5-461d-9167-25e5a27daacc","Type":"ContainerStarted","Data":"45dac08cfe6b60f4a79913434e0ccfdae6616a9fdc7d59a90a7f84a49125e5e0"} Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.854697 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.855401 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.858059 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" event={"ID":"0648e0c9-8537-4093-9f70-f91e62f1ff62","Type":"ContainerStarted","Data":"3143c8c8733bc8ccd0ce9cea7012ad9dd17a7e2d395b38d4187f1a38633880ac"} Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.866605 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f99df5645-r62ts" event={"ID":"7128dcbe-8297-4c05-ae36-ec9d12364871","Type":"ContainerStarted","Data":"0bcb906c852a26aae92d941c5708d973ad44435a1f3748398aeed82afe9171ba"} Feb 25 08:41:48 crc kubenswrapper[4978]: I0225 08:41:48.887272 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-66b759cf9d-cxrn2" podStartSLOduration=1.8872471960000001 podStartE2EDuration="1.887247196s" podCreationTimestamp="2026-02-25 08:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:41:48.88154776 +0000 UTC m=+7002.320804219" watchObservedRunningTime="2026-02-25 08:41:48.887247196 +0000 UTC m=+7002.326503655" Feb 25 08:41:49 crc kubenswrapper[4978]: I0225 08:41:49.876503 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-959668997-qwvf2" event={"ID":"7c9fc22b-74e3-40f1-b51d-77b380048ab7","Type":"ContainerStarted","Data":"701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7"} Feb 25 08:41:49 crc kubenswrapper[4978]: I0225 08:41:49.876821 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:49 crc kubenswrapper[4978]: I0225 08:41:49.879245 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" event={"ID":"0648e0c9-8537-4093-9f70-f91e62f1ff62","Type":"ContainerStarted","Data":"edd86bd4dee0dd8f4eef806d9bfdb845a863f8c9d701b7ce6f15d86c8135d1ad"} Feb 25 08:41:49 crc kubenswrapper[4978]: I0225 08:41:49.880758 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f99df5645-r62ts" event={"ID":"7128dcbe-8297-4c05-ae36-ec9d12364871","Type":"ContainerStarted","Data":"3d19ea96124f3192fcff8a7ea0600ba8a5312f164da82336afa5b64903779e36"} Feb 25 08:41:49 crc kubenswrapper[4978]: I0225 08:41:49.893495 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-959668997-qwvf2" podStartSLOduration=2.893476188 podStartE2EDuration="2.893476188s" podCreationTimestamp="2026-02-25 08:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:41:49.891234318 +0000 UTC m=+7003.330490787" watchObservedRunningTime="2026-02-25 08:41:49.893476188 +0000 UTC m=+7003.332732657" Feb 25 08:41:49 crc kubenswrapper[4978]: I0225 08:41:49.922461 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-f99df5645-r62ts" podStartSLOduration=1.343621334 podStartE2EDuration="2.922437686s" podCreationTimestamp="2026-02-25 08:41:47 +0000 UTC" firstStartedPulling="2026-02-25 08:41:47.901405356 +0000 UTC m=+7001.340661825" lastFinishedPulling="2026-02-25 08:41:49.480221708 +0000 UTC m=+7002.919478177" observedRunningTime="2026-02-25 08:41:49.912476276 +0000 UTC m=+7003.351732745" watchObservedRunningTime="2026-02-25 08:41:49.922437686 +0000 UTC m=+7003.361694175" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.381042 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-c85b7bd74-wf92q"] Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.382388 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.386576 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.387435 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.401992 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c85b7bd74-wf92q"] Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.496913 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-config-data-custom\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.497035 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-public-tls-certs\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.497149 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngvjz\" (UniqueName: \"kubernetes.io/projected/38283175-934a-4e62-a5af-d873db0c22aa-kube-api-access-ngvjz\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.497208 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-internal-tls-certs\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.497231 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-combined-ca-bundle\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.497263 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-config-data\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.497320 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38283175-934a-4e62-a5af-d873db0c22aa-logs\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.599195 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-public-tls-certs\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.599811 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngvjz\" (UniqueName: \"kubernetes.io/projected/38283175-934a-4e62-a5af-d873db0c22aa-kube-api-access-ngvjz\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.600002 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-internal-tls-certs\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.600192 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-combined-ca-bundle\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.600404 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-config-data\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.600626 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38283175-934a-4e62-a5af-d873db0c22aa-logs\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.600798 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-config-data-custom\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.603686 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-internal-tls-certs\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.603739 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38283175-934a-4e62-a5af-d873db0c22aa-logs\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.607838 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-public-tls-certs\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.611804 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-config-data-custom\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.613357 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-config-data\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.615783 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38283175-934a-4e62-a5af-d873db0c22aa-combined-ca-bundle\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.625130 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngvjz\" (UniqueName: \"kubernetes.io/projected/38283175-934a-4e62-a5af-d873db0c22aa-kube-api-access-ngvjz\") pod \"barbican-api-c85b7bd74-wf92q\" (UID: \"38283175-934a-4e62-a5af-d873db0c22aa\") " pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.737203 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.912545 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" event={"ID":"0648e0c9-8537-4093-9f70-f91e62f1ff62","Type":"ContainerStarted","Data":"0948db27c03218b898b951d306c0037264684f27f796abcee7b2556192344680"} Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.924052 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f99df5645-r62ts" event={"ID":"7128dcbe-8297-4c05-ae36-ec9d12364871","Type":"ContainerStarted","Data":"02f9a9c6b5aff6c6e79de3e2f16fc2c49a50cdbe7703dd77e25fdb522c7fbe65"} Feb 25 08:41:50 crc kubenswrapper[4978]: I0225 08:41:50.940155 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5684bb8b46-xpzqh" podStartSLOduration=2.447849354 podStartE2EDuration="3.940130083s" podCreationTimestamp="2026-02-25 08:41:47 +0000 UTC" firstStartedPulling="2026-02-25 08:41:47.988809825 +0000 UTC m=+7001.428066284" lastFinishedPulling="2026-02-25 08:41:49.481090554 +0000 UTC m=+7002.920347013" observedRunningTime="2026-02-25 08:41:50.930396881 +0000 UTC m=+7004.369653350" watchObservedRunningTime="2026-02-25 08:41:50.940130083 +0000 UTC m=+7004.379386542" Feb 25 08:41:51 crc kubenswrapper[4978]: I0225 08:41:51.254677 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-c85b7bd74-wf92q"] Feb 25 08:41:51 crc kubenswrapper[4978]: W0225 08:41:51.260733 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38283175_934a_4e62_a5af_d873db0c22aa.slice/crio-6f618a10ceb7bc8c844d51e6820e2e9ff6677c5cb163c6905a1130c37e2ea14b WatchSource:0}: Error finding container 6f618a10ceb7bc8c844d51e6820e2e9ff6677c5cb163c6905a1130c37e2ea14b: Status 404 returned error can't find the container with id 6f618a10ceb7bc8c844d51e6820e2e9ff6677c5cb163c6905a1130c37e2ea14b Feb 25 08:41:51 crc kubenswrapper[4978]: I0225 08:41:51.936685 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c85b7bd74-wf92q" event={"ID":"38283175-934a-4e62-a5af-d873db0c22aa","Type":"ContainerStarted","Data":"e1981b6947830664d51efd18cd2219c4e665e2875e5915abbedd82b07a2f0ba1"} Feb 25 08:41:51 crc kubenswrapper[4978]: I0225 08:41:51.937198 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c85b7bd74-wf92q" event={"ID":"38283175-934a-4e62-a5af-d873db0c22aa","Type":"ContainerStarted","Data":"0b485fb428a8347b728567a59bbdc586728ca9a887d344717d46bf5a22446150"} Feb 25 08:41:51 crc kubenswrapper[4978]: I0225 08:41:51.937209 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-c85b7bd74-wf92q" event={"ID":"38283175-934a-4e62-a5af-d873db0c22aa","Type":"ContainerStarted","Data":"6f618a10ceb7bc8c844d51e6820e2e9ff6677c5cb163c6905a1130c37e2ea14b"} Feb 25 08:41:51 crc kubenswrapper[4978]: I0225 08:41:51.975627 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-c85b7bd74-wf92q" podStartSLOduration=1.975605601 podStartE2EDuration="1.975605601s" podCreationTimestamp="2026-02-25 08:41:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:41:51.967594973 +0000 UTC m=+7005.406851432" watchObservedRunningTime="2026-02-25 08:41:51.975605601 +0000 UTC m=+7005.414862070" Feb 25 08:41:52 crc kubenswrapper[4978]: I0225 08:41:52.947461 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:52 crc kubenswrapper[4978]: I0225 08:41:52.947538 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:41:57 crc kubenswrapper[4978]: I0225 08:41:57.502589 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:41:57 crc kubenswrapper[4978]: I0225 08:41:57.590072 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869c9ccd85-7hn7z"] Feb 25 08:41:57 crc kubenswrapper[4978]: I0225 08:41:57.590420 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" podUID="9b375187-4921-42b6-a5e3-00fc2cb0c881" containerName="dnsmasq-dns" containerID="cri-o://c48acaea28da976ead56105c2ebbc4573da9737059f67fbb13863a0241a1db5f" gracePeriod=10 Feb 25 08:41:57 crc kubenswrapper[4978]: I0225 08:41:57.998725 4978 generic.go:334] "Generic (PLEG): container finished" podID="9b375187-4921-42b6-a5e3-00fc2cb0c881" containerID="c48acaea28da976ead56105c2ebbc4573da9737059f67fbb13863a0241a1db5f" exitCode=0 Feb 25 08:41:57 crc kubenswrapper[4978]: I0225 08:41:57.999057 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" event={"ID":"9b375187-4921-42b6-a5e3-00fc2cb0c881","Type":"ContainerDied","Data":"c48acaea28da976ead56105c2ebbc4573da9737059f67fbb13863a0241a1db5f"} Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.092863 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.168287 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-nb\") pod \"9b375187-4921-42b6-a5e3-00fc2cb0c881\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.168345 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwnmn\" (UniqueName: \"kubernetes.io/projected/9b375187-4921-42b6-a5e3-00fc2cb0c881-kube-api-access-hwnmn\") pod \"9b375187-4921-42b6-a5e3-00fc2cb0c881\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.168379 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-sb\") pod \"9b375187-4921-42b6-a5e3-00fc2cb0c881\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.168431 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-config\") pod \"9b375187-4921-42b6-a5e3-00fc2cb0c881\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.168453 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-dns-svc\") pod \"9b375187-4921-42b6-a5e3-00fc2cb0c881\" (UID: \"9b375187-4921-42b6-a5e3-00fc2cb0c881\") " Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.174551 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b375187-4921-42b6-a5e3-00fc2cb0c881-kube-api-access-hwnmn" (OuterVolumeSpecName: "kube-api-access-hwnmn") pod "9b375187-4921-42b6-a5e3-00fc2cb0c881" (UID: "9b375187-4921-42b6-a5e3-00fc2cb0c881"). InnerVolumeSpecName "kube-api-access-hwnmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.209037 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9b375187-4921-42b6-a5e3-00fc2cb0c881" (UID: "9b375187-4921-42b6-a5e3-00fc2cb0c881"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.209329 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9b375187-4921-42b6-a5e3-00fc2cb0c881" (UID: "9b375187-4921-42b6-a5e3-00fc2cb0c881"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.210165 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9b375187-4921-42b6-a5e3-00fc2cb0c881" (UID: "9b375187-4921-42b6-a5e3-00fc2cb0c881"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.211211 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-config" (OuterVolumeSpecName: "config") pod "9b375187-4921-42b6-a5e3-00fc2cb0c881" (UID: "9b375187-4921-42b6-a5e3-00fc2cb0c881"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.270304 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.270337 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.270347 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.270359 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwnmn\" (UniqueName: \"kubernetes.io/projected/9b375187-4921-42b6-a5e3-00fc2cb0c881-kube-api-access-hwnmn\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.270379 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b375187-4921-42b6-a5e3-00fc2cb0c881-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:41:58 crc kubenswrapper[4978]: I0225 08:41:58.936424 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:59 crc kubenswrapper[4978]: I0225 08:41:59.014263 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" event={"ID":"9b375187-4921-42b6-a5e3-00fc2cb0c881","Type":"ContainerDied","Data":"9ee140e2a0ab6e00f47ad91c3b13336f5397f6a4264e2a0982d05e04ad50b978"} Feb 25 08:41:59 crc kubenswrapper[4978]: I0225 08:41:59.014398 4978 scope.go:117] "RemoveContainer" containerID="c48acaea28da976ead56105c2ebbc4573da9737059f67fbb13863a0241a1db5f" Feb 25 08:41:59 crc kubenswrapper[4978]: I0225 08:41:59.014449 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869c9ccd85-7hn7z" Feb 25 08:41:59 crc kubenswrapper[4978]: I0225 08:41:59.062600 4978 scope.go:117] "RemoveContainer" containerID="42de0ce6b0ed4214ac56d8732f38bcb69f51d2dcfbad218c53ff0c5e10431866" Feb 25 08:41:59 crc kubenswrapper[4978]: I0225 08:41:59.079510 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869c9ccd85-7hn7z"] Feb 25 08:41:59 crc kubenswrapper[4978]: I0225 08:41:59.098540 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869c9ccd85-7hn7z"] Feb 25 08:41:59 crc kubenswrapper[4978]: I0225 08:41:59.194492 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:41:59 crc kubenswrapper[4978]: I0225 08:41:59.341269 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b375187-4921-42b6-a5e3-00fc2cb0c881" path="/var/lib/kubelet/pods/9b375187-4921-42b6-a5e3-00fc2cb0c881/volumes" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.140420 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533482-wvthz"] Feb 25 08:42:00 crc kubenswrapper[4978]: E0225 08:42:00.140976 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b375187-4921-42b6-a5e3-00fc2cb0c881" containerName="init" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.140990 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b375187-4921-42b6-a5e3-00fc2cb0c881" containerName="init" Feb 25 08:42:00 crc kubenswrapper[4978]: E0225 08:42:00.141002 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b375187-4921-42b6-a5e3-00fc2cb0c881" containerName="dnsmasq-dns" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.141009 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b375187-4921-42b6-a5e3-00fc2cb0c881" containerName="dnsmasq-dns" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.141211 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b375187-4921-42b6-a5e3-00fc2cb0c881" containerName="dnsmasq-dns" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.141828 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533482-wvthz" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.144170 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.144416 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.144853 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.153344 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533482-wvthz"] Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.211490 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md8rh\" (UniqueName: \"kubernetes.io/projected/fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a-kube-api-access-md8rh\") pod \"auto-csr-approver-29533482-wvthz\" (UID: \"fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a\") " pod="openshift-infra/auto-csr-approver-29533482-wvthz" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.313326 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md8rh\" (UniqueName: \"kubernetes.io/projected/fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a-kube-api-access-md8rh\") pod \"auto-csr-approver-29533482-wvthz\" (UID: \"fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a\") " pod="openshift-infra/auto-csr-approver-29533482-wvthz" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.356072 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md8rh\" (UniqueName: \"kubernetes.io/projected/fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a-kube-api-access-md8rh\") pod \"auto-csr-approver-29533482-wvthz\" (UID: \"fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a\") " pod="openshift-infra/auto-csr-approver-29533482-wvthz" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.520985 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533482-wvthz" Feb 25 08:42:00 crc kubenswrapper[4978]: I0225 08:42:00.995999 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533482-wvthz"] Feb 25 08:42:00 crc kubenswrapper[4978]: W0225 08:42:00.998934 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcdffbf5_f70a_4dd9_acc9_2a93f5c9bf2a.slice/crio-52007774004c673222e2b42fa1899e3e5089957582da265f8354b6b8043d4ea5 WatchSource:0}: Error finding container 52007774004c673222e2b42fa1899e3e5089957582da265f8354b6b8043d4ea5: Status 404 returned error can't find the container with id 52007774004c673222e2b42fa1899e3e5089957582da265f8354b6b8043d4ea5 Feb 25 08:42:01 crc kubenswrapper[4978]: I0225 08:42:01.041580 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533482-wvthz" event={"ID":"fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a","Type":"ContainerStarted","Data":"52007774004c673222e2b42fa1899e3e5089957582da265f8354b6b8043d4ea5"} Feb 25 08:42:01 crc kubenswrapper[4978]: I0225 08:42:01.985979 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:42:02 crc kubenswrapper[4978]: I0225 08:42:02.130068 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-c85b7bd74-wf92q" Feb 25 08:42:02 crc kubenswrapper[4978]: I0225 08:42:02.193343 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-66b759cf9d-cxrn2"] Feb 25 08:42:02 crc kubenswrapper[4978]: I0225 08:42:02.193586 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-66b759cf9d-cxrn2" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" containerName="barbican-api-log" containerID="cri-o://09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8" gracePeriod=30 Feb 25 08:42:02 crc kubenswrapper[4978]: I0225 08:42:02.193668 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-66b759cf9d-cxrn2" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" containerName="barbican-api" containerID="cri-o://9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268" gracePeriod=30 Feb 25 08:42:03 crc kubenswrapper[4978]: I0225 08:42:03.075355 4978 generic.go:334] "Generic (PLEG): container finished" podID="02717b6a-29d5-461d-9167-25e5a27daacc" containerID="09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8" exitCode=143 Feb 25 08:42:03 crc kubenswrapper[4978]: I0225 08:42:03.075427 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66b759cf9d-cxrn2" event={"ID":"02717b6a-29d5-461d-9167-25e5a27daacc","Type":"ContainerDied","Data":"09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8"} Feb 25 08:42:03 crc kubenswrapper[4978]: I0225 08:42:03.078100 4978 generic.go:334] "Generic (PLEG): container finished" podID="fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a" containerID="1d5d7c730c018c8087296a28ee707405223e3b0b7b5323b3f21b49ee7c1e7234" exitCode=0 Feb 25 08:42:03 crc kubenswrapper[4978]: I0225 08:42:03.078153 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533482-wvthz" event={"ID":"fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a","Type":"ContainerDied","Data":"1d5d7c730c018c8087296a28ee707405223e3b0b7b5323b3f21b49ee7c1e7234"} Feb 25 08:42:04 crc kubenswrapper[4978]: I0225 08:42:04.592489 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533482-wvthz" Feb 25 08:42:04 crc kubenswrapper[4978]: I0225 08:42:04.711576 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md8rh\" (UniqueName: \"kubernetes.io/projected/fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a-kube-api-access-md8rh\") pod \"fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a\" (UID: \"fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a\") " Feb 25 08:42:04 crc kubenswrapper[4978]: I0225 08:42:04.717771 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a-kube-api-access-md8rh" (OuterVolumeSpecName: "kube-api-access-md8rh") pod "fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a" (UID: "fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a"). InnerVolumeSpecName "kube-api-access-md8rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:42:04 crc kubenswrapper[4978]: I0225 08:42:04.814255 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md8rh\" (UniqueName: \"kubernetes.io/projected/fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a-kube-api-access-md8rh\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:05 crc kubenswrapper[4978]: I0225 08:42:05.108472 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533482-wvthz" event={"ID":"fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a","Type":"ContainerDied","Data":"52007774004c673222e2b42fa1899e3e5089957582da265f8354b6b8043d4ea5"} Feb 25 08:42:05 crc kubenswrapper[4978]: I0225 08:42:05.108535 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52007774004c673222e2b42fa1899e3e5089957582da265f8354b6b8043d4ea5" Feb 25 08:42:05 crc kubenswrapper[4978]: I0225 08:42:05.108576 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533482-wvthz" Feb 25 08:42:05 crc kubenswrapper[4978]: I0225 08:42:05.356121 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-66b759cf9d-cxrn2" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.94:9311/healthcheck\": read tcp 10.217.0.2:53432->10.217.1.94:9311: read: connection reset by peer" Feb 25 08:42:05 crc kubenswrapper[4978]: I0225 08:42:05.356302 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-66b759cf9d-cxrn2" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.94:9311/healthcheck\": read tcp 10.217.0.2:53444->10.217.1.94:9311: read: connection reset by peer" Feb 25 08:42:05 crc kubenswrapper[4978]: I0225 08:42:05.659323 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533476-wz5xc"] Feb 25 08:42:05 crc kubenswrapper[4978]: I0225 08:42:05.666453 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533476-wz5xc"] Feb 25 08:42:05 crc kubenswrapper[4978]: I0225 08:42:05.863780 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.038247 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-combined-ca-bundle\") pod \"02717b6a-29d5-461d-9167-25e5a27daacc\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.038299 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02717b6a-29d5-461d-9167-25e5a27daacc-logs\") pod \"02717b6a-29d5-461d-9167-25e5a27daacc\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.038437 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2thm\" (UniqueName: \"kubernetes.io/projected/02717b6a-29d5-461d-9167-25e5a27daacc-kube-api-access-c2thm\") pod \"02717b6a-29d5-461d-9167-25e5a27daacc\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.038479 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data\") pod \"02717b6a-29d5-461d-9167-25e5a27daacc\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.038564 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data-custom\") pod \"02717b6a-29d5-461d-9167-25e5a27daacc\" (UID: \"02717b6a-29d5-461d-9167-25e5a27daacc\") " Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.040408 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02717b6a-29d5-461d-9167-25e5a27daacc-logs" (OuterVolumeSpecName: "logs") pod "02717b6a-29d5-461d-9167-25e5a27daacc" (UID: "02717b6a-29d5-461d-9167-25e5a27daacc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.048959 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "02717b6a-29d5-461d-9167-25e5a27daacc" (UID: "02717b6a-29d5-461d-9167-25e5a27daacc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.049806 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02717b6a-29d5-461d-9167-25e5a27daacc-kube-api-access-c2thm" (OuterVolumeSpecName: "kube-api-access-c2thm") pod "02717b6a-29d5-461d-9167-25e5a27daacc" (UID: "02717b6a-29d5-461d-9167-25e5a27daacc"). InnerVolumeSpecName "kube-api-access-c2thm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.083569 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02717b6a-29d5-461d-9167-25e5a27daacc" (UID: "02717b6a-29d5-461d-9167-25e5a27daacc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.102271 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data" (OuterVolumeSpecName: "config-data") pod "02717b6a-29d5-461d-9167-25e5a27daacc" (UID: "02717b6a-29d5-461d-9167-25e5a27daacc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.119464 4978 generic.go:334] "Generic (PLEG): container finished" podID="02717b6a-29d5-461d-9167-25e5a27daacc" containerID="9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268" exitCode=0 Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.119521 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66b759cf9d-cxrn2" event={"ID":"02717b6a-29d5-461d-9167-25e5a27daacc","Type":"ContainerDied","Data":"9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268"} Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.119561 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-66b759cf9d-cxrn2" event={"ID":"02717b6a-29d5-461d-9167-25e5a27daacc","Type":"ContainerDied","Data":"45dac08cfe6b60f4a79913434e0ccfdae6616a9fdc7d59a90a7f84a49125e5e0"} Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.119568 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-66b759cf9d-cxrn2" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.119592 4978 scope.go:117] "RemoveContainer" containerID="9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.140293 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.140321 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.140332 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02717b6a-29d5-461d-9167-25e5a27daacc-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.140341 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2thm\" (UniqueName: \"kubernetes.io/projected/02717b6a-29d5-461d-9167-25e5a27daacc-kube-api-access-c2thm\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.140350 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02717b6a-29d5-461d-9167-25e5a27daacc-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.198035 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-66b759cf9d-cxrn2"] Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.203920 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-66b759cf9d-cxrn2"] Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.206297 4978 scope.go:117] "RemoveContainer" containerID="09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.255618 4978 scope.go:117] "RemoveContainer" containerID="9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268" Feb 25 08:42:06 crc kubenswrapper[4978]: E0225 08:42:06.268054 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268\": container with ID starting with 9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268 not found: ID does not exist" containerID="9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.268142 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268"} err="failed to get container status \"9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268\": rpc error: code = NotFound desc = could not find container \"9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268\": container with ID starting with 9d33814737aea39011c5936327c79f85e60ebeb67ef8c3dea560a4e4fe8d9268 not found: ID does not exist" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.268191 4978 scope.go:117] "RemoveContainer" containerID="09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8" Feb 25 08:42:06 crc kubenswrapper[4978]: E0225 08:42:06.271767 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8\": container with ID starting with 09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8 not found: ID does not exist" containerID="09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8" Feb 25 08:42:06 crc kubenswrapper[4978]: I0225 08:42:06.271800 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8"} err="failed to get container status \"09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8\": rpc error: code = NotFound desc = could not find container \"09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8\": container with ID starting with 09551aaa2059e1fce5db0f059f763b72e9b6c13ceecee9288b02757c844f98e8 not found: ID does not exist" Feb 25 08:42:07 crc kubenswrapper[4978]: I0225 08:42:07.359553 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" path="/var/lib/kubelet/pods/02717b6a-29d5-461d-9167-25e5a27daacc/volumes" Feb 25 08:42:07 crc kubenswrapper[4978]: I0225 08:42:07.361759 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55f0901c-b44d-47e6-9dc4-245e937029a6" path="/var/lib/kubelet/pods/55f0901c-b44d-47e6-9dc4-245e937029a6/volumes" Feb 25 08:42:16 crc kubenswrapper[4978]: I0225 08:42:16.540428 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:42:16 crc kubenswrapper[4978]: I0225 08:42:16.541102 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:42:27 crc kubenswrapper[4978]: I0225 08:42:27.731718 4978 scope.go:117] "RemoveContainer" containerID="9e3f906d24a1336c48c4c76cb0fec7f366bdf78252141d85a4a99800e7d37ce1" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.801116 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-75j64"] Feb 25 08:42:44 crc kubenswrapper[4978]: E0225 08:42:44.802819 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" containerName="barbican-api-log" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.802909 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" containerName="barbican-api-log" Feb 25 08:42:44 crc kubenswrapper[4978]: E0225 08:42:44.802996 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" containerName="barbican-api" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.803057 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" containerName="barbican-api" Feb 25 08:42:44 crc kubenswrapper[4978]: E0225 08:42:44.803125 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a" containerName="oc" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.803185 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a" containerName="oc" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.803400 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" containerName="barbican-api" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.803481 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="02717b6a-29d5-461d-9167-25e5a27daacc" containerName="barbican-api-log" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.803551 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a" containerName="oc" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.804134 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-75j64" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.814145 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-75j64"] Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.910341 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-51be-account-create-update-xvkcz"] Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.911359 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-51be-account-create-update-xvkcz" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.913645 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.922556 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-51be-account-create-update-xvkcz"] Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.959625 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvxb9\" (UniqueName: \"kubernetes.io/projected/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-kube-api-access-gvxb9\") pod \"neutron-db-create-75j64\" (UID: \"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5\") " pod="openstack/neutron-db-create-75j64" Feb 25 08:42:44 crc kubenswrapper[4978]: I0225 08:42:44.959975 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-operator-scripts\") pod \"neutron-db-create-75j64\" (UID: \"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5\") " pod="openstack/neutron-db-create-75j64" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.062508 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvxb9\" (UniqueName: \"kubernetes.io/projected/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-kube-api-access-gvxb9\") pod \"neutron-db-create-75j64\" (UID: \"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5\") " pod="openstack/neutron-db-create-75j64" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.062648 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvvw5\" (UniqueName: \"kubernetes.io/projected/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-kube-api-access-rvvw5\") pod \"neutron-51be-account-create-update-xvkcz\" (UID: \"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f\") " pod="openstack/neutron-51be-account-create-update-xvkcz" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.062734 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-operator-scripts\") pod \"neutron-51be-account-create-update-xvkcz\" (UID: \"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f\") " pod="openstack/neutron-51be-account-create-update-xvkcz" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.062790 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-operator-scripts\") pod \"neutron-db-create-75j64\" (UID: \"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5\") " pod="openstack/neutron-db-create-75j64" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.063870 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-operator-scripts\") pod \"neutron-db-create-75j64\" (UID: \"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5\") " pod="openstack/neutron-db-create-75j64" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.080696 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvxb9\" (UniqueName: \"kubernetes.io/projected/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-kube-api-access-gvxb9\") pod \"neutron-db-create-75j64\" (UID: \"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5\") " pod="openstack/neutron-db-create-75j64" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.121009 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-75j64" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.164606 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-operator-scripts\") pod \"neutron-51be-account-create-update-xvkcz\" (UID: \"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f\") " pod="openstack/neutron-51be-account-create-update-xvkcz" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.165014 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvvw5\" (UniqueName: \"kubernetes.io/projected/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-kube-api-access-rvvw5\") pod \"neutron-51be-account-create-update-xvkcz\" (UID: \"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f\") " pod="openstack/neutron-51be-account-create-update-xvkcz" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.165304 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-operator-scripts\") pod \"neutron-51be-account-create-update-xvkcz\" (UID: \"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f\") " pod="openstack/neutron-51be-account-create-update-xvkcz" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.190281 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvvw5\" (UniqueName: \"kubernetes.io/projected/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-kube-api-access-rvvw5\") pod \"neutron-51be-account-create-update-xvkcz\" (UID: \"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f\") " pod="openstack/neutron-51be-account-create-update-xvkcz" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.232378 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-51be-account-create-update-xvkcz" Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.616944 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-75j64"] Feb 25 08:42:45 crc kubenswrapper[4978]: I0225 08:42:45.709771 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-51be-account-create-update-xvkcz"] Feb 25 08:42:45 crc kubenswrapper[4978]: W0225 08:42:45.720878 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fbb51bf_dcc5_47b6_85fd_85b6ed46da1f.slice/crio-5963137d282d3074f212309d11822f539c77b65d919d91cef6de21311993a127 WatchSource:0}: Error finding container 5963137d282d3074f212309d11822f539c77b65d919d91cef6de21311993a127: Status 404 returned error can't find the container with id 5963137d282d3074f212309d11822f539c77b65d919d91cef6de21311993a127 Feb 25 08:42:46 crc kubenswrapper[4978]: I0225 08:42:46.541276 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:42:46 crc kubenswrapper[4978]: I0225 08:42:46.541757 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:42:46 crc kubenswrapper[4978]: I0225 08:42:46.579671 4978 generic.go:334] "Generic (PLEG): container finished" podID="15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5" containerID="708d5bf1739b5b5ad5f990362a7480ef8b7459894fd84af9395a21209a64bc29" exitCode=0 Feb 25 08:42:46 crc kubenswrapper[4978]: I0225 08:42:46.579751 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-75j64" event={"ID":"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5","Type":"ContainerDied","Data":"708d5bf1739b5b5ad5f990362a7480ef8b7459894fd84af9395a21209a64bc29"} Feb 25 08:42:46 crc kubenswrapper[4978]: I0225 08:42:46.579883 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-75j64" event={"ID":"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5","Type":"ContainerStarted","Data":"b48d782ac46e664d70ff188971ab12843e149e1b44813a91e69c9476db507be5"} Feb 25 08:42:46 crc kubenswrapper[4978]: I0225 08:42:46.582075 4978 generic.go:334] "Generic (PLEG): container finished" podID="6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f" containerID="deb0808a6b3beb6afba2ae303af557f432a49b862af7aa8bc0c89290061078fd" exitCode=0 Feb 25 08:42:46 crc kubenswrapper[4978]: I0225 08:42:46.582123 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-51be-account-create-update-xvkcz" event={"ID":"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f","Type":"ContainerDied","Data":"deb0808a6b3beb6afba2ae303af557f432a49b862af7aa8bc0c89290061078fd"} Feb 25 08:42:46 crc kubenswrapper[4978]: I0225 08:42:46.582152 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-51be-account-create-update-xvkcz" event={"ID":"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f","Type":"ContainerStarted","Data":"5963137d282d3074f212309d11822f539c77b65d919d91cef6de21311993a127"} Feb 25 08:42:47 crc kubenswrapper[4978]: I0225 08:42:47.965782 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-75j64" Feb 25 08:42:47 crc kubenswrapper[4978]: I0225 08:42:47.969993 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-51be-account-create-update-xvkcz" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.130702 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-operator-scripts\") pod \"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f\" (UID: \"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f\") " Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.130908 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvxb9\" (UniqueName: \"kubernetes.io/projected/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-kube-api-access-gvxb9\") pod \"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5\" (UID: \"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5\") " Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.130975 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvvw5\" (UniqueName: \"kubernetes.io/projected/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-kube-api-access-rvvw5\") pod \"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f\" (UID: \"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f\") " Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.131021 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-operator-scripts\") pod \"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5\" (UID: \"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5\") " Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.132272 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5" (UID: "15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.132282 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f" (UID: "6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.141592 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-kube-api-access-gvxb9" (OuterVolumeSpecName: "kube-api-access-gvxb9") pod "15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5" (UID: "15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5"). InnerVolumeSpecName "kube-api-access-gvxb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.151894 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-kube-api-access-rvvw5" (OuterVolumeSpecName: "kube-api-access-rvvw5") pod "6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f" (UID: "6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f"). InnerVolumeSpecName "kube-api-access-rvvw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.233592 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvxb9\" (UniqueName: \"kubernetes.io/projected/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-kube-api-access-gvxb9\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.233622 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvvw5\" (UniqueName: \"kubernetes.io/projected/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-kube-api-access-rvvw5\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.233641 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.233651 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.621023 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-75j64" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.621030 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-75j64" event={"ID":"15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5","Type":"ContainerDied","Data":"b48d782ac46e664d70ff188971ab12843e149e1b44813a91e69c9476db507be5"} Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.621623 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b48d782ac46e664d70ff188971ab12843e149e1b44813a91e69c9476db507be5" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.625269 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-51be-account-create-update-xvkcz" event={"ID":"6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f","Type":"ContainerDied","Data":"5963137d282d3074f212309d11822f539c77b65d919d91cef6de21311993a127"} Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.625319 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5963137d282d3074f212309d11822f539c77b65d919d91cef6de21311993a127" Feb 25 08:42:48 crc kubenswrapper[4978]: I0225 08:42:48.625347 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-51be-account-create-update-xvkcz" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.227685 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-f7v86"] Feb 25 08:42:50 crc kubenswrapper[4978]: E0225 08:42:50.228262 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f" containerName="mariadb-account-create-update" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.228286 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f" containerName="mariadb-account-create-update" Feb 25 08:42:50 crc kubenswrapper[4978]: E0225 08:42:50.228345 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5" containerName="mariadb-database-create" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.228359 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5" containerName="mariadb-database-create" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.228665 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5" containerName="mariadb-database-create" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.228722 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f" containerName="mariadb-account-create-update" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.229706 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.234063 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ck88c" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.234543 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.235732 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.236854 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-f7v86"] Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.271781 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-config\") pod \"neutron-db-sync-f7v86\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.271872 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw2kz\" (UniqueName: \"kubernetes.io/projected/5d02cc3c-4121-46d0-ae30-7842907c283f-kube-api-access-vw2kz\") pod \"neutron-db-sync-f7v86\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.271949 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-combined-ca-bundle\") pod \"neutron-db-sync-f7v86\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.373758 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-combined-ca-bundle\") pod \"neutron-db-sync-f7v86\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.374129 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-config\") pod \"neutron-db-sync-f7v86\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.374258 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw2kz\" (UniqueName: \"kubernetes.io/projected/5d02cc3c-4121-46d0-ae30-7842907c283f-kube-api-access-vw2kz\") pod \"neutron-db-sync-f7v86\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.380015 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-combined-ca-bundle\") pod \"neutron-db-sync-f7v86\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.380189 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-config\") pod \"neutron-db-sync-f7v86\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.393022 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw2kz\" (UniqueName: \"kubernetes.io/projected/5d02cc3c-4121-46d0-ae30-7842907c283f-kube-api-access-vw2kz\") pod \"neutron-db-sync-f7v86\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:50 crc kubenswrapper[4978]: I0225 08:42:50.589901 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:51 crc kubenswrapper[4978]: I0225 08:42:51.090530 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-f7v86"] Feb 25 08:42:51 crc kubenswrapper[4978]: W0225 08:42:51.098504 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d02cc3c_4121_46d0_ae30_7842907c283f.slice/crio-0c69786dc6906117c8c78bd5a246482c9f10f01d5e15b37afe198de04e76ef0b WatchSource:0}: Error finding container 0c69786dc6906117c8c78bd5a246482c9f10f01d5e15b37afe198de04e76ef0b: Status 404 returned error can't find the container with id 0c69786dc6906117c8c78bd5a246482c9f10f01d5e15b37afe198de04e76ef0b Feb 25 08:42:51 crc kubenswrapper[4978]: I0225 08:42:51.658182 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-f7v86" event={"ID":"5d02cc3c-4121-46d0-ae30-7842907c283f","Type":"ContainerStarted","Data":"9d3d8fd493ab143cb7a9ab8336c2c8d387260cf743db5bfb345c38c00a8a4783"} Feb 25 08:42:51 crc kubenswrapper[4978]: I0225 08:42:51.659857 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-f7v86" event={"ID":"5d02cc3c-4121-46d0-ae30-7842907c283f","Type":"ContainerStarted","Data":"0c69786dc6906117c8c78bd5a246482c9f10f01d5e15b37afe198de04e76ef0b"} Feb 25 08:42:55 crc kubenswrapper[4978]: I0225 08:42:55.707901 4978 generic.go:334] "Generic (PLEG): container finished" podID="5d02cc3c-4121-46d0-ae30-7842907c283f" containerID="9d3d8fd493ab143cb7a9ab8336c2c8d387260cf743db5bfb345c38c00a8a4783" exitCode=0 Feb 25 08:42:55 crc kubenswrapper[4978]: I0225 08:42:55.707985 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-f7v86" event={"ID":"5d02cc3c-4121-46d0-ae30-7842907c283f","Type":"ContainerDied","Data":"9d3d8fd493ab143cb7a9ab8336c2c8d387260cf743db5bfb345c38c00a8a4783"} Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.192549 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.308115 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-config\") pod \"5d02cc3c-4121-46d0-ae30-7842907c283f\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.308278 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw2kz\" (UniqueName: \"kubernetes.io/projected/5d02cc3c-4121-46d0-ae30-7842907c283f-kube-api-access-vw2kz\") pod \"5d02cc3c-4121-46d0-ae30-7842907c283f\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.308412 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-combined-ca-bundle\") pod \"5d02cc3c-4121-46d0-ae30-7842907c283f\" (UID: \"5d02cc3c-4121-46d0-ae30-7842907c283f\") " Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.318475 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d02cc3c-4121-46d0-ae30-7842907c283f-kube-api-access-vw2kz" (OuterVolumeSpecName: "kube-api-access-vw2kz") pod "5d02cc3c-4121-46d0-ae30-7842907c283f" (UID: "5d02cc3c-4121-46d0-ae30-7842907c283f"). InnerVolumeSpecName "kube-api-access-vw2kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.359412 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d02cc3c-4121-46d0-ae30-7842907c283f" (UID: "5d02cc3c-4121-46d0-ae30-7842907c283f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.362319 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-config" (OuterVolumeSpecName: "config") pod "5d02cc3c-4121-46d0-ae30-7842907c283f" (UID: "5d02cc3c-4121-46d0-ae30-7842907c283f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.410759 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw2kz\" (UniqueName: \"kubernetes.io/projected/5d02cc3c-4121-46d0-ae30-7842907c283f-kube-api-access-vw2kz\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.410815 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.410834 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d02cc3c-4121-46d0-ae30-7842907c283f-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.732129 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-f7v86" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.732137 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-f7v86" event={"ID":"5d02cc3c-4121-46d0-ae30-7842907c283f","Type":"ContainerDied","Data":"0c69786dc6906117c8c78bd5a246482c9f10f01d5e15b37afe198de04e76ef0b"} Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.732197 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c69786dc6906117c8c78bd5a246482c9f10f01d5e15b37afe198de04e76ef0b" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.900817 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-88f9d946c-zq7ns"] Feb 25 08:42:57 crc kubenswrapper[4978]: E0225 08:42:57.901185 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d02cc3c-4121-46d0-ae30-7842907c283f" containerName="neutron-db-sync" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.901196 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d02cc3c-4121-46d0-ae30-7842907c283f" containerName="neutron-db-sync" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.901384 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d02cc3c-4121-46d0-ae30-7842907c283f" containerName="neutron-db-sync" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.902254 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:57 crc kubenswrapper[4978]: I0225 08:42:57.926654 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-88f9d946c-zq7ns"] Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.044321 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8546f94b5d-dnjws"] Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.044836 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-dns-svc\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.044890 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-sb\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.044987 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-config\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.045032 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c99zr\" (UniqueName: \"kubernetes.io/projected/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-kube-api-access-c99zr\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.045095 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-nb\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.046194 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.048782 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.050965 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.051097 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ck88c" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.051232 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.053714 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8546f94b5d-dnjws"] Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.146249 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c99zr\" (UniqueName: \"kubernetes.io/projected/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-kube-api-access-c99zr\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.146298 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-combined-ca-bundle\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.146340 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-httpd-config\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.146491 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-nb\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.146551 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-config\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.146658 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-dns-svc\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.146684 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-sb\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.146806 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-ovndb-tls-certs\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.146849 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wglw8\" (UniqueName: \"kubernetes.io/projected/11829354-abeb-4753-a75a-47f9100fb784-kube-api-access-wglw8\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.146914 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-config\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.147329 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-nb\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.147501 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-sb\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.147707 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-config\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.148167 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-dns-svc\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.164486 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c99zr\" (UniqueName: \"kubernetes.io/projected/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-kube-api-access-c99zr\") pod \"dnsmasq-dns-88f9d946c-zq7ns\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.248166 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-ovndb-tls-certs\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.249307 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wglw8\" (UniqueName: \"kubernetes.io/projected/11829354-abeb-4753-a75a-47f9100fb784-kube-api-access-wglw8\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.249518 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-combined-ca-bundle\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.249655 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-httpd-config\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.249828 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-config\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.253459 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-config\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.253618 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-ovndb-tls-certs\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.254246 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-combined-ca-bundle\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.261122 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-httpd-config\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.264960 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.270075 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wglw8\" (UniqueName: \"kubernetes.io/projected/11829354-abeb-4753-a75a-47f9100fb784-kube-api-access-wglw8\") pod \"neutron-8546f94b5d-dnjws\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.373777 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.532811 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-88f9d946c-zq7ns"] Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.747304 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" event={"ID":"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7","Type":"ContainerStarted","Data":"010482d1aa6e7dd1b33fe88d9c4301d653bf5006fd65a06cff9488eb24d52cc5"} Feb 25 08:42:58 crc kubenswrapper[4978]: I0225 08:42:58.940151 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8546f94b5d-dnjws"] Feb 25 08:42:59 crc kubenswrapper[4978]: I0225 08:42:59.756503 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8546f94b5d-dnjws" event={"ID":"11829354-abeb-4753-a75a-47f9100fb784","Type":"ContainerStarted","Data":"6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce"} Feb 25 08:42:59 crc kubenswrapper[4978]: I0225 08:42:59.757175 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:42:59 crc kubenswrapper[4978]: I0225 08:42:59.757193 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8546f94b5d-dnjws" event={"ID":"11829354-abeb-4753-a75a-47f9100fb784","Type":"ContainerStarted","Data":"563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2"} Feb 25 08:42:59 crc kubenswrapper[4978]: I0225 08:42:59.757206 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8546f94b5d-dnjws" event={"ID":"11829354-abeb-4753-a75a-47f9100fb784","Type":"ContainerStarted","Data":"91e084eb2f6ce873ac0d490568fd79e491ebbf75c06aeee9babf1619786b336a"} Feb 25 08:42:59 crc kubenswrapper[4978]: I0225 08:42:59.758481 4978 generic.go:334] "Generic (PLEG): container finished" podID="ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" containerID="8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5" exitCode=0 Feb 25 08:42:59 crc kubenswrapper[4978]: I0225 08:42:59.758511 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" event={"ID":"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7","Type":"ContainerDied","Data":"8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5"} Feb 25 08:42:59 crc kubenswrapper[4978]: I0225 08:42:59.807323 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8546f94b5d-dnjws" podStartSLOduration=1.807304583 podStartE2EDuration="1.807304583s" podCreationTimestamp="2026-02-25 08:42:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:42:59.789507011 +0000 UTC m=+7073.228763500" watchObservedRunningTime="2026-02-25 08:42:59.807304583 +0000 UTC m=+7073.246561042" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.475859 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7cc578599c-fmh6q"] Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.478043 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.480887 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.481100 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.489836 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7cc578599c-fmh6q"] Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.594177 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-combined-ca-bundle\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.594227 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-config\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.594288 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-internal-tls-certs\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.594312 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-public-tls-certs\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.594518 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-httpd-config\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.594576 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-ovndb-tls-certs\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.594599 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2mf7\" (UniqueName: \"kubernetes.io/projected/882b5c88-3ecd-4baa-91c6-682450eb37fe-kube-api-access-f2mf7\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.695633 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-combined-ca-bundle\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.695690 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-config\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.695768 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-internal-tls-certs\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.695796 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-public-tls-certs\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.695855 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-httpd-config\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.695887 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-ovndb-tls-certs\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.695911 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2mf7\" (UniqueName: \"kubernetes.io/projected/882b5c88-3ecd-4baa-91c6-682450eb37fe-kube-api-access-f2mf7\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.702242 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-internal-tls-certs\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.702273 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-httpd-config\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.702555 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-public-tls-certs\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.706054 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-combined-ca-bundle\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.706418 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-ovndb-tls-certs\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.706861 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/882b5c88-3ecd-4baa-91c6-682450eb37fe-config\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.720084 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2mf7\" (UniqueName: \"kubernetes.io/projected/882b5c88-3ecd-4baa-91c6-682450eb37fe-kube-api-access-f2mf7\") pod \"neutron-7cc578599c-fmh6q\" (UID: \"882b5c88-3ecd-4baa-91c6-682450eb37fe\") " pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.768915 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" event={"ID":"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7","Type":"ContainerStarted","Data":"4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd"} Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.769015 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.793897 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" podStartSLOduration=3.793872616 podStartE2EDuration="3.793872616s" podCreationTimestamp="2026-02-25 08:42:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:43:00.786700573 +0000 UTC m=+7074.225957042" watchObservedRunningTime="2026-02-25 08:43:00.793872616 +0000 UTC m=+7074.233129075" Feb 25 08:43:00 crc kubenswrapper[4978]: I0225 08:43:00.838333 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:01 crc kubenswrapper[4978]: I0225 08:43:01.344179 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7cc578599c-fmh6q"] Feb 25 08:43:01 crc kubenswrapper[4978]: I0225 08:43:01.778117 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cc578599c-fmh6q" event={"ID":"882b5c88-3ecd-4baa-91c6-682450eb37fe","Type":"ContainerStarted","Data":"e0e9109e4d087f7ba82490f24c6ec7802a9507e6780728db319b48c632be49e4"} Feb 25 08:43:01 crc kubenswrapper[4978]: I0225 08:43:01.778355 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:01 crc kubenswrapper[4978]: I0225 08:43:01.778383 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cc578599c-fmh6q" event={"ID":"882b5c88-3ecd-4baa-91c6-682450eb37fe","Type":"ContainerStarted","Data":"cade0975bcf31977241104ec10af6b497d24377fc151f41efbfd4038788eb875"} Feb 25 08:43:01 crc kubenswrapper[4978]: I0225 08:43:01.778393 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7cc578599c-fmh6q" event={"ID":"882b5c88-3ecd-4baa-91c6-682450eb37fe","Type":"ContainerStarted","Data":"80265f7808a14b0f504e8ab34514eab4d3d0ef23af0021c229e81ec70461ff54"} Feb 25 08:43:01 crc kubenswrapper[4978]: I0225 08:43:01.802815 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7cc578599c-fmh6q" podStartSLOduration=1.802798331 podStartE2EDuration="1.802798331s" podCreationTimestamp="2026-02-25 08:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:43:01.798066635 +0000 UTC m=+7075.237323094" watchObservedRunningTime="2026-02-25 08:43:01.802798331 +0000 UTC m=+7075.242054790" Feb 25 08:43:08 crc kubenswrapper[4978]: I0225 08:43:08.266866 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:43:08 crc kubenswrapper[4978]: I0225 08:43:08.349718 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-959668997-qwvf2"] Feb 25 08:43:08 crc kubenswrapper[4978]: I0225 08:43:08.850473 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-959668997-qwvf2" podUID="7c9fc22b-74e3-40f1-b51d-77b380048ab7" containerName="dnsmasq-dns" containerID="cri-o://701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7" gracePeriod=10 Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.305203 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.377143 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-nb\") pod \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.377200 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-dns-svc\") pod \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.377293 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzglq\" (UniqueName: \"kubernetes.io/projected/7c9fc22b-74e3-40f1-b51d-77b380048ab7-kube-api-access-lzglq\") pod \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.377329 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-sb\") pod \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.377357 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-config\") pod \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\" (UID: \"7c9fc22b-74e3-40f1-b51d-77b380048ab7\") " Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.392323 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c9fc22b-74e3-40f1-b51d-77b380048ab7-kube-api-access-lzglq" (OuterVolumeSpecName: "kube-api-access-lzglq") pod "7c9fc22b-74e3-40f1-b51d-77b380048ab7" (UID: "7c9fc22b-74e3-40f1-b51d-77b380048ab7"). InnerVolumeSpecName "kube-api-access-lzglq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.422597 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7c9fc22b-74e3-40f1-b51d-77b380048ab7" (UID: "7c9fc22b-74e3-40f1-b51d-77b380048ab7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.427030 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-config" (OuterVolumeSpecName: "config") pod "7c9fc22b-74e3-40f1-b51d-77b380048ab7" (UID: "7c9fc22b-74e3-40f1-b51d-77b380048ab7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.430660 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7c9fc22b-74e3-40f1-b51d-77b380048ab7" (UID: "7c9fc22b-74e3-40f1-b51d-77b380048ab7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.443408 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7c9fc22b-74e3-40f1-b51d-77b380048ab7" (UID: "7c9fc22b-74e3-40f1-b51d-77b380048ab7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.479811 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.479848 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzglq\" (UniqueName: \"kubernetes.io/projected/7c9fc22b-74e3-40f1-b51d-77b380048ab7-kube-api-access-lzglq\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.479862 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.479876 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.479888 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c9fc22b-74e3-40f1-b51d-77b380048ab7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.865666 4978 generic.go:334] "Generic (PLEG): container finished" podID="7c9fc22b-74e3-40f1-b51d-77b380048ab7" containerID="701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7" exitCode=0 Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.865726 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-959668997-qwvf2" event={"ID":"7c9fc22b-74e3-40f1-b51d-77b380048ab7","Type":"ContainerDied","Data":"701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7"} Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.865764 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-959668997-qwvf2" event={"ID":"7c9fc22b-74e3-40f1-b51d-77b380048ab7","Type":"ContainerDied","Data":"5241b6cf0ca781a761bdf19233e1576a5ff1e4da7daca4c357f2f770e109e43a"} Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.865792 4978 scope.go:117] "RemoveContainer" containerID="701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.865856 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-959668997-qwvf2" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.897804 4978 scope.go:117] "RemoveContainer" containerID="da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.923420 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-959668997-qwvf2"] Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.936579 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-959668997-qwvf2"] Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.945223 4978 scope.go:117] "RemoveContainer" containerID="701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7" Feb 25 08:43:09 crc kubenswrapper[4978]: E0225 08:43:09.945867 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7\": container with ID starting with 701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7 not found: ID does not exist" containerID="701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.945920 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7"} err="failed to get container status \"701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7\": rpc error: code = NotFound desc = could not find container \"701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7\": container with ID starting with 701008ab4ee568c2cd69e116de0411e144cab16d39881881aef87348cf02b0a7 not found: ID does not exist" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.945943 4978 scope.go:117] "RemoveContainer" containerID="da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc" Feb 25 08:43:09 crc kubenswrapper[4978]: E0225 08:43:09.946343 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc\": container with ID starting with da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc not found: ID does not exist" containerID="da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc" Feb 25 08:43:09 crc kubenswrapper[4978]: I0225 08:43:09.946401 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc"} err="failed to get container status \"da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc\": rpc error: code = NotFound desc = could not find container \"da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc\": container with ID starting with da9200c4c9fc18d8affc2ed456d961648199d258b7b8b829cac447a6c8f488cc not found: ID does not exist" Feb 25 08:43:11 crc kubenswrapper[4978]: I0225 08:43:11.344068 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c9fc22b-74e3-40f1-b51d-77b380048ab7" path="/var/lib/kubelet/pods/7c9fc22b-74e3-40f1-b51d-77b380048ab7/volumes" Feb 25 08:43:16 crc kubenswrapper[4978]: I0225 08:43:16.540639 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:43:16 crc kubenswrapper[4978]: I0225 08:43:16.541283 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:43:16 crc kubenswrapper[4978]: I0225 08:43:16.541347 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 08:43:16 crc kubenswrapper[4978]: I0225 08:43:16.542453 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 08:43:16 crc kubenswrapper[4978]: I0225 08:43:16.542552 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" gracePeriod=600 Feb 25 08:43:16 crc kubenswrapper[4978]: E0225 08:43:16.669962 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:43:16 crc kubenswrapper[4978]: I0225 08:43:16.947793 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" exitCode=0 Feb 25 08:43:16 crc kubenswrapper[4978]: I0225 08:43:16.947901 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b"} Feb 25 08:43:16 crc kubenswrapper[4978]: I0225 08:43:16.948207 4978 scope.go:117] "RemoveContainer" containerID="914d9830f2d8d5aaf326a3efa925940acc7734a6c345c2cd844ec55b8a401b2b" Feb 25 08:43:16 crc kubenswrapper[4978]: I0225 08:43:16.949238 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:43:16 crc kubenswrapper[4978]: E0225 08:43:16.950016 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:43:28 crc kubenswrapper[4978]: I0225 08:43:28.391156 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:43:30 crc kubenswrapper[4978]: I0225 08:43:30.874790 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7cc578599c-fmh6q" Feb 25 08:43:30 crc kubenswrapper[4978]: I0225 08:43:30.949317 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8546f94b5d-dnjws"] Feb 25 08:43:30 crc kubenswrapper[4978]: I0225 08:43:30.949599 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8546f94b5d-dnjws" podUID="11829354-abeb-4753-a75a-47f9100fb784" containerName="neutron-api" containerID="cri-o://563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2" gracePeriod=30 Feb 25 08:43:30 crc kubenswrapper[4978]: I0225 08:43:30.949782 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8546f94b5d-dnjws" podUID="11829354-abeb-4753-a75a-47f9100fb784" containerName="neutron-httpd" containerID="cri-o://6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce" gracePeriod=30 Feb 25 08:43:31 crc kubenswrapper[4978]: I0225 08:43:31.116207 4978 generic.go:334] "Generic (PLEG): container finished" podID="11829354-abeb-4753-a75a-47f9100fb784" containerID="6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce" exitCode=0 Feb 25 08:43:31 crc kubenswrapper[4978]: I0225 08:43:31.116254 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8546f94b5d-dnjws" event={"ID":"11829354-abeb-4753-a75a-47f9100fb784","Type":"ContainerDied","Data":"6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce"} Feb 25 08:43:32 crc kubenswrapper[4978]: I0225 08:43:32.327594 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:43:32 crc kubenswrapper[4978]: E0225 08:43:32.328167 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:43:33 crc kubenswrapper[4978]: I0225 08:43:33.879097 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:43:33 crc kubenswrapper[4978]: I0225 08:43:33.977043 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-combined-ca-bundle\") pod \"11829354-abeb-4753-a75a-47f9100fb784\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " Feb 25 08:43:33 crc kubenswrapper[4978]: I0225 08:43:33.977264 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-config\") pod \"11829354-abeb-4753-a75a-47f9100fb784\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " Feb 25 08:43:33 crc kubenswrapper[4978]: I0225 08:43:33.977358 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wglw8\" (UniqueName: \"kubernetes.io/projected/11829354-abeb-4753-a75a-47f9100fb784-kube-api-access-wglw8\") pod \"11829354-abeb-4753-a75a-47f9100fb784\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " Feb 25 08:43:33 crc kubenswrapper[4978]: I0225 08:43:33.977652 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-httpd-config\") pod \"11829354-abeb-4753-a75a-47f9100fb784\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " Feb 25 08:43:33 crc kubenswrapper[4978]: I0225 08:43:33.977739 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-ovndb-tls-certs\") pod \"11829354-abeb-4753-a75a-47f9100fb784\" (UID: \"11829354-abeb-4753-a75a-47f9100fb784\") " Feb 25 08:43:33 crc kubenswrapper[4978]: I0225 08:43:33.983907 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11829354-abeb-4753-a75a-47f9100fb784-kube-api-access-wglw8" (OuterVolumeSpecName: "kube-api-access-wglw8") pod "11829354-abeb-4753-a75a-47f9100fb784" (UID: "11829354-abeb-4753-a75a-47f9100fb784"). InnerVolumeSpecName "kube-api-access-wglw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:43:33 crc kubenswrapper[4978]: I0225 08:43:33.984949 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "11829354-abeb-4753-a75a-47f9100fb784" (UID: "11829354-abeb-4753-a75a-47f9100fb784"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.055820 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11829354-abeb-4753-a75a-47f9100fb784" (UID: "11829354-abeb-4753-a75a-47f9100fb784"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.058640 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-config" (OuterVolumeSpecName: "config") pod "11829354-abeb-4753-a75a-47f9100fb784" (UID: "11829354-abeb-4753-a75a-47f9100fb784"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.080349 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.080412 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.080429 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.080441 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wglw8\" (UniqueName: \"kubernetes.io/projected/11829354-abeb-4753-a75a-47f9100fb784-kube-api-access-wglw8\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.084828 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "11829354-abeb-4753-a75a-47f9100fb784" (UID: "11829354-abeb-4753-a75a-47f9100fb784"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.151506 4978 generic.go:334] "Generic (PLEG): container finished" podID="11829354-abeb-4753-a75a-47f9100fb784" containerID="563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2" exitCode=0 Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.151611 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8546f94b5d-dnjws" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.151639 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8546f94b5d-dnjws" event={"ID":"11829354-abeb-4753-a75a-47f9100fb784","Type":"ContainerDied","Data":"563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2"} Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.152775 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8546f94b5d-dnjws" event={"ID":"11829354-abeb-4753-a75a-47f9100fb784","Type":"ContainerDied","Data":"91e084eb2f6ce873ac0d490568fd79e491ebbf75c06aeee9babf1619786b336a"} Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.152806 4978 scope.go:117] "RemoveContainer" containerID="6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.182582 4978 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/11829354-abeb-4753-a75a-47f9100fb784-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.190202 4978 scope.go:117] "RemoveContainer" containerID="563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.206868 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8546f94b5d-dnjws"] Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.213220 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8546f94b5d-dnjws"] Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.228322 4978 scope.go:117] "RemoveContainer" containerID="6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce" Feb 25 08:43:34 crc kubenswrapper[4978]: E0225 08:43:34.229886 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce\": container with ID starting with 6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce not found: ID does not exist" containerID="6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.229936 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce"} err="failed to get container status \"6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce\": rpc error: code = NotFound desc = could not find container \"6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce\": container with ID starting with 6d8372f1ff2cef9d88a86734b4bf0910b3ddfdf2421a3981466ea60c9433c8ce not found: ID does not exist" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.229967 4978 scope.go:117] "RemoveContainer" containerID="563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2" Feb 25 08:43:34 crc kubenswrapper[4978]: E0225 08:43:34.230642 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2\": container with ID starting with 563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2 not found: ID does not exist" containerID="563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2" Feb 25 08:43:34 crc kubenswrapper[4978]: I0225 08:43:34.230671 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2"} err="failed to get container status \"563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2\": rpc error: code = NotFound desc = could not find container \"563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2\": container with ID starting with 563fe75ac8d7c9b229462809fd340badaf42166277592ea9b50ac411075756b2 not found: ID does not exist" Feb 25 08:43:35 crc kubenswrapper[4978]: I0225 08:43:35.361535 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11829354-abeb-4753-a75a-47f9100fb784" path="/var/lib/kubelet/pods/11829354-abeb-4753-a75a-47f9100fb784/volumes" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.893212 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-hv77m"] Feb 25 08:43:40 crc kubenswrapper[4978]: E0225 08:43:40.894044 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11829354-abeb-4753-a75a-47f9100fb784" containerName="neutron-httpd" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.894058 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="11829354-abeb-4753-a75a-47f9100fb784" containerName="neutron-httpd" Feb 25 08:43:40 crc kubenswrapper[4978]: E0225 08:43:40.894073 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c9fc22b-74e3-40f1-b51d-77b380048ab7" containerName="init" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.894080 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c9fc22b-74e3-40f1-b51d-77b380048ab7" containerName="init" Feb 25 08:43:40 crc kubenswrapper[4978]: E0225 08:43:40.894109 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c9fc22b-74e3-40f1-b51d-77b380048ab7" containerName="dnsmasq-dns" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.894120 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c9fc22b-74e3-40f1-b51d-77b380048ab7" containerName="dnsmasq-dns" Feb 25 08:43:40 crc kubenswrapper[4978]: E0225 08:43:40.894134 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11829354-abeb-4753-a75a-47f9100fb784" containerName="neutron-api" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.894141 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="11829354-abeb-4753-a75a-47f9100fb784" containerName="neutron-api" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.894286 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c9fc22b-74e3-40f1-b51d-77b380048ab7" containerName="dnsmasq-dns" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.894300 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="11829354-abeb-4753-a75a-47f9100fb784" containerName="neutron-api" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.894310 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="11829354-abeb-4753-a75a-47f9100fb784" containerName="neutron-httpd" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.894996 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.897344 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-8ccpt" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.897345 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.902480 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.902810 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.902852 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.919218 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9fw44"] Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.920481 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.935244 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-hv77m"] Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.965439 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9fw44"] Feb 25 08:43:40 crc kubenswrapper[4978]: I0225 08:43:40.977126 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-hv77m"] Feb 25 08:43:40 crc kubenswrapper[4978]: E0225 08:43:40.978697 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-f462d ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-hv77m" podUID="5b46fdab-0b93-4cf3-9cb0-78967d582c50" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.013180 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57b6bf9479-cm8f2"] Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.014634 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022203 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-scripts\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022265 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-ring-data-devices\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022301 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-ring-data-devices\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022319 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f462d\" (UniqueName: \"kubernetes.io/projected/5b46fdab-0b93-4cf3-9cb0-78967d582c50-kube-api-access-f462d\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022346 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-swiftconf\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022384 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-swiftconf\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022417 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-scripts\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022434 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69kx5\" (UniqueName: \"kubernetes.io/projected/f0976f41-a781-43a1-be78-efbb126c9910-kube-api-access-69kx5\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022455 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-dispersionconf\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022470 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-combined-ca-bundle\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022479 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57b6bf9479-cm8f2"] Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022510 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0976f41-a781-43a1-be78-efbb126c9910-etc-swift\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022529 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-combined-ca-bundle\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022545 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5b46fdab-0b93-4cf3-9cb0-78967d582c50-etc-swift\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.022671 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-dispersionconf\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124434 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-swiftconf\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124499 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-swiftconf\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124522 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-sb\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124582 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhdhx\" (UniqueName: \"kubernetes.io/projected/1b190635-d02f-41ac-bcce-282a73f66d40-kube-api-access-mhdhx\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124604 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-scripts\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124620 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69kx5\" (UniqueName: \"kubernetes.io/projected/f0976f41-a781-43a1-be78-efbb126c9910-kube-api-access-69kx5\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124640 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-dispersionconf\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124659 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-combined-ca-bundle\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124676 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-config\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124720 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0976f41-a781-43a1-be78-efbb126c9910-etc-swift\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124735 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-combined-ca-bundle\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124750 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5b46fdab-0b93-4cf3-9cb0-78967d582c50-etc-swift\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124773 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-dispersionconf\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124792 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-dns-svc\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124817 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-scripts\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124837 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-nb\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124874 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-ring-data-devices\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124903 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-ring-data-devices\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.124919 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f462d\" (UniqueName: \"kubernetes.io/projected/5b46fdab-0b93-4cf3-9cb0-78967d582c50-kube-api-access-f462d\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.127914 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0976f41-a781-43a1-be78-efbb126c9910-etc-swift\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.130690 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-ring-data-devices\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.130932 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5b46fdab-0b93-4cf3-9cb0-78967d582c50-etc-swift\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.133938 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-scripts\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.135097 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-ring-data-devices\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.135670 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-scripts\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.145196 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-dispersionconf\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.145254 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f462d\" (UniqueName: \"kubernetes.io/projected/5b46fdab-0b93-4cf3-9cb0-78967d582c50-kube-api-access-f462d\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.154044 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-combined-ca-bundle\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.154316 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-dispersionconf\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.154598 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-swiftconf\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.155730 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-combined-ca-bundle\") pod \"swift-ring-rebalance-hv77m\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.161796 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-swiftconf\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.162933 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69kx5\" (UniqueName: \"kubernetes.io/projected/f0976f41-a781-43a1-be78-efbb126c9910-kube-api-access-69kx5\") pod \"swift-ring-rebalance-9fw44\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.228232 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-config\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.228587 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-dns-svc\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.228630 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-nb\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.228705 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-sb\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.228733 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhdhx\" (UniqueName: \"kubernetes.io/projected/1b190635-d02f-41ac-bcce-282a73f66d40-kube-api-access-mhdhx\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.229755 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-config\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.230234 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-dns-svc\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.230728 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-nb\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.231205 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-sb\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.237088 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.247566 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.285566 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.290836 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhdhx\" (UniqueName: \"kubernetes.io/projected/1b190635-d02f-41ac-bcce-282a73f66d40-kube-api-access-mhdhx\") pod \"dnsmasq-dns-57b6bf9479-cm8f2\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.347940 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.434393 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-scripts\") pod \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.434472 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-ring-data-devices\") pod \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.434566 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-dispersionconf\") pod \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.434602 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-combined-ca-bundle\") pod \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.434623 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5b46fdab-0b93-4cf3-9cb0-78967d582c50-etc-swift\") pod \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.434660 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f462d\" (UniqueName: \"kubernetes.io/projected/5b46fdab-0b93-4cf3-9cb0-78967d582c50-kube-api-access-f462d\") pod \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.434722 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-swiftconf\") pod \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\" (UID: \"5b46fdab-0b93-4cf3-9cb0-78967d582c50\") " Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.436925 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b46fdab-0b93-4cf3-9cb0-78967d582c50-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5b46fdab-0b93-4cf3-9cb0-78967d582c50" (UID: "5b46fdab-0b93-4cf3-9cb0-78967d582c50"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.437882 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-scripts" (OuterVolumeSpecName: "scripts") pod "5b46fdab-0b93-4cf3-9cb0-78967d582c50" (UID: "5b46fdab-0b93-4cf3-9cb0-78967d582c50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.438245 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5b46fdab-0b93-4cf3-9cb0-78967d582c50" (UID: "5b46fdab-0b93-4cf3-9cb0-78967d582c50"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.439552 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5b46fdab-0b93-4cf3-9cb0-78967d582c50" (UID: "5b46fdab-0b93-4cf3-9cb0-78967d582c50"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.441308 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b46fdab-0b93-4cf3-9cb0-78967d582c50" (UID: "5b46fdab-0b93-4cf3-9cb0-78967d582c50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.441538 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5b46fdab-0b93-4cf3-9cb0-78967d582c50" (UID: "5b46fdab-0b93-4cf3-9cb0-78967d582c50"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.444176 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b46fdab-0b93-4cf3-9cb0-78967d582c50-kube-api-access-f462d" (OuterVolumeSpecName: "kube-api-access-f462d") pod "5b46fdab-0b93-4cf3-9cb0-78967d582c50" (UID: "5b46fdab-0b93-4cf3-9cb0-78967d582c50"). InnerVolumeSpecName "kube-api-access-f462d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.536672 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.536720 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5b46fdab-0b93-4cf3-9cb0-78967d582c50-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.536734 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f462d\" (UniqueName: \"kubernetes.io/projected/5b46fdab-0b93-4cf3-9cb0-78967d582c50-kube-api-access-f462d\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.536748 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.536760 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.536794 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5b46fdab-0b93-4cf3-9cb0-78967d582c50-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.536801 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5b46fdab-0b93-4cf3-9cb0-78967d582c50-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.763144 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9fw44"] Feb 25 08:43:41 crc kubenswrapper[4978]: I0225 08:43:41.845058 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57b6bf9479-cm8f2"] Feb 25 08:43:41 crc kubenswrapper[4978]: W0225 08:43:41.849468 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b190635_d02f_41ac_bcce_282a73f66d40.slice/crio-758cb580f0fea5016abaae5b06362a2fbfc852d8e644185dd573104af022e99b WatchSource:0}: Error finding container 758cb580f0fea5016abaae5b06362a2fbfc852d8e644185dd573104af022e99b: Status 404 returned error can't find the container with id 758cb580f0fea5016abaae5b06362a2fbfc852d8e644185dd573104af022e99b Feb 25 08:43:42 crc kubenswrapper[4978]: I0225 08:43:42.248872 4978 generic.go:334] "Generic (PLEG): container finished" podID="1b190635-d02f-41ac-bcce-282a73f66d40" containerID="76f0115e6d9038f7e9d58a18039109ea26258971f5e059062456304cdfa92e31" exitCode=0 Feb 25 08:43:42 crc kubenswrapper[4978]: I0225 08:43:42.249071 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" event={"ID":"1b190635-d02f-41ac-bcce-282a73f66d40","Type":"ContainerDied","Data":"76f0115e6d9038f7e9d58a18039109ea26258971f5e059062456304cdfa92e31"} Feb 25 08:43:42 crc kubenswrapper[4978]: I0225 08:43:42.249308 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" event={"ID":"1b190635-d02f-41ac-bcce-282a73f66d40","Type":"ContainerStarted","Data":"758cb580f0fea5016abaae5b06362a2fbfc852d8e644185dd573104af022e99b"} Feb 25 08:43:42 crc kubenswrapper[4978]: I0225 08:43:42.251969 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-hv77m" Feb 25 08:43:42 crc kubenswrapper[4978]: I0225 08:43:42.251981 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fw44" event={"ID":"f0976f41-a781-43a1-be78-efbb126c9910","Type":"ContainerStarted","Data":"ce8b1b9054aa62af067f203612789adfcccd1cb4a4152ce50c6e89a2ea5acffc"} Feb 25 08:43:42 crc kubenswrapper[4978]: I0225 08:43:42.438164 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-hv77m"] Feb 25 08:43:42 crc kubenswrapper[4978]: I0225 08:43:42.447427 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-hv77m"] Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.051232 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7b89cfcdf8-pgk9t"] Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.080726 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.085724 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.099679 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b89cfcdf8-pgk9t"] Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.181333 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgmgs\" (UniqueName: \"kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-kube-api-access-qgmgs\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.181399 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-combined-ca-bundle\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.181443 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-log-httpd\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.181520 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-etc-swift\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.181572 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-config-data\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.181609 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-run-httpd\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.262684 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" event={"ID":"1b190635-d02f-41ac-bcce-282a73f66d40","Type":"ContainerStarted","Data":"75bcb874ea68ae0d4be3642cfc3b4c24733abbfd3402a2fb29189fde4f2eca81"} Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.263594 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.285210 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-etc-swift\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.285294 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-config-data\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.285350 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-run-httpd\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.285577 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgmgs\" (UniqueName: \"kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-kube-api-access-qgmgs\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.285618 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-combined-ca-bundle\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.285677 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-log-httpd\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.285878 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" podStartSLOduration=3.285864719 podStartE2EDuration="3.285864719s" podCreationTimestamp="2026-02-25 08:43:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:43:43.280178243 +0000 UTC m=+7116.719434702" watchObservedRunningTime="2026-02-25 08:43:43.285864719 +0000 UTC m=+7116.725121178" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.286565 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-log-httpd\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.286986 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-run-httpd\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.292051 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-etc-swift\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.306215 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-config-data\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.306893 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgmgs\" (UniqueName: \"kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-kube-api-access-qgmgs\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.307484 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-combined-ca-bundle\") pod \"swift-proxy-7b89cfcdf8-pgk9t\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.331929 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:43:43 crc kubenswrapper[4978]: E0225 08:43:43.332107 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.343890 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b46fdab-0b93-4cf3-9cb0-78967d582c50" path="/var/lib/kubelet/pods/5b46fdab-0b93-4cf3-9cb0-78967d582c50/volumes" Feb 25 08:43:43 crc kubenswrapper[4978]: I0225 08:43:43.400515 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:44 crc kubenswrapper[4978]: I0225 08:43:44.014989 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b89cfcdf8-pgk9t"] Feb 25 08:43:45 crc kubenswrapper[4978]: W0225 08:43:45.415897 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode86aa8c2_41be_42b5_a79a_402834ca2f9b.slice/crio-89836922478ce5eb6f7e4b54545a9f0752410e830d326740c09c715a69b6eb8c WatchSource:0}: Error finding container 89836922478ce5eb6f7e4b54545a9f0752410e830d326740c09c715a69b6eb8c: Status 404 returned error can't find the container with id 89836922478ce5eb6f7e4b54545a9f0752410e830d326740c09c715a69b6eb8c Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.643001 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-66cf8f6986-prvmn"] Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.647552 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.651764 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.652498 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.676675 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-66cf8f6986-prvmn"] Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.732151 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-public-tls-certs\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.732298 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0d7bad3-d58a-4225-8390-e2785235d412-run-httpd\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.732433 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-combined-ca-bundle\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.732467 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-internal-tls-certs\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.732664 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpwkm\" (UniqueName: \"kubernetes.io/projected/a0d7bad3-d58a-4225-8390-e2785235d412-kube-api-access-tpwkm\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.732719 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0d7bad3-d58a-4225-8390-e2785235d412-log-httpd\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.732755 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0d7bad3-d58a-4225-8390-e2785235d412-etc-swift\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.732792 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-config-data\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.834241 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-public-tls-certs\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.834299 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0d7bad3-d58a-4225-8390-e2785235d412-run-httpd\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.834335 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-combined-ca-bundle\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.834357 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-internal-tls-certs\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.834454 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpwkm\" (UniqueName: \"kubernetes.io/projected/a0d7bad3-d58a-4225-8390-e2785235d412-kube-api-access-tpwkm\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.834493 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0d7bad3-d58a-4225-8390-e2785235d412-log-httpd\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.834520 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0d7bad3-d58a-4225-8390-e2785235d412-etc-swift\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.834540 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-config-data\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.838832 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0d7bad3-d58a-4225-8390-e2785235d412-run-httpd\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.839233 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0d7bad3-d58a-4225-8390-e2785235d412-log-httpd\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.839499 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-config-data\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.841972 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-public-tls-certs\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.842243 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a0d7bad3-d58a-4225-8390-e2785235d412-etc-swift\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.842903 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-combined-ca-bundle\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.845120 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0d7bad3-d58a-4225-8390-e2785235d412-internal-tls-certs\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:45 crc kubenswrapper[4978]: I0225 08:43:45.866409 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpwkm\" (UniqueName: \"kubernetes.io/projected/a0d7bad3-d58a-4225-8390-e2785235d412-kube-api-access-tpwkm\") pod \"swift-proxy-66cf8f6986-prvmn\" (UID: \"a0d7bad3-d58a-4225-8390-e2785235d412\") " pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:46 crc kubenswrapper[4978]: I0225 08:43:46.001661 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:46 crc kubenswrapper[4978]: I0225 08:43:46.287345 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" event={"ID":"e86aa8c2-41be-42b5-a79a-402834ca2f9b","Type":"ContainerStarted","Data":"e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff"} Feb 25 08:43:46 crc kubenswrapper[4978]: I0225 08:43:46.287689 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:46 crc kubenswrapper[4978]: I0225 08:43:46.287701 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:46 crc kubenswrapper[4978]: I0225 08:43:46.287709 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" event={"ID":"e86aa8c2-41be-42b5-a79a-402834ca2f9b","Type":"ContainerStarted","Data":"e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63"} Feb 25 08:43:46 crc kubenswrapper[4978]: I0225 08:43:46.287719 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" event={"ID":"e86aa8c2-41be-42b5-a79a-402834ca2f9b","Type":"ContainerStarted","Data":"89836922478ce5eb6f7e4b54545a9f0752410e830d326740c09c715a69b6eb8c"} Feb 25 08:43:46 crc kubenswrapper[4978]: I0225 08:43:46.288472 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fw44" event={"ID":"f0976f41-a781-43a1-be78-efbb126c9910","Type":"ContainerStarted","Data":"ba1fd6abbe8ff9ab0430c4fd0d93d1f9f6a597a2bc2ec7efeaaa32902a86ea89"} Feb 25 08:43:46 crc kubenswrapper[4978]: I0225 08:43:46.316765 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" podStartSLOduration=3.107371722 podStartE2EDuration="3.316744672s" podCreationTimestamp="2026-02-25 08:43:43 +0000 UTC" firstStartedPulling="2026-02-25 08:43:45.421127839 +0000 UTC m=+7118.860384338" lastFinishedPulling="2026-02-25 08:43:45.630500829 +0000 UTC m=+7119.069757288" observedRunningTime="2026-02-25 08:43:46.30248242 +0000 UTC m=+7119.741738889" watchObservedRunningTime="2026-02-25 08:43:46.316744672 +0000 UTC m=+7119.756001141" Feb 25 08:43:46 crc kubenswrapper[4978]: I0225 08:43:46.330697 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-9fw44" podStartSLOduration=2.635596161 podStartE2EDuration="6.330677274s" podCreationTimestamp="2026-02-25 08:43:40 +0000 UTC" firstStartedPulling="2026-02-25 08:43:41.768844713 +0000 UTC m=+7115.208101172" lastFinishedPulling="2026-02-25 08:43:45.463925826 +0000 UTC m=+7118.903182285" observedRunningTime="2026-02-25 08:43:46.321897592 +0000 UTC m=+7119.761154061" watchObservedRunningTime="2026-02-25 08:43:46.330677274 +0000 UTC m=+7119.769933733" Feb 25 08:43:46 crc kubenswrapper[4978]: W0225 08:43:46.655468 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0d7bad3_d58a_4225_8390_e2785235d412.slice/crio-c3ff6bcb35b78e3a0f06e7c539e416478fb23a1f8a23321c1f674b7687cfa22f WatchSource:0}: Error finding container c3ff6bcb35b78e3a0f06e7c539e416478fb23a1f8a23321c1f674b7687cfa22f: Status 404 returned error can't find the container with id c3ff6bcb35b78e3a0f06e7c539e416478fb23a1f8a23321c1f674b7687cfa22f Feb 25 08:43:46 crc kubenswrapper[4978]: I0225 08:43:46.661906 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-66cf8f6986-prvmn"] Feb 25 08:43:47 crc kubenswrapper[4978]: I0225 08:43:47.299456 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66cf8f6986-prvmn" event={"ID":"a0d7bad3-d58a-4225-8390-e2785235d412","Type":"ContainerStarted","Data":"c9031532d31fb1e9740ebc76c33fa8368cabd7d200416c5367b583fb5146e61a"} Feb 25 08:43:47 crc kubenswrapper[4978]: I0225 08:43:47.300000 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:47 crc kubenswrapper[4978]: I0225 08:43:47.300029 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:47 crc kubenswrapper[4978]: I0225 08:43:47.300042 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66cf8f6986-prvmn" event={"ID":"a0d7bad3-d58a-4225-8390-e2785235d412","Type":"ContainerStarted","Data":"682ff18b76072b10abdf649f64eb6a0e3a2d4f6974971723be716a1af4a61106"} Feb 25 08:43:47 crc kubenswrapper[4978]: I0225 08:43:47.300058 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-66cf8f6986-prvmn" event={"ID":"a0d7bad3-d58a-4225-8390-e2785235d412","Type":"ContainerStarted","Data":"c3ff6bcb35b78e3a0f06e7c539e416478fb23a1f8a23321c1f674b7687cfa22f"} Feb 25 08:43:47 crc kubenswrapper[4978]: I0225 08:43:47.328898 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-66cf8f6986-prvmn" podStartSLOduration=2.328882347 podStartE2EDuration="2.328882347s" podCreationTimestamp="2026-02-25 08:43:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:43:47.325958457 +0000 UTC m=+7120.765214986" watchObservedRunningTime="2026-02-25 08:43:47.328882347 +0000 UTC m=+7120.768138806" Feb 25 08:43:50 crc kubenswrapper[4978]: I0225 08:43:50.328197 4978 generic.go:334] "Generic (PLEG): container finished" podID="f0976f41-a781-43a1-be78-efbb126c9910" containerID="ba1fd6abbe8ff9ab0430c4fd0d93d1f9f6a597a2bc2ec7efeaaa32902a86ea89" exitCode=0 Feb 25 08:43:50 crc kubenswrapper[4978]: I0225 08:43:50.328239 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fw44" event={"ID":"f0976f41-a781-43a1-be78-efbb126c9910","Type":"ContainerDied","Data":"ba1fd6abbe8ff9ab0430c4fd0d93d1f9f6a597a2bc2ec7efeaaa32902a86ea89"} Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.015640 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.350647 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.442851 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-88f9d946c-zq7ns"] Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.443222 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" podUID="ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" containerName="dnsmasq-dns" containerID="cri-o://4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd" gracePeriod=10 Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.819986 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.956913 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-dispersionconf\") pod \"f0976f41-a781-43a1-be78-efbb126c9910\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.956956 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-combined-ca-bundle\") pod \"f0976f41-a781-43a1-be78-efbb126c9910\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.956978 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-swiftconf\") pod \"f0976f41-a781-43a1-be78-efbb126c9910\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.956996 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-ring-data-devices\") pod \"f0976f41-a781-43a1-be78-efbb126c9910\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.957020 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69kx5\" (UniqueName: \"kubernetes.io/projected/f0976f41-a781-43a1-be78-efbb126c9910-kube-api-access-69kx5\") pod \"f0976f41-a781-43a1-be78-efbb126c9910\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.957091 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0976f41-a781-43a1-be78-efbb126c9910-etc-swift\") pod \"f0976f41-a781-43a1-be78-efbb126c9910\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.957198 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-scripts\") pod \"f0976f41-a781-43a1-be78-efbb126c9910\" (UID: \"f0976f41-a781-43a1-be78-efbb126c9910\") " Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.957628 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f0976f41-a781-43a1-be78-efbb126c9910" (UID: "f0976f41-a781-43a1-be78-efbb126c9910"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.958411 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0976f41-a781-43a1-be78-efbb126c9910-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f0976f41-a781-43a1-be78-efbb126c9910" (UID: "f0976f41-a781-43a1-be78-efbb126c9910"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.962387 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0976f41-a781-43a1-be78-efbb126c9910-kube-api-access-69kx5" (OuterVolumeSpecName: "kube-api-access-69kx5") pod "f0976f41-a781-43a1-be78-efbb126c9910" (UID: "f0976f41-a781-43a1-be78-efbb126c9910"). InnerVolumeSpecName "kube-api-access-69kx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.967553 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f0976f41-a781-43a1-be78-efbb126c9910" (UID: "f0976f41-a781-43a1-be78-efbb126c9910"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.970838 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.981822 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-scripts" (OuterVolumeSpecName: "scripts") pod "f0976f41-a781-43a1-be78-efbb126c9910" (UID: "f0976f41-a781-43a1-be78-efbb126c9910"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.982728 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f0976f41-a781-43a1-be78-efbb126c9910" (UID: "f0976f41-a781-43a1-be78-efbb126c9910"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:51 crc kubenswrapper[4978]: I0225 08:43:51.999675 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0976f41-a781-43a1-be78-efbb126c9910" (UID: "f0976f41-a781-43a1-be78-efbb126c9910"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.058511 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-config\") pod \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.058563 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-dns-svc\") pod \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.058618 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c99zr\" (UniqueName: \"kubernetes.io/projected/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-kube-api-access-c99zr\") pod \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.058637 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-nb\") pod \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.058719 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-sb\") pod \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\" (UID: \"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7\") " Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.059107 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.059121 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.059131 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.059141 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f0976f41-a781-43a1-be78-efbb126c9910-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.059149 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f0976f41-a781-43a1-be78-efbb126c9910-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.059158 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69kx5\" (UniqueName: \"kubernetes.io/projected/f0976f41-a781-43a1-be78-efbb126c9910-kube-api-access-69kx5\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.059167 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f0976f41-a781-43a1-be78-efbb126c9910-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.061483 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-kube-api-access-c99zr" (OuterVolumeSpecName: "kube-api-access-c99zr") pod "ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" (UID: "ef1d6157-4d4e-455e-9e48-d3ae5fe753c7"). InnerVolumeSpecName "kube-api-access-c99zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.102339 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" (UID: "ef1d6157-4d4e-455e-9e48-d3ae5fe753c7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.103758 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" (UID: "ef1d6157-4d4e-455e-9e48-d3ae5fe753c7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.109001 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" (UID: "ef1d6157-4d4e-455e-9e48-d3ae5fe753c7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.127626 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-config" (OuterVolumeSpecName: "config") pod "ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" (UID: "ef1d6157-4d4e-455e-9e48-d3ae5fe753c7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.160520 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c99zr\" (UniqueName: \"kubernetes.io/projected/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-kube-api-access-c99zr\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.160556 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.160568 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.160582 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.160596 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.348814 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fw44" event={"ID":"f0976f41-a781-43a1-be78-efbb126c9910","Type":"ContainerDied","Data":"ce8b1b9054aa62af067f203612789adfcccd1cb4a4152ce50c6e89a2ea5acffc"} Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.348855 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce8b1b9054aa62af067f203612789adfcccd1cb4a4152ce50c6e89a2ea5acffc" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.349078 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fw44" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.350569 4978 generic.go:334] "Generic (PLEG): container finished" podID="ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" containerID="4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd" exitCode=0 Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.350609 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" event={"ID":"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7","Type":"ContainerDied","Data":"4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd"} Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.350639 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" event={"ID":"ef1d6157-4d4e-455e-9e48-d3ae5fe753c7","Type":"ContainerDied","Data":"010482d1aa6e7dd1b33fe88d9c4301d653bf5006fd65a06cff9488eb24d52cc5"} Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.350676 4978 scope.go:117] "RemoveContainer" containerID="4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.350702 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-88f9d946c-zq7ns" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.377735 4978 scope.go:117] "RemoveContainer" containerID="8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.416752 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-88f9d946c-zq7ns"] Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.425204 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-88f9d946c-zq7ns"] Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.426848 4978 scope.go:117] "RemoveContainer" containerID="4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd" Feb 25 08:43:52 crc kubenswrapper[4978]: E0225 08:43:52.427439 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd\": container with ID starting with 4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd not found: ID does not exist" containerID="4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.427492 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd"} err="failed to get container status \"4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd\": rpc error: code = NotFound desc = could not find container \"4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd\": container with ID starting with 4734494fa3479bc0e688e926c7970f1d43a4f589d281d319f567d82e18cb5cbd not found: ID does not exist" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.427531 4978 scope.go:117] "RemoveContainer" containerID="8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5" Feb 25 08:43:52 crc kubenswrapper[4978]: E0225 08:43:52.428008 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5\": container with ID starting with 8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5 not found: ID does not exist" containerID="8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5" Feb 25 08:43:52 crc kubenswrapper[4978]: I0225 08:43:52.428132 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5"} err="failed to get container status \"8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5\": rpc error: code = NotFound desc = could not find container \"8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5\": container with ID starting with 8c9040ffaee68f8c7a06da48db4f14cad1b45468b7a3e6d5c5f5856f7de89ff5 not found: ID does not exist" Feb 25 08:43:53 crc kubenswrapper[4978]: I0225 08:43:53.349057 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" path="/var/lib/kubelet/pods/ef1d6157-4d4e-455e-9e48-d3ae5fe753c7/volumes" Feb 25 08:43:53 crc kubenswrapper[4978]: I0225 08:43:53.404027 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:53 crc kubenswrapper[4978]: I0225 08:43:53.407572 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:55 crc kubenswrapper[4978]: I0225 08:43:55.327949 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:43:55 crc kubenswrapper[4978]: E0225 08:43:55.328898 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:43:56 crc kubenswrapper[4978]: I0225 08:43:56.013337 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-66cf8f6986-prvmn" Feb 25 08:43:56 crc kubenswrapper[4978]: I0225 08:43:56.124595 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7b89cfcdf8-pgk9t"] Feb 25 08:43:56 crc kubenswrapper[4978]: I0225 08:43:56.124937 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" podUID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" containerName="proxy-httpd" containerID="cri-o://e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63" gracePeriod=30 Feb 25 08:43:56 crc kubenswrapper[4978]: I0225 08:43:56.125564 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" podUID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" containerName="proxy-server" containerID="cri-o://e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff" gracePeriod=30 Feb 25 08:43:56 crc kubenswrapper[4978]: I0225 08:43:56.387726 4978 generic.go:334] "Generic (PLEG): container finished" podID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" containerID="e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63" exitCode=0 Feb 25 08:43:56 crc kubenswrapper[4978]: I0225 08:43:56.387766 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" event={"ID":"e86aa8c2-41be-42b5-a79a-402834ca2f9b","Type":"ContainerDied","Data":"e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63"} Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.106693 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.255656 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-run-httpd\") pod \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.255727 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-config-data\") pod \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.255817 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-etc-swift\") pod \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.255843 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-combined-ca-bundle\") pod \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.255915 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgmgs\" (UniqueName: \"kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-kube-api-access-qgmgs\") pod \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.255936 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-log-httpd\") pod \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\" (UID: \"e86aa8c2-41be-42b5-a79a-402834ca2f9b\") " Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.256167 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e86aa8c2-41be-42b5-a79a-402834ca2f9b" (UID: "e86aa8c2-41be-42b5-a79a-402834ca2f9b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.256392 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.256413 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e86aa8c2-41be-42b5-a79a-402834ca2f9b" (UID: "e86aa8c2-41be-42b5-a79a-402834ca2f9b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.260725 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-kube-api-access-qgmgs" (OuterVolumeSpecName: "kube-api-access-qgmgs") pod "e86aa8c2-41be-42b5-a79a-402834ca2f9b" (UID: "e86aa8c2-41be-42b5-a79a-402834ca2f9b"). InnerVolumeSpecName "kube-api-access-qgmgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.276673 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e86aa8c2-41be-42b5-a79a-402834ca2f9b" (UID: "e86aa8c2-41be-42b5-a79a-402834ca2f9b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.309948 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-config-data" (OuterVolumeSpecName: "config-data") pod "e86aa8c2-41be-42b5-a79a-402834ca2f9b" (UID: "e86aa8c2-41be-42b5-a79a-402834ca2f9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.317352 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e86aa8c2-41be-42b5-a79a-402834ca2f9b" (UID: "e86aa8c2-41be-42b5-a79a-402834ca2f9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.358959 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.358996 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.359008 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e86aa8c2-41be-42b5-a79a-402834ca2f9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.359024 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgmgs\" (UniqueName: \"kubernetes.io/projected/e86aa8c2-41be-42b5-a79a-402834ca2f9b-kube-api-access-qgmgs\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.359035 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e86aa8c2-41be-42b5-a79a-402834ca2f9b-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.398836 4978 generic.go:334] "Generic (PLEG): container finished" podID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" containerID="e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff" exitCode=0 Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.398881 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.398902 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" event={"ID":"e86aa8c2-41be-42b5-a79a-402834ca2f9b","Type":"ContainerDied","Data":"e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff"} Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.399819 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b89cfcdf8-pgk9t" event={"ID":"e86aa8c2-41be-42b5-a79a-402834ca2f9b","Type":"ContainerDied","Data":"89836922478ce5eb6f7e4b54545a9f0752410e830d326740c09c715a69b6eb8c"} Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.399916 4978 scope.go:117] "RemoveContainer" containerID="e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.426920 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7b89cfcdf8-pgk9t"] Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.429281 4978 scope.go:117] "RemoveContainer" containerID="e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.434979 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-7b89cfcdf8-pgk9t"] Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.455555 4978 scope.go:117] "RemoveContainer" containerID="e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff" Feb 25 08:43:57 crc kubenswrapper[4978]: E0225 08:43:57.456258 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff\": container with ID starting with e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff not found: ID does not exist" containerID="e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.456325 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff"} err="failed to get container status \"e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff\": rpc error: code = NotFound desc = could not find container \"e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff\": container with ID starting with e25d702dd82f8cb59e69bcb627ec4cf33d1bc94f6181c870c255728ff86a15ff not found: ID does not exist" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.456417 4978 scope.go:117] "RemoveContainer" containerID="e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63" Feb 25 08:43:57 crc kubenswrapper[4978]: E0225 08:43:57.457836 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63\": container with ID starting with e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63 not found: ID does not exist" containerID="e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63" Feb 25 08:43:57 crc kubenswrapper[4978]: I0225 08:43:57.457882 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63"} err="failed to get container status \"e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63\": rpc error: code = NotFound desc = could not find container \"e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63\": container with ID starting with e08855a5a4d2eca06375d1c2d4023e1d860b44c0a655646981336c7a42f24d63 not found: ID does not exist" Feb 25 08:43:59 crc kubenswrapper[4978]: I0225 08:43:59.344569 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" path="/var/lib/kubelet/pods/e86aa8c2-41be-42b5-a79a-402834ca2f9b/volumes" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.130989 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533484-55fdb"] Feb 25 08:44:00 crc kubenswrapper[4978]: E0225 08:44:00.131315 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" containerName="init" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.131332 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" containerName="init" Feb 25 08:44:00 crc kubenswrapper[4978]: E0225 08:44:00.131345 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0976f41-a781-43a1-be78-efbb126c9910" containerName="swift-ring-rebalance" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.131351 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0976f41-a781-43a1-be78-efbb126c9910" containerName="swift-ring-rebalance" Feb 25 08:44:00 crc kubenswrapper[4978]: E0225 08:44:00.131745 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" containerName="dnsmasq-dns" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.131758 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" containerName="dnsmasq-dns" Feb 25 08:44:00 crc kubenswrapper[4978]: E0225 08:44:00.131772 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" containerName="proxy-httpd" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.131778 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" containerName="proxy-httpd" Feb 25 08:44:00 crc kubenswrapper[4978]: E0225 08:44:00.131786 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" containerName="proxy-server" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.131791 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" containerName="proxy-server" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.132060 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0976f41-a781-43a1-be78-efbb126c9910" containerName="swift-ring-rebalance" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.132076 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" containerName="proxy-httpd" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.132093 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e86aa8c2-41be-42b5-a79a-402834ca2f9b" containerName="proxy-server" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.132100 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef1d6157-4d4e-455e-9e48-d3ae5fe753c7" containerName="dnsmasq-dns" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.132671 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533484-55fdb" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.134871 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.135680 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.135815 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.152110 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533484-55fdb"] Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.212977 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm5j6\" (UniqueName: \"kubernetes.io/projected/763ee025-3136-4ba3-8d91-53d980116649-kube-api-access-tm5j6\") pod \"auto-csr-approver-29533484-55fdb\" (UID: \"763ee025-3136-4ba3-8d91-53d980116649\") " pod="openshift-infra/auto-csr-approver-29533484-55fdb" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.314402 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm5j6\" (UniqueName: \"kubernetes.io/projected/763ee025-3136-4ba3-8d91-53d980116649-kube-api-access-tm5j6\") pod \"auto-csr-approver-29533484-55fdb\" (UID: \"763ee025-3136-4ba3-8d91-53d980116649\") " pod="openshift-infra/auto-csr-approver-29533484-55fdb" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.345799 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm5j6\" (UniqueName: \"kubernetes.io/projected/763ee025-3136-4ba3-8d91-53d980116649-kube-api-access-tm5j6\") pod \"auto-csr-approver-29533484-55fdb\" (UID: \"763ee025-3136-4ba3-8d91-53d980116649\") " pod="openshift-infra/auto-csr-approver-29533484-55fdb" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.452798 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533484-55fdb" Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.890740 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533484-55fdb"] Feb 25 08:44:00 crc kubenswrapper[4978]: I0225 08:44:00.910730 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 08:44:00 crc kubenswrapper[4978]: E0225 08:44:00.973784 4978 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.146:55500->38.102.83.146:35177: read tcp 38.102.83.146:55500->38.102.83.146:35177: read: connection reset by peer Feb 25 08:44:01 crc kubenswrapper[4978]: I0225 08:44:01.442499 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533484-55fdb" event={"ID":"763ee025-3136-4ba3-8d91-53d980116649","Type":"ContainerStarted","Data":"2f5a01039e4c4ae5dada000ed6dc8a054e24b083334855686e423ce3bbe0a616"} Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.184703 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-2cvxh"] Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.185990 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2cvxh" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.200554 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-2cvxh"] Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.248859 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8280b490-fc6f-4ff5-a758-cda686b2483d-operator-scripts\") pod \"cinder-db-create-2cvxh\" (UID: \"8280b490-fc6f-4ff5-a758-cda686b2483d\") " pod="openstack/cinder-db-create-2cvxh" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.248930 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjchv\" (UniqueName: \"kubernetes.io/projected/8280b490-fc6f-4ff5-a758-cda686b2483d-kube-api-access-hjchv\") pod \"cinder-db-create-2cvxh\" (UID: \"8280b490-fc6f-4ff5-a758-cda686b2483d\") " pod="openstack/cinder-db-create-2cvxh" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.288009 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-da55-account-create-update-zksjb"] Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.289130 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-da55-account-create-update-zksjb" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.291602 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.296838 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-da55-account-create-update-zksjb"] Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.350265 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8280b490-fc6f-4ff5-a758-cda686b2483d-operator-scripts\") pod \"cinder-db-create-2cvxh\" (UID: \"8280b490-fc6f-4ff5-a758-cda686b2483d\") " pod="openstack/cinder-db-create-2cvxh" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.350316 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjchv\" (UniqueName: \"kubernetes.io/projected/8280b490-fc6f-4ff5-a758-cda686b2483d-kube-api-access-hjchv\") pod \"cinder-db-create-2cvxh\" (UID: \"8280b490-fc6f-4ff5-a758-cda686b2483d\") " pod="openstack/cinder-db-create-2cvxh" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.350348 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7c59b20-c701-4e2c-8383-03f6d93b5551-operator-scripts\") pod \"cinder-da55-account-create-update-zksjb\" (UID: \"f7c59b20-c701-4e2c-8383-03f6d93b5551\") " pod="openstack/cinder-da55-account-create-update-zksjb" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.350399 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glk4x\" (UniqueName: \"kubernetes.io/projected/f7c59b20-c701-4e2c-8383-03f6d93b5551-kube-api-access-glk4x\") pod \"cinder-da55-account-create-update-zksjb\" (UID: \"f7c59b20-c701-4e2c-8383-03f6d93b5551\") " pod="openstack/cinder-da55-account-create-update-zksjb" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.351174 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8280b490-fc6f-4ff5-a758-cda686b2483d-operator-scripts\") pod \"cinder-db-create-2cvxh\" (UID: \"8280b490-fc6f-4ff5-a758-cda686b2483d\") " pod="openstack/cinder-db-create-2cvxh" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.371586 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjchv\" (UniqueName: \"kubernetes.io/projected/8280b490-fc6f-4ff5-a758-cda686b2483d-kube-api-access-hjchv\") pod \"cinder-db-create-2cvxh\" (UID: \"8280b490-fc6f-4ff5-a758-cda686b2483d\") " pod="openstack/cinder-db-create-2cvxh" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.451643 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7c59b20-c701-4e2c-8383-03f6d93b5551-operator-scripts\") pod \"cinder-da55-account-create-update-zksjb\" (UID: \"f7c59b20-c701-4e2c-8383-03f6d93b5551\") " pod="openstack/cinder-da55-account-create-update-zksjb" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.452009 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glk4x\" (UniqueName: \"kubernetes.io/projected/f7c59b20-c701-4e2c-8383-03f6d93b5551-kube-api-access-glk4x\") pod \"cinder-da55-account-create-update-zksjb\" (UID: \"f7c59b20-c701-4e2c-8383-03f6d93b5551\") " pod="openstack/cinder-da55-account-create-update-zksjb" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.453275 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7c59b20-c701-4e2c-8383-03f6d93b5551-operator-scripts\") pod \"cinder-da55-account-create-update-zksjb\" (UID: \"f7c59b20-c701-4e2c-8383-03f6d93b5551\") " pod="openstack/cinder-da55-account-create-update-zksjb" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.453548 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533484-55fdb" event={"ID":"763ee025-3136-4ba3-8d91-53d980116649","Type":"ContainerStarted","Data":"63d433593b576c533c54000588288ce057fe6ee2c09db10b46ac262f5e311ef2"} Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.471810 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glk4x\" (UniqueName: \"kubernetes.io/projected/f7c59b20-c701-4e2c-8383-03f6d93b5551-kube-api-access-glk4x\") pod \"cinder-da55-account-create-update-zksjb\" (UID: \"f7c59b20-c701-4e2c-8383-03f6d93b5551\") " pod="openstack/cinder-da55-account-create-update-zksjb" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.520969 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2cvxh" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.612255 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-da55-account-create-update-zksjb" Feb 25 08:44:02 crc kubenswrapper[4978]: I0225 08:44:02.952466 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-2cvxh"] Feb 25 08:44:02 crc kubenswrapper[4978]: W0225 08:44:02.954318 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8280b490_fc6f_4ff5_a758_cda686b2483d.slice/crio-4b00c0c944108b53319513d78d3988361c9e1c9ca0e0fb65bfc94acb004eb6d1 WatchSource:0}: Error finding container 4b00c0c944108b53319513d78d3988361c9e1c9ca0e0fb65bfc94acb004eb6d1: Status 404 returned error can't find the container with id 4b00c0c944108b53319513d78d3988361c9e1c9ca0e0fb65bfc94acb004eb6d1 Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.103483 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-da55-account-create-update-zksjb"] Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.465652 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-da55-account-create-update-zksjb" event={"ID":"f7c59b20-c701-4e2c-8383-03f6d93b5551","Type":"ContainerStarted","Data":"6c8d45c68e3b713679a3be8b8b70d0e574a9c5c04b9e8ccf0e07b2c867369d79"} Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.465989 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-da55-account-create-update-zksjb" event={"ID":"f7c59b20-c701-4e2c-8383-03f6d93b5551","Type":"ContainerStarted","Data":"12445f352db15611f139ace3a6bec09979d2e5fccf0c7b1c974af5feab2e4ce2"} Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.468167 4978 generic.go:334] "Generic (PLEG): container finished" podID="763ee025-3136-4ba3-8d91-53d980116649" containerID="63d433593b576c533c54000588288ce057fe6ee2c09db10b46ac262f5e311ef2" exitCode=0 Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.468263 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533484-55fdb" event={"ID":"763ee025-3136-4ba3-8d91-53d980116649","Type":"ContainerDied","Data":"63d433593b576c533c54000588288ce057fe6ee2c09db10b46ac262f5e311ef2"} Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.470750 4978 generic.go:334] "Generic (PLEG): container finished" podID="8280b490-fc6f-4ff5-a758-cda686b2483d" containerID="662c76f3f813c122197536b2b51356488bce382ec99be2af22cab57a70a7432a" exitCode=0 Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.470821 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2cvxh" event={"ID":"8280b490-fc6f-4ff5-a758-cda686b2483d","Type":"ContainerDied","Data":"662c76f3f813c122197536b2b51356488bce382ec99be2af22cab57a70a7432a"} Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.470855 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2cvxh" event={"ID":"8280b490-fc6f-4ff5-a758-cda686b2483d","Type":"ContainerStarted","Data":"4b00c0c944108b53319513d78d3988361c9e1c9ca0e0fb65bfc94acb004eb6d1"} Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.499038 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-da55-account-create-update-zksjb" podStartSLOduration=1.499013593 podStartE2EDuration="1.499013593s" podCreationTimestamp="2026-02-25 08:44:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:44:03.486499844 +0000 UTC m=+7136.925756313" watchObservedRunningTime="2026-02-25 08:44:03.499013593 +0000 UTC m=+7136.938270072" Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.798401 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533484-55fdb" Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.878517 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm5j6\" (UniqueName: \"kubernetes.io/projected/763ee025-3136-4ba3-8d91-53d980116649-kube-api-access-tm5j6\") pod \"763ee025-3136-4ba3-8d91-53d980116649\" (UID: \"763ee025-3136-4ba3-8d91-53d980116649\") " Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.883204 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/763ee025-3136-4ba3-8d91-53d980116649-kube-api-access-tm5j6" (OuterVolumeSpecName: "kube-api-access-tm5j6") pod "763ee025-3136-4ba3-8d91-53d980116649" (UID: "763ee025-3136-4ba3-8d91-53d980116649"). InnerVolumeSpecName "kube-api-access-tm5j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:44:03 crc kubenswrapper[4978]: I0225 08:44:03.981135 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm5j6\" (UniqueName: \"kubernetes.io/projected/763ee025-3136-4ba3-8d91-53d980116649-kube-api-access-tm5j6\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:04 crc kubenswrapper[4978]: I0225 08:44:04.493803 4978 generic.go:334] "Generic (PLEG): container finished" podID="f7c59b20-c701-4e2c-8383-03f6d93b5551" containerID="6c8d45c68e3b713679a3be8b8b70d0e574a9c5c04b9e8ccf0e07b2c867369d79" exitCode=0 Feb 25 08:44:04 crc kubenswrapper[4978]: I0225 08:44:04.494329 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-da55-account-create-update-zksjb" event={"ID":"f7c59b20-c701-4e2c-8383-03f6d93b5551","Type":"ContainerDied","Data":"6c8d45c68e3b713679a3be8b8b70d0e574a9c5c04b9e8ccf0e07b2c867369d79"} Feb 25 08:44:04 crc kubenswrapper[4978]: I0225 08:44:04.503038 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533484-55fdb" Feb 25 08:44:04 crc kubenswrapper[4978]: I0225 08:44:04.503064 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533484-55fdb" event={"ID":"763ee025-3136-4ba3-8d91-53d980116649","Type":"ContainerDied","Data":"2f5a01039e4c4ae5dada000ed6dc8a054e24b083334855686e423ce3bbe0a616"} Feb 25 08:44:04 crc kubenswrapper[4978]: I0225 08:44:04.503130 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f5a01039e4c4ae5dada000ed6dc8a054e24b083334855686e423ce3bbe0a616" Feb 25 08:44:04 crc kubenswrapper[4978]: I0225 08:44:04.891932 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533478-vvt88"] Feb 25 08:44:04 crc kubenswrapper[4978]: I0225 08:44:04.898561 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533478-vvt88"] Feb 25 08:44:04 crc kubenswrapper[4978]: I0225 08:44:04.952695 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2cvxh" Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.000929 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8280b490-fc6f-4ff5-a758-cda686b2483d-operator-scripts\") pod \"8280b490-fc6f-4ff5-a758-cda686b2483d\" (UID: \"8280b490-fc6f-4ff5-a758-cda686b2483d\") " Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.001118 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjchv\" (UniqueName: \"kubernetes.io/projected/8280b490-fc6f-4ff5-a758-cda686b2483d-kube-api-access-hjchv\") pod \"8280b490-fc6f-4ff5-a758-cda686b2483d\" (UID: \"8280b490-fc6f-4ff5-a758-cda686b2483d\") " Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.001801 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8280b490-fc6f-4ff5-a758-cda686b2483d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8280b490-fc6f-4ff5-a758-cda686b2483d" (UID: "8280b490-fc6f-4ff5-a758-cda686b2483d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.005401 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8280b490-fc6f-4ff5-a758-cda686b2483d-kube-api-access-hjchv" (OuterVolumeSpecName: "kube-api-access-hjchv") pod "8280b490-fc6f-4ff5-a758-cda686b2483d" (UID: "8280b490-fc6f-4ff5-a758-cda686b2483d"). InnerVolumeSpecName "kube-api-access-hjchv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.102857 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjchv\" (UniqueName: \"kubernetes.io/projected/8280b490-fc6f-4ff5-a758-cda686b2483d-kube-api-access-hjchv\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.103071 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8280b490-fc6f-4ff5-a758-cda686b2483d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.337510 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fbf1727-1311-44ae-a207-0e856f22c017" path="/var/lib/kubelet/pods/7fbf1727-1311-44ae-a207-0e856f22c017/volumes" Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.513916 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-2cvxh" event={"ID":"8280b490-fc6f-4ff5-a758-cda686b2483d","Type":"ContainerDied","Data":"4b00c0c944108b53319513d78d3988361c9e1c9ca0e0fb65bfc94acb004eb6d1"} Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.513968 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b00c0c944108b53319513d78d3988361c9e1c9ca0e0fb65bfc94acb004eb6d1" Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.513941 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-2cvxh" Feb 25 08:44:05 crc kubenswrapper[4978]: I0225 08:44:05.923398 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-da55-account-create-update-zksjb" Feb 25 08:44:06 crc kubenswrapper[4978]: I0225 08:44:06.023840 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7c59b20-c701-4e2c-8383-03f6d93b5551-operator-scripts\") pod \"f7c59b20-c701-4e2c-8383-03f6d93b5551\" (UID: \"f7c59b20-c701-4e2c-8383-03f6d93b5551\") " Feb 25 08:44:06 crc kubenswrapper[4978]: I0225 08:44:06.024048 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glk4x\" (UniqueName: \"kubernetes.io/projected/f7c59b20-c701-4e2c-8383-03f6d93b5551-kube-api-access-glk4x\") pod \"f7c59b20-c701-4e2c-8383-03f6d93b5551\" (UID: \"f7c59b20-c701-4e2c-8383-03f6d93b5551\") " Feb 25 08:44:06 crc kubenswrapper[4978]: I0225 08:44:06.025264 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7c59b20-c701-4e2c-8383-03f6d93b5551-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7c59b20-c701-4e2c-8383-03f6d93b5551" (UID: "f7c59b20-c701-4e2c-8383-03f6d93b5551"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:44:06 crc kubenswrapper[4978]: I0225 08:44:06.028963 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7c59b20-c701-4e2c-8383-03f6d93b5551-kube-api-access-glk4x" (OuterVolumeSpecName: "kube-api-access-glk4x") pod "f7c59b20-c701-4e2c-8383-03f6d93b5551" (UID: "f7c59b20-c701-4e2c-8383-03f6d93b5551"). InnerVolumeSpecName "kube-api-access-glk4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:44:06 crc kubenswrapper[4978]: I0225 08:44:06.127198 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7c59b20-c701-4e2c-8383-03f6d93b5551-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:06 crc kubenswrapper[4978]: I0225 08:44:06.127250 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glk4x\" (UniqueName: \"kubernetes.io/projected/f7c59b20-c701-4e2c-8383-03f6d93b5551-kube-api-access-glk4x\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:06 crc kubenswrapper[4978]: I0225 08:44:06.327316 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:44:06 crc kubenswrapper[4978]: E0225 08:44:06.327654 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:44:06 crc kubenswrapper[4978]: I0225 08:44:06.528817 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-da55-account-create-update-zksjb" event={"ID":"f7c59b20-c701-4e2c-8383-03f6d93b5551","Type":"ContainerDied","Data":"12445f352db15611f139ace3a6bec09979d2e5fccf0c7b1c974af5feab2e4ce2"} Feb 25 08:44:06 crc kubenswrapper[4978]: I0225 08:44:06.528890 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12445f352db15611f139ace3a6bec09979d2e5fccf0c7b1c974af5feab2e4ce2" Feb 25 08:44:06 crc kubenswrapper[4978]: I0225 08:44:06.528921 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-da55-account-create-update-zksjb" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.542523 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-426w4"] Feb 25 08:44:07 crc kubenswrapper[4978]: E0225 08:44:07.543659 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8280b490-fc6f-4ff5-a758-cda686b2483d" containerName="mariadb-database-create" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.543694 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8280b490-fc6f-4ff5-a758-cda686b2483d" containerName="mariadb-database-create" Feb 25 08:44:07 crc kubenswrapper[4978]: E0225 08:44:07.543745 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="763ee025-3136-4ba3-8d91-53d980116649" containerName="oc" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.543764 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="763ee025-3136-4ba3-8d91-53d980116649" containerName="oc" Feb 25 08:44:07 crc kubenswrapper[4978]: E0225 08:44:07.543789 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c59b20-c701-4e2c-8383-03f6d93b5551" containerName="mariadb-account-create-update" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.543806 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c59b20-c701-4e2c-8383-03f6d93b5551" containerName="mariadb-account-create-update" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.544179 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="763ee025-3136-4ba3-8d91-53d980116649" containerName="oc" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.544223 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8280b490-fc6f-4ff5-a758-cda686b2483d" containerName="mariadb-database-create" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.544245 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c59b20-c701-4e2c-8383-03f6d93b5551" containerName="mariadb-account-create-update" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.545484 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.551645 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kkjr8" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.552116 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.553002 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.568906 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-426w4"] Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.654862 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-combined-ca-bundle\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.654910 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-scripts\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.655309 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-config-data\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.655671 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-db-sync-config-data\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.655747 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgw89\" (UniqueName: \"kubernetes.io/projected/09d826f1-229c-4a05-bfac-68f2c3f66c70-kube-api-access-xgw89\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.655920 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/09d826f1-229c-4a05-bfac-68f2c3f66c70-etc-machine-id\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.758231 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-config-data\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.758332 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-db-sync-config-data\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.758358 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgw89\" (UniqueName: \"kubernetes.io/projected/09d826f1-229c-4a05-bfac-68f2c3f66c70-kube-api-access-xgw89\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.758446 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/09d826f1-229c-4a05-bfac-68f2c3f66c70-etc-machine-id\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.758501 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-combined-ca-bundle\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.758521 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-scripts\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.758686 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/09d826f1-229c-4a05-bfac-68f2c3f66c70-etc-machine-id\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.767459 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-db-sync-config-data\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.768304 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-scripts\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.770764 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-combined-ca-bundle\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.772336 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-config-data\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.780357 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgw89\" (UniqueName: \"kubernetes.io/projected/09d826f1-229c-4a05-bfac-68f2c3f66c70-kube-api-access-xgw89\") pod \"cinder-db-sync-426w4\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:07 crc kubenswrapper[4978]: I0225 08:44:07.871345 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:08 crc kubenswrapper[4978]: I0225 08:44:08.817978 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-426w4"] Feb 25 08:44:08 crc kubenswrapper[4978]: W0225 08:44:08.834345 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09d826f1_229c_4a05_bfac_68f2c3f66c70.slice/crio-d3ba4b0427771cd594f0eb4ec679cf41f2e2bc49d44a67369d489f19822f3513 WatchSource:0}: Error finding container d3ba4b0427771cd594f0eb4ec679cf41f2e2bc49d44a67369d489f19822f3513: Status 404 returned error can't find the container with id d3ba4b0427771cd594f0eb4ec679cf41f2e2bc49d44a67369d489f19822f3513 Feb 25 08:44:09 crc kubenswrapper[4978]: I0225 08:44:09.569170 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-426w4" event={"ID":"09d826f1-229c-4a05-bfac-68f2c3f66c70","Type":"ContainerStarted","Data":"d3ba4b0427771cd594f0eb4ec679cf41f2e2bc49d44a67369d489f19822f3513"} Feb 25 08:44:19 crc kubenswrapper[4978]: I0225 08:44:19.327201 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:44:19 crc kubenswrapper[4978]: E0225 08:44:19.328006 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:44:27 crc kubenswrapper[4978]: I0225 08:44:27.917359 4978 scope.go:117] "RemoveContainer" containerID="f409e122243cfd8dbfca58352923255a229ee32e8796f4f38d263375c901945e" Feb 25 08:44:28 crc kubenswrapper[4978]: I0225 08:44:28.224039 4978 scope.go:117] "RemoveContainer" containerID="16b71432cdae2d7aa2358d36a154f1cab03d2f076ca387cf97dcd7116d85126d" Feb 25 08:44:29 crc kubenswrapper[4978]: I0225 08:44:29.771837 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-426w4" event={"ID":"09d826f1-229c-4a05-bfac-68f2c3f66c70","Type":"ContainerStarted","Data":"aaccef42f90917448f09edc3c29b0e8dc12daba8eb38d20a87e3a4184d0b8445"} Feb 25 08:44:29 crc kubenswrapper[4978]: I0225 08:44:29.816579 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-426w4" podStartSLOduration=3.320364649 podStartE2EDuration="22.816547943s" podCreationTimestamp="2026-02-25 08:44:07 +0000 UTC" firstStartedPulling="2026-02-25 08:44:08.837283052 +0000 UTC m=+7142.276539511" lastFinishedPulling="2026-02-25 08:44:28.333466346 +0000 UTC m=+7161.772722805" observedRunningTime="2026-02-25 08:44:29.803329163 +0000 UTC m=+7163.242585692" watchObservedRunningTime="2026-02-25 08:44:29.816547943 +0000 UTC m=+7163.255804472" Feb 25 08:44:30 crc kubenswrapper[4978]: I0225 08:44:30.328530 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:44:30 crc kubenswrapper[4978]: E0225 08:44:30.329346 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:44:31 crc kubenswrapper[4978]: I0225 08:44:31.806039 4978 generic.go:334] "Generic (PLEG): container finished" podID="09d826f1-229c-4a05-bfac-68f2c3f66c70" containerID="aaccef42f90917448f09edc3c29b0e8dc12daba8eb38d20a87e3a4184d0b8445" exitCode=0 Feb 25 08:44:31 crc kubenswrapper[4978]: I0225 08:44:31.806147 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-426w4" event={"ID":"09d826f1-229c-4a05-bfac-68f2c3f66c70","Type":"ContainerDied","Data":"aaccef42f90917448f09edc3c29b0e8dc12daba8eb38d20a87e3a4184d0b8445"} Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.208333 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.274351 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-scripts\") pod \"09d826f1-229c-4a05-bfac-68f2c3f66c70\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.275574 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/09d826f1-229c-4a05-bfac-68f2c3f66c70-etc-machine-id\") pod \"09d826f1-229c-4a05-bfac-68f2c3f66c70\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.275631 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-combined-ca-bundle\") pod \"09d826f1-229c-4a05-bfac-68f2c3f66c70\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.275686 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09d826f1-229c-4a05-bfac-68f2c3f66c70-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "09d826f1-229c-4a05-bfac-68f2c3f66c70" (UID: "09d826f1-229c-4a05-bfac-68f2c3f66c70"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.275790 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgw89\" (UniqueName: \"kubernetes.io/projected/09d826f1-229c-4a05-bfac-68f2c3f66c70-kube-api-access-xgw89\") pod \"09d826f1-229c-4a05-bfac-68f2c3f66c70\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.275861 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-db-sync-config-data\") pod \"09d826f1-229c-4a05-bfac-68f2c3f66c70\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.275890 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-config-data\") pod \"09d826f1-229c-4a05-bfac-68f2c3f66c70\" (UID: \"09d826f1-229c-4a05-bfac-68f2c3f66c70\") " Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.277094 4978 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/09d826f1-229c-4a05-bfac-68f2c3f66c70-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.281774 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "09d826f1-229c-4a05-bfac-68f2c3f66c70" (UID: "09d826f1-229c-4a05-bfac-68f2c3f66c70"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.282068 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09d826f1-229c-4a05-bfac-68f2c3f66c70-kube-api-access-xgw89" (OuterVolumeSpecName: "kube-api-access-xgw89") pod "09d826f1-229c-4a05-bfac-68f2c3f66c70" (UID: "09d826f1-229c-4a05-bfac-68f2c3f66c70"). InnerVolumeSpecName "kube-api-access-xgw89". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.282661 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-scripts" (OuterVolumeSpecName: "scripts") pod "09d826f1-229c-4a05-bfac-68f2c3f66c70" (UID: "09d826f1-229c-4a05-bfac-68f2c3f66c70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.322065 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09d826f1-229c-4a05-bfac-68f2c3f66c70" (UID: "09d826f1-229c-4a05-bfac-68f2c3f66c70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.361445 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-config-data" (OuterVolumeSpecName: "config-data") pod "09d826f1-229c-4a05-bfac-68f2c3f66c70" (UID: "09d826f1-229c-4a05-bfac-68f2c3f66c70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.378521 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgw89\" (UniqueName: \"kubernetes.io/projected/09d826f1-229c-4a05-bfac-68f2c3f66c70-kube-api-access-xgw89\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.378565 4978 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.378579 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.378590 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.378602 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09d826f1-229c-4a05-bfac-68f2c3f66c70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.836950 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-426w4" event={"ID":"09d826f1-229c-4a05-bfac-68f2c3f66c70","Type":"ContainerDied","Data":"d3ba4b0427771cd594f0eb4ec679cf41f2e2bc49d44a67369d489f19822f3513"} Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.836990 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3ba4b0427771cd594f0eb4ec679cf41f2e2bc49d44a67369d489f19822f3513" Feb 25 08:44:33 crc kubenswrapper[4978]: I0225 08:44:33.837056 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-426w4" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.226423 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cbfc78d5f-7rc2t"] Feb 25 08:44:34 crc kubenswrapper[4978]: E0225 08:44:34.226951 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09d826f1-229c-4a05-bfac-68f2c3f66c70" containerName="cinder-db-sync" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.226962 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="09d826f1-229c-4a05-bfac-68f2c3f66c70" containerName="cinder-db-sync" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.227131 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="09d826f1-229c-4a05-bfac-68f2c3f66c70" containerName="cinder-db-sync" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.227987 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.246948 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cbfc78d5f-7rc2t"] Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.295386 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txlw2\" (UniqueName: \"kubernetes.io/projected/e04d46cc-c0c6-4157-8701-c3b18996ade3-kube-api-access-txlw2\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.295444 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-sb\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.295473 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-config\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.295622 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-dns-svc\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.295930 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-nb\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.338265 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.343301 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.346755 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kkjr8" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.346984 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.347130 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.350844 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.352481 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.397465 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-dns-svc\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.397705 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c7cc55eb-3049-4768-b112-077f07d7eb1a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.397778 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-scripts\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.397870 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data-custom\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.397967 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvzd8\" (UniqueName: \"kubernetes.io/projected/c7cc55eb-3049-4768-b112-077f07d7eb1a-kube-api-access-kvzd8\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.398037 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.398113 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.398201 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-nb\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.398305 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txlw2\" (UniqueName: \"kubernetes.io/projected/e04d46cc-c0c6-4157-8701-c3b18996ade3-kube-api-access-txlw2\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.398399 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7cc55eb-3049-4768-b112-077f07d7eb1a-logs\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.398460 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-sb\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.398529 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-config\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.398872 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-nb\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.399047 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-sb\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.399160 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-dns-svc\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.399554 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-config\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.415967 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txlw2\" (UniqueName: \"kubernetes.io/projected/e04d46cc-c0c6-4157-8701-c3b18996ade3-kube-api-access-txlw2\") pod \"dnsmasq-dns-5cbfc78d5f-7rc2t\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.499964 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvzd8\" (UniqueName: \"kubernetes.io/projected/c7cc55eb-3049-4768-b112-077f07d7eb1a-kube-api-access-kvzd8\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.500003 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.500036 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.500085 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7cc55eb-3049-4768-b112-077f07d7eb1a-logs\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.500131 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c7cc55eb-3049-4768-b112-077f07d7eb1a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.500150 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-scripts\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.500190 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data-custom\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.500330 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c7cc55eb-3049-4768-b112-077f07d7eb1a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.500861 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7cc55eb-3049-4768-b112-077f07d7eb1a-logs\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.505183 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.508770 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data-custom\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.514764 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-scripts\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.515388 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.518746 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvzd8\" (UniqueName: \"kubernetes.io/projected/c7cc55eb-3049-4768-b112-077f07d7eb1a-kube-api-access-kvzd8\") pod \"cinder-api-0\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " pod="openstack/cinder-api-0" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.559497 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:34 crc kubenswrapper[4978]: I0225 08:44:34.660254 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 08:44:35 crc kubenswrapper[4978]: I0225 08:44:35.009768 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:44:35 crc kubenswrapper[4978]: I0225 08:44:35.099926 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cbfc78d5f-7rc2t"] Feb 25 08:44:35 crc kubenswrapper[4978]: W0225 08:44:35.110267 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode04d46cc_c0c6_4157_8701_c3b18996ade3.slice/crio-005dccfcdd5e53f873a2320d2f82622388c8442b341d413fb2e5b7d62a077e9f WatchSource:0}: Error finding container 005dccfcdd5e53f873a2320d2f82622388c8442b341d413fb2e5b7d62a077e9f: Status 404 returned error can't find the container with id 005dccfcdd5e53f873a2320d2f82622388c8442b341d413fb2e5b7d62a077e9f Feb 25 08:44:35 crc kubenswrapper[4978]: I0225 08:44:35.858557 4978 generic.go:334] "Generic (PLEG): container finished" podID="e04d46cc-c0c6-4157-8701-c3b18996ade3" containerID="71d48a7c2e9a1e190bdf3bf1a6d5eecaa5bc940f175870eecb06001dfabe7eb1" exitCode=0 Feb 25 08:44:35 crc kubenswrapper[4978]: I0225 08:44:35.858829 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" event={"ID":"e04d46cc-c0c6-4157-8701-c3b18996ade3","Type":"ContainerDied","Data":"71d48a7c2e9a1e190bdf3bf1a6d5eecaa5bc940f175870eecb06001dfabe7eb1"} Feb 25 08:44:35 crc kubenswrapper[4978]: I0225 08:44:35.858960 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" event={"ID":"e04d46cc-c0c6-4157-8701-c3b18996ade3","Type":"ContainerStarted","Data":"005dccfcdd5e53f873a2320d2f82622388c8442b341d413fb2e5b7d62a077e9f"} Feb 25 08:44:35 crc kubenswrapper[4978]: I0225 08:44:35.886605 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c7cc55eb-3049-4768-b112-077f07d7eb1a","Type":"ContainerStarted","Data":"479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29"} Feb 25 08:44:35 crc kubenswrapper[4978]: I0225 08:44:35.886653 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c7cc55eb-3049-4768-b112-077f07d7eb1a","Type":"ContainerStarted","Data":"3888377817ecdc59e3c85851003d3119808b18a199ef764fac2c2968b7551b47"} Feb 25 08:44:36 crc kubenswrapper[4978]: I0225 08:44:36.643707 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:44:36 crc kubenswrapper[4978]: I0225 08:44:36.897872 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" event={"ID":"e04d46cc-c0c6-4157-8701-c3b18996ade3","Type":"ContainerStarted","Data":"48b47601de0639b6ff473276bc6c39289033f3d6d6411fc14103fe670738b562"} Feb 25 08:44:36 crc kubenswrapper[4978]: I0225 08:44:36.899057 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:36 crc kubenswrapper[4978]: I0225 08:44:36.902016 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c7cc55eb-3049-4768-b112-077f07d7eb1a","Type":"ContainerStarted","Data":"4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9"} Feb 25 08:44:36 crc kubenswrapper[4978]: I0225 08:44:36.902108 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 25 08:44:36 crc kubenswrapper[4978]: I0225 08:44:36.922210 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" podStartSLOduration=2.922186042 podStartE2EDuration="2.922186042s" podCreationTimestamp="2026-02-25 08:44:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:44:36.916466155 +0000 UTC m=+7170.355722634" watchObservedRunningTime="2026-02-25 08:44:36.922186042 +0000 UTC m=+7170.361442501" Feb 25 08:44:36 crc kubenswrapper[4978]: I0225 08:44:36.953919 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.9538979850000002 podStartE2EDuration="2.953897985s" podCreationTimestamp="2026-02-25 08:44:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:44:36.942736319 +0000 UTC m=+7170.381992788" watchObservedRunningTime="2026-02-25 08:44:36.953897985 +0000 UTC m=+7170.393154444" Feb 25 08:44:37 crc kubenswrapper[4978]: I0225 08:44:37.909810 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c7cc55eb-3049-4768-b112-077f07d7eb1a" containerName="cinder-api-log" containerID="cri-o://479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29" gracePeriod=30 Feb 25 08:44:37 crc kubenswrapper[4978]: I0225 08:44:37.909878 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="c7cc55eb-3049-4768-b112-077f07d7eb1a" containerName="cinder-api" containerID="cri-o://4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9" gracePeriod=30 Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.473756 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.531776 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvzd8\" (UniqueName: \"kubernetes.io/projected/c7cc55eb-3049-4768-b112-077f07d7eb1a-kube-api-access-kvzd8\") pod \"c7cc55eb-3049-4768-b112-077f07d7eb1a\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.531871 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c7cc55eb-3049-4768-b112-077f07d7eb1a-etc-machine-id\") pod \"c7cc55eb-3049-4768-b112-077f07d7eb1a\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.531917 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7cc55eb-3049-4768-b112-077f07d7eb1a-logs\") pod \"c7cc55eb-3049-4768-b112-077f07d7eb1a\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.531954 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-scripts\") pod \"c7cc55eb-3049-4768-b112-077f07d7eb1a\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.532066 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data-custom\") pod \"c7cc55eb-3049-4768-b112-077f07d7eb1a\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.532097 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data\") pod \"c7cc55eb-3049-4768-b112-077f07d7eb1a\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.532172 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-combined-ca-bundle\") pod \"c7cc55eb-3049-4768-b112-077f07d7eb1a\" (UID: \"c7cc55eb-3049-4768-b112-077f07d7eb1a\") " Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.532526 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c7cc55eb-3049-4768-b112-077f07d7eb1a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c7cc55eb-3049-4768-b112-077f07d7eb1a" (UID: "c7cc55eb-3049-4768-b112-077f07d7eb1a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.532595 4978 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c7cc55eb-3049-4768-b112-077f07d7eb1a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.532772 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7cc55eb-3049-4768-b112-077f07d7eb1a-logs" (OuterVolumeSpecName: "logs") pod "c7cc55eb-3049-4768-b112-077f07d7eb1a" (UID: "c7cc55eb-3049-4768-b112-077f07d7eb1a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.539071 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7cc55eb-3049-4768-b112-077f07d7eb1a-kube-api-access-kvzd8" (OuterVolumeSpecName: "kube-api-access-kvzd8") pod "c7cc55eb-3049-4768-b112-077f07d7eb1a" (UID: "c7cc55eb-3049-4768-b112-077f07d7eb1a"). InnerVolumeSpecName "kube-api-access-kvzd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.544685 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-scripts" (OuterVolumeSpecName: "scripts") pod "c7cc55eb-3049-4768-b112-077f07d7eb1a" (UID: "c7cc55eb-3049-4768-b112-077f07d7eb1a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.557527 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c7cc55eb-3049-4768-b112-077f07d7eb1a" (UID: "c7cc55eb-3049-4768-b112-077f07d7eb1a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.576697 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7cc55eb-3049-4768-b112-077f07d7eb1a" (UID: "c7cc55eb-3049-4768-b112-077f07d7eb1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.622962 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data" (OuterVolumeSpecName: "config-data") pod "c7cc55eb-3049-4768-b112-077f07d7eb1a" (UID: "c7cc55eb-3049-4768-b112-077f07d7eb1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.634127 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.634161 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.634171 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.634181 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvzd8\" (UniqueName: \"kubernetes.io/projected/c7cc55eb-3049-4768-b112-077f07d7eb1a-kube-api-access-kvzd8\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.634190 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c7cc55eb-3049-4768-b112-077f07d7eb1a-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.634198 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7cc55eb-3049-4768-b112-077f07d7eb1a-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.920216 4978 generic.go:334] "Generic (PLEG): container finished" podID="c7cc55eb-3049-4768-b112-077f07d7eb1a" containerID="4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9" exitCode=0 Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.920245 4978 generic.go:334] "Generic (PLEG): container finished" podID="c7cc55eb-3049-4768-b112-077f07d7eb1a" containerID="479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29" exitCode=143 Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.920257 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c7cc55eb-3049-4768-b112-077f07d7eb1a","Type":"ContainerDied","Data":"4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9"} Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.920300 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c7cc55eb-3049-4768-b112-077f07d7eb1a","Type":"ContainerDied","Data":"479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29"} Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.920310 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c7cc55eb-3049-4768-b112-077f07d7eb1a","Type":"ContainerDied","Data":"3888377817ecdc59e3c85851003d3119808b18a199ef764fac2c2968b7551b47"} Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.920325 4978 scope.go:117] "RemoveContainer" containerID="4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.920463 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.955348 4978 scope.go:117] "RemoveContainer" containerID="479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.964838 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.976005 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.982176 4978 scope.go:117] "RemoveContainer" containerID="4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9" Feb 25 08:44:38 crc kubenswrapper[4978]: E0225 08:44:38.984914 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9\": container with ID starting with 4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9 not found: ID does not exist" containerID="4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.984994 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9"} err="failed to get container status \"4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9\": rpc error: code = NotFound desc = could not find container \"4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9\": container with ID starting with 4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9 not found: ID does not exist" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.985036 4978 scope.go:117] "RemoveContainer" containerID="479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29" Feb 25 08:44:38 crc kubenswrapper[4978]: E0225 08:44:38.985523 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29\": container with ID starting with 479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29 not found: ID does not exist" containerID="479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.985558 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29"} err="failed to get container status \"479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29\": rpc error: code = NotFound desc = could not find container \"479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29\": container with ID starting with 479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29 not found: ID does not exist" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.985581 4978 scope.go:117] "RemoveContainer" containerID="4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.985828 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9"} err="failed to get container status \"4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9\": rpc error: code = NotFound desc = could not find container \"4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9\": container with ID starting with 4f8f9c9c4895f343f7e6eec051f0b7886198e964bcebae5ae659b76f0217b2c9 not found: ID does not exist" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.985857 4978 scope.go:117] "RemoveContainer" containerID="479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.986236 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29"} err="failed to get container status \"479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29\": rpc error: code = NotFound desc = could not find container \"479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29\": container with ID starting with 479f48205978af1f7aee92a258828e11fc1fd72e4773801231670d7857f93a29 not found: ID does not exist" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.989709 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:44:38 crc kubenswrapper[4978]: E0225 08:44:38.990036 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cc55eb-3049-4768-b112-077f07d7eb1a" containerName="cinder-api" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.990052 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cc55eb-3049-4768-b112-077f07d7eb1a" containerName="cinder-api" Feb 25 08:44:38 crc kubenswrapper[4978]: E0225 08:44:38.990080 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cc55eb-3049-4768-b112-077f07d7eb1a" containerName="cinder-api-log" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.990087 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cc55eb-3049-4768-b112-077f07d7eb1a" containerName="cinder-api-log" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.990233 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cc55eb-3049-4768-b112-077f07d7eb1a" containerName="cinder-api-log" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.990244 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cc55eb-3049-4768-b112-077f07d7eb1a" containerName="cinder-api" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.991098 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.993669 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.993785 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.993956 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.998284 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.998534 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 25 08:44:38 crc kubenswrapper[4978]: I0225 08:44:38.998706 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kkjr8" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.017677 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.042945 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.043008 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lgsg\" (UniqueName: \"kubernetes.io/projected/34469660-662f-451a-b4b5-b002f087312c-kube-api-access-5lgsg\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.043092 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data-custom\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.043123 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.043176 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34469660-662f-451a-b4b5-b002f087312c-logs\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.043233 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.043263 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.043290 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34469660-662f-451a-b4b5-b002f087312c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.043337 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-scripts\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.144554 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data-custom\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.144616 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.144669 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34469660-662f-451a-b4b5-b002f087312c-logs\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.144740 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.144776 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.144811 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34469660-662f-451a-b4b5-b002f087312c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.144837 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-scripts\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.144871 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.144913 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lgsg\" (UniqueName: \"kubernetes.io/projected/34469660-662f-451a-b4b5-b002f087312c-kube-api-access-5lgsg\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.145148 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34469660-662f-451a-b4b5-b002f087312c-logs\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.145208 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34469660-662f-451a-b4b5-b002f087312c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.151105 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.153266 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.157328 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data-custom\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.157762 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.157880 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-scripts\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.160141 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.160823 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lgsg\" (UniqueName: \"kubernetes.io/projected/34469660-662f-451a-b4b5-b002f087312c-kube-api-access-5lgsg\") pod \"cinder-api-0\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.316309 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.346820 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7cc55eb-3049-4768-b112-077f07d7eb1a" path="/var/lib/kubelet/pods/c7cc55eb-3049-4768-b112-077f07d7eb1a/volumes" Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.790086 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:44:39 crc kubenswrapper[4978]: W0225 08:44:39.793806 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34469660_662f_451a_b4b5_b002f087312c.slice/crio-f2dd6c2bcff587a18a465be692c4052f76ae15138ae9a98ee11e40aa4db8b3ca WatchSource:0}: Error finding container f2dd6c2bcff587a18a465be692c4052f76ae15138ae9a98ee11e40aa4db8b3ca: Status 404 returned error can't find the container with id f2dd6c2bcff587a18a465be692c4052f76ae15138ae9a98ee11e40aa4db8b3ca Feb 25 08:44:39 crc kubenswrapper[4978]: I0225 08:44:39.934541 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34469660-662f-451a-b4b5-b002f087312c","Type":"ContainerStarted","Data":"f2dd6c2bcff587a18a465be692c4052f76ae15138ae9a98ee11e40aa4db8b3ca"} Feb 25 08:44:40 crc kubenswrapper[4978]: I0225 08:44:40.981563 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34469660-662f-451a-b4b5-b002f087312c","Type":"ContainerStarted","Data":"26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408"} Feb 25 08:44:41 crc kubenswrapper[4978]: I0225 08:44:41.997320 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34469660-662f-451a-b4b5-b002f087312c","Type":"ContainerStarted","Data":"6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8"} Feb 25 08:44:41 crc kubenswrapper[4978]: I0225 08:44:41.997738 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 25 08:44:42 crc kubenswrapper[4978]: I0225 08:44:42.034144 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.034122024 podStartE2EDuration="4.034122024s" podCreationTimestamp="2026-02-25 08:44:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:44:42.028158959 +0000 UTC m=+7175.467415448" watchObservedRunningTime="2026-02-25 08:44:42.034122024 +0000 UTC m=+7175.473378493" Feb 25 08:44:42 crc kubenswrapper[4978]: I0225 08:44:42.328249 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:44:42 crc kubenswrapper[4978]: E0225 08:44:42.328731 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:44:44 crc kubenswrapper[4978]: I0225 08:44:44.561513 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:44:44 crc kubenswrapper[4978]: I0225 08:44:44.642494 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57b6bf9479-cm8f2"] Feb 25 08:44:44 crc kubenswrapper[4978]: I0225 08:44:44.642723 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" podUID="1b190635-d02f-41ac-bcce-282a73f66d40" containerName="dnsmasq-dns" containerID="cri-o://75bcb874ea68ae0d4be3642cfc3b4c24733abbfd3402a2fb29189fde4f2eca81" gracePeriod=10 Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.037868 4978 generic.go:334] "Generic (PLEG): container finished" podID="1b190635-d02f-41ac-bcce-282a73f66d40" containerID="75bcb874ea68ae0d4be3642cfc3b4c24733abbfd3402a2fb29189fde4f2eca81" exitCode=0 Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.038080 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" event={"ID":"1b190635-d02f-41ac-bcce-282a73f66d40","Type":"ContainerDied","Data":"75bcb874ea68ae0d4be3642cfc3b4c24733abbfd3402a2fb29189fde4f2eca81"} Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.038133 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" event={"ID":"1b190635-d02f-41ac-bcce-282a73f66d40","Type":"ContainerDied","Data":"758cb580f0fea5016abaae5b06362a2fbfc852d8e644185dd573104af022e99b"} Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.038149 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="758cb580f0fea5016abaae5b06362a2fbfc852d8e644185dd573104af022e99b" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.108062 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.260222 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-nb\") pod \"1b190635-d02f-41ac-bcce-282a73f66d40\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.260297 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-dns-svc\") pod \"1b190635-d02f-41ac-bcce-282a73f66d40\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.260561 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhdhx\" (UniqueName: \"kubernetes.io/projected/1b190635-d02f-41ac-bcce-282a73f66d40-kube-api-access-mhdhx\") pod \"1b190635-d02f-41ac-bcce-282a73f66d40\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.260690 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-config\") pod \"1b190635-d02f-41ac-bcce-282a73f66d40\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.260796 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-sb\") pod \"1b190635-d02f-41ac-bcce-282a73f66d40\" (UID: \"1b190635-d02f-41ac-bcce-282a73f66d40\") " Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.267790 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b190635-d02f-41ac-bcce-282a73f66d40-kube-api-access-mhdhx" (OuterVolumeSpecName: "kube-api-access-mhdhx") pod "1b190635-d02f-41ac-bcce-282a73f66d40" (UID: "1b190635-d02f-41ac-bcce-282a73f66d40"). InnerVolumeSpecName "kube-api-access-mhdhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.320278 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b190635-d02f-41ac-bcce-282a73f66d40" (UID: "1b190635-d02f-41ac-bcce-282a73f66d40"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.320606 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-config" (OuterVolumeSpecName: "config") pod "1b190635-d02f-41ac-bcce-282a73f66d40" (UID: "1b190635-d02f-41ac-bcce-282a73f66d40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.327754 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b190635-d02f-41ac-bcce-282a73f66d40" (UID: "1b190635-d02f-41ac-bcce-282a73f66d40"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.329121 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b190635-d02f-41ac-bcce-282a73f66d40" (UID: "1b190635-d02f-41ac-bcce-282a73f66d40"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.363315 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.363358 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.363389 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.363400 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b190635-d02f-41ac-bcce-282a73f66d40-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:45 crc kubenswrapper[4978]: I0225 08:44:45.363413 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhdhx\" (UniqueName: \"kubernetes.io/projected/1b190635-d02f-41ac-bcce-282a73f66d40-kube-api-access-mhdhx\") on node \"crc\" DevicePath \"\"" Feb 25 08:44:46 crc kubenswrapper[4978]: I0225 08:44:46.047715 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b6bf9479-cm8f2" Feb 25 08:44:46 crc kubenswrapper[4978]: I0225 08:44:46.076339 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57b6bf9479-cm8f2"] Feb 25 08:44:46 crc kubenswrapper[4978]: I0225 08:44:46.084322 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57b6bf9479-cm8f2"] Feb 25 08:44:47 crc kubenswrapper[4978]: I0225 08:44:47.346276 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b190635-d02f-41ac-bcce-282a73f66d40" path="/var/lib/kubelet/pods/1b190635-d02f-41ac-bcce-282a73f66d40/volumes" Feb 25 08:44:51 crc kubenswrapper[4978]: I0225 08:44:51.090486 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 25 08:44:55 crc kubenswrapper[4978]: I0225 08:44:55.329265 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:44:55 crc kubenswrapper[4978]: E0225 08:44:55.330797 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.175057 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc"] Feb 25 08:45:00 crc kubenswrapper[4978]: E0225 08:45:00.177142 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b190635-d02f-41ac-bcce-282a73f66d40" containerName="init" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.177240 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b190635-d02f-41ac-bcce-282a73f66d40" containerName="init" Feb 25 08:45:00 crc kubenswrapper[4978]: E0225 08:45:00.177347 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b190635-d02f-41ac-bcce-282a73f66d40" containerName="dnsmasq-dns" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.177478 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b190635-d02f-41ac-bcce-282a73f66d40" containerName="dnsmasq-dns" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.177803 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b190635-d02f-41ac-bcce-282a73f66d40" containerName="dnsmasq-dns" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.178703 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.181799 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.189083 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.208412 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc"] Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.275401 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5wsd\" (UniqueName: \"kubernetes.io/projected/7a15832c-a661-450a-911f-8d6294511806-kube-api-access-m5wsd\") pod \"collect-profiles-29533485-s78gc\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.275533 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a15832c-a661-450a-911f-8d6294511806-config-volume\") pod \"collect-profiles-29533485-s78gc\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.275567 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a15832c-a661-450a-911f-8d6294511806-secret-volume\") pod \"collect-profiles-29533485-s78gc\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.376783 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5wsd\" (UniqueName: \"kubernetes.io/projected/7a15832c-a661-450a-911f-8d6294511806-kube-api-access-m5wsd\") pod \"collect-profiles-29533485-s78gc\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.376887 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a15832c-a661-450a-911f-8d6294511806-config-volume\") pod \"collect-profiles-29533485-s78gc\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.376914 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a15832c-a661-450a-911f-8d6294511806-secret-volume\") pod \"collect-profiles-29533485-s78gc\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.377808 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a15832c-a661-450a-911f-8d6294511806-config-volume\") pod \"collect-profiles-29533485-s78gc\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.386511 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a15832c-a661-450a-911f-8d6294511806-secret-volume\") pod \"collect-profiles-29533485-s78gc\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.396654 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5wsd\" (UniqueName: \"kubernetes.io/projected/7a15832c-a661-450a-911f-8d6294511806-kube-api-access-m5wsd\") pod \"collect-profiles-29533485-s78gc\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.512407 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:00 crc kubenswrapper[4978]: I0225 08:45:00.779146 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc"] Feb 25 08:45:01 crc kubenswrapper[4978]: I0225 08:45:01.261172 4978 generic.go:334] "Generic (PLEG): container finished" podID="7a15832c-a661-450a-911f-8d6294511806" containerID="246de34d3d0c2174bd838d7222114a0ea1d253bd9c2e26f8f73b99f1000ad6af" exitCode=0 Feb 25 08:45:01 crc kubenswrapper[4978]: I0225 08:45:01.261280 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" event={"ID":"7a15832c-a661-450a-911f-8d6294511806","Type":"ContainerDied","Data":"246de34d3d0c2174bd838d7222114a0ea1d253bd9c2e26f8f73b99f1000ad6af"} Feb 25 08:45:01 crc kubenswrapper[4978]: I0225 08:45:01.261539 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" event={"ID":"7a15832c-a661-450a-911f-8d6294511806","Type":"ContainerStarted","Data":"ee7ce4a5b74b3e8b474cf18cd354b52c3e70dc512443754bca9526feaa0d10ad"} Feb 25 08:45:01 crc kubenswrapper[4978]: E0225 08:45:01.317106 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a15832c_a661_450a_911f_8d6294511806.slice/crio-conmon-246de34d3d0c2174bd838d7222114a0ea1d253bd9c2e26f8f73b99f1000ad6af.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a15832c_a661_450a_911f_8d6294511806.slice/crio-246de34d3d0c2174bd838d7222114a0ea1d253bd9c2e26f8f73b99f1000ad6af.scope\": RecentStats: unable to find data in memory cache]" Feb 25 08:45:02 crc kubenswrapper[4978]: I0225 08:45:02.678705 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:02 crc kubenswrapper[4978]: I0225 08:45:02.821771 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5wsd\" (UniqueName: \"kubernetes.io/projected/7a15832c-a661-450a-911f-8d6294511806-kube-api-access-m5wsd\") pod \"7a15832c-a661-450a-911f-8d6294511806\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " Feb 25 08:45:02 crc kubenswrapper[4978]: I0225 08:45:02.822195 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a15832c-a661-450a-911f-8d6294511806-secret-volume\") pod \"7a15832c-a661-450a-911f-8d6294511806\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " Feb 25 08:45:02 crc kubenswrapper[4978]: I0225 08:45:02.822317 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a15832c-a661-450a-911f-8d6294511806-config-volume\") pod \"7a15832c-a661-450a-911f-8d6294511806\" (UID: \"7a15832c-a661-450a-911f-8d6294511806\") " Feb 25 08:45:02 crc kubenswrapper[4978]: I0225 08:45:02.822813 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a15832c-a661-450a-911f-8d6294511806-config-volume" (OuterVolumeSpecName: "config-volume") pod "7a15832c-a661-450a-911f-8d6294511806" (UID: "7a15832c-a661-450a-911f-8d6294511806"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:45:02 crc kubenswrapper[4978]: I0225 08:45:02.829664 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a15832c-a661-450a-911f-8d6294511806-kube-api-access-m5wsd" (OuterVolumeSpecName: "kube-api-access-m5wsd") pod "7a15832c-a661-450a-911f-8d6294511806" (UID: "7a15832c-a661-450a-911f-8d6294511806"). InnerVolumeSpecName "kube-api-access-m5wsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:45:02 crc kubenswrapper[4978]: I0225 08:45:02.831350 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a15832c-a661-450a-911f-8d6294511806-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7a15832c-a661-450a-911f-8d6294511806" (UID: "7a15832c-a661-450a-911f-8d6294511806"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:02 crc kubenswrapper[4978]: I0225 08:45:02.924753 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7a15832c-a661-450a-911f-8d6294511806-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:02 crc kubenswrapper[4978]: I0225 08:45:02.924789 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7a15832c-a661-450a-911f-8d6294511806-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:02 crc kubenswrapper[4978]: I0225 08:45:02.924800 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5wsd\" (UniqueName: \"kubernetes.io/projected/7a15832c-a661-450a-911f-8d6294511806-kube-api-access-m5wsd\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:03 crc kubenswrapper[4978]: I0225 08:45:03.291010 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" event={"ID":"7a15832c-a661-450a-911f-8d6294511806","Type":"ContainerDied","Data":"ee7ce4a5b74b3e8b474cf18cd354b52c3e70dc512443754bca9526feaa0d10ad"} Feb 25 08:45:03 crc kubenswrapper[4978]: I0225 08:45:03.291053 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee7ce4a5b74b3e8b474cf18cd354b52c3e70dc512443754bca9526feaa0d10ad" Feb 25 08:45:03 crc kubenswrapper[4978]: I0225 08:45:03.291073 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc" Feb 25 08:45:03 crc kubenswrapper[4978]: I0225 08:45:03.760087 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn"] Feb 25 08:45:03 crc kubenswrapper[4978]: I0225 08:45:03.767577 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533440-hvsjn"] Feb 25 08:45:05 crc kubenswrapper[4978]: I0225 08:45:05.339086 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad8c7810-e508-400f-86de-6aae3c2b99c2" path="/var/lib/kubelet/pods/ad8c7810-e508-400f-86de-6aae3c2b99c2/volumes" Feb 25 08:45:06 crc kubenswrapper[4978]: I0225 08:45:06.330342 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:45:06 crc kubenswrapper[4978]: E0225 08:45:06.330843 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.723966 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 08:45:12 crc kubenswrapper[4978]: E0225 08:45:12.725072 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a15832c-a661-450a-911f-8d6294511806" containerName="collect-profiles" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.725095 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a15832c-a661-450a-911f-8d6294511806" containerName="collect-profiles" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.725499 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a15832c-a661-450a-911f-8d6294511806" containerName="collect-profiles" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.727036 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.731400 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.741413 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.824150 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.824230 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.824320 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.824462 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/123dbda5-f0ac-465e-bd80-cbf8ede8722b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.824531 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-scripts\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.824590 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6tt2\" (UniqueName: \"kubernetes.io/projected/123dbda5-f0ac-465e-bd80-cbf8ede8722b-kube-api-access-s6tt2\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.926384 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6tt2\" (UniqueName: \"kubernetes.io/projected/123dbda5-f0ac-465e-bd80-cbf8ede8722b-kube-api-access-s6tt2\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.926457 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.926495 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.926556 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.926606 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/123dbda5-f0ac-465e-bd80-cbf8ede8722b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.926662 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-scripts\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.928068 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/123dbda5-f0ac-465e-bd80-cbf8ede8722b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.933006 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.933175 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.934669 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.949432 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-scripts\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:12 crc kubenswrapper[4978]: I0225 08:45:12.954478 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6tt2\" (UniqueName: \"kubernetes.io/projected/123dbda5-f0ac-465e-bd80-cbf8ede8722b-kube-api-access-s6tt2\") pod \"cinder-scheduler-0\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:13 crc kubenswrapper[4978]: I0225 08:45:13.077191 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 08:45:13 crc kubenswrapper[4978]: I0225 08:45:13.634661 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 08:45:13 crc kubenswrapper[4978]: W0225 08:45:13.636973 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod123dbda5_f0ac_465e_bd80_cbf8ede8722b.slice/crio-7bb093dc4dec1d1193cc30fc0889f87c94f4750a9eb6556a10f3fe35d02db6b0 WatchSource:0}: Error finding container 7bb093dc4dec1d1193cc30fc0889f87c94f4750a9eb6556a10f3fe35d02db6b0: Status 404 returned error can't find the container with id 7bb093dc4dec1d1193cc30fc0889f87c94f4750a9eb6556a10f3fe35d02db6b0 Feb 25 08:45:14 crc kubenswrapper[4978]: I0225 08:45:14.369890 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:45:14 crc kubenswrapper[4978]: I0225 08:45:14.370461 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="34469660-662f-451a-b4b5-b002f087312c" containerName="cinder-api-log" containerID="cri-o://26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408" gracePeriod=30 Feb 25 08:45:14 crc kubenswrapper[4978]: I0225 08:45:14.371748 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="34469660-662f-451a-b4b5-b002f087312c" containerName="cinder-api" containerID="cri-o://6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8" gracePeriod=30 Feb 25 08:45:14 crc kubenswrapper[4978]: I0225 08:45:14.418156 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"123dbda5-f0ac-465e-bd80-cbf8ede8722b","Type":"ContainerStarted","Data":"7bb093dc4dec1d1193cc30fc0889f87c94f4750a9eb6556a10f3fe35d02db6b0"} Feb 25 08:45:15 crc kubenswrapper[4978]: I0225 08:45:15.433172 4978 generic.go:334] "Generic (PLEG): container finished" podID="34469660-662f-451a-b4b5-b002f087312c" containerID="26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408" exitCode=143 Feb 25 08:45:15 crc kubenswrapper[4978]: I0225 08:45:15.433305 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34469660-662f-451a-b4b5-b002f087312c","Type":"ContainerDied","Data":"26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408"} Feb 25 08:45:15 crc kubenswrapper[4978]: I0225 08:45:15.436531 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"123dbda5-f0ac-465e-bd80-cbf8ede8722b","Type":"ContainerStarted","Data":"68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2"} Feb 25 08:45:15 crc kubenswrapper[4978]: I0225 08:45:15.436579 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"123dbda5-f0ac-465e-bd80-cbf8ede8722b","Type":"ContainerStarted","Data":"f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf"} Feb 25 08:45:15 crc kubenswrapper[4978]: I0225 08:45:15.466596 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.182275914 podStartE2EDuration="3.466572532s" podCreationTimestamp="2026-02-25 08:45:12 +0000 UTC" firstStartedPulling="2026-02-25 08:45:13.639301931 +0000 UTC m=+7207.078558390" lastFinishedPulling="2026-02-25 08:45:13.923598509 +0000 UTC m=+7207.362855008" observedRunningTime="2026-02-25 08:45:15.459816312 +0000 UTC m=+7208.899072831" watchObservedRunningTime="2026-02-25 08:45:15.466572532 +0000 UTC m=+7208.905829051" Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.764297 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bpd6r"] Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.766601 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.799677 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bpd6r"] Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.857265 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-utilities\") pod \"redhat-operators-bpd6r\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.857364 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jltjc\" (UniqueName: \"kubernetes.io/projected/50c3dcff-1b40-4cbd-a680-11a44a18c26f-kube-api-access-jltjc\") pod \"redhat-operators-bpd6r\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.857660 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-catalog-content\") pod \"redhat-operators-bpd6r\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.960016 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-catalog-content\") pod \"redhat-operators-bpd6r\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.960158 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-utilities\") pod \"redhat-operators-bpd6r\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.960212 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jltjc\" (UniqueName: \"kubernetes.io/projected/50c3dcff-1b40-4cbd-a680-11a44a18c26f-kube-api-access-jltjc\") pod \"redhat-operators-bpd6r\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.960712 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-catalog-content\") pod \"redhat-operators-bpd6r\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.960826 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-utilities\") pod \"redhat-operators-bpd6r\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:16 crc kubenswrapper[4978]: I0225 08:45:16.983696 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jltjc\" (UniqueName: \"kubernetes.io/projected/50c3dcff-1b40-4cbd-a680-11a44a18c26f-kube-api-access-jltjc\") pod \"redhat-operators-bpd6r\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.096413 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.332495 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:45:17 crc kubenswrapper[4978]: E0225 08:45:17.332943 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.548523 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bpd6r"] Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.903179 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.976118 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34469660-662f-451a-b4b5-b002f087312c-logs\") pod \"34469660-662f-451a-b4b5-b002f087312c\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.976215 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-combined-ca-bundle\") pod \"34469660-662f-451a-b4b5-b002f087312c\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.976286 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-scripts\") pod \"34469660-662f-451a-b4b5-b002f087312c\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.976356 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-internal-tls-certs\") pod \"34469660-662f-451a-b4b5-b002f087312c\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.976460 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-public-tls-certs\") pod \"34469660-662f-451a-b4b5-b002f087312c\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.976510 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data-custom\") pod \"34469660-662f-451a-b4b5-b002f087312c\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.976557 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data\") pod \"34469660-662f-451a-b4b5-b002f087312c\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.976649 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lgsg\" (UniqueName: \"kubernetes.io/projected/34469660-662f-451a-b4b5-b002f087312c-kube-api-access-5lgsg\") pod \"34469660-662f-451a-b4b5-b002f087312c\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.976675 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34469660-662f-451a-b4b5-b002f087312c-etc-machine-id\") pod \"34469660-662f-451a-b4b5-b002f087312c\" (UID: \"34469660-662f-451a-b4b5-b002f087312c\") " Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.976707 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34469660-662f-451a-b4b5-b002f087312c-logs" (OuterVolumeSpecName: "logs") pod "34469660-662f-451a-b4b5-b002f087312c" (UID: "34469660-662f-451a-b4b5-b002f087312c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.977073 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34469660-662f-451a-b4b5-b002f087312c-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.977137 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34469660-662f-451a-b4b5-b002f087312c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "34469660-662f-451a-b4b5-b002f087312c" (UID: "34469660-662f-451a-b4b5-b002f087312c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.983599 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34469660-662f-451a-b4b5-b002f087312c-kube-api-access-5lgsg" (OuterVolumeSpecName: "kube-api-access-5lgsg") pod "34469660-662f-451a-b4b5-b002f087312c" (UID: "34469660-662f-451a-b4b5-b002f087312c"). InnerVolumeSpecName "kube-api-access-5lgsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.992266 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "34469660-662f-451a-b4b5-b002f087312c" (UID: "34469660-662f-451a-b4b5-b002f087312c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:17 crc kubenswrapper[4978]: I0225 08:45:17.993781 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-scripts" (OuterVolumeSpecName: "scripts") pod "34469660-662f-451a-b4b5-b002f087312c" (UID: "34469660-662f-451a-b4b5-b002f087312c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.012564 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34469660-662f-451a-b4b5-b002f087312c" (UID: "34469660-662f-451a-b4b5-b002f087312c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.045879 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data" (OuterVolumeSpecName: "config-data") pod "34469660-662f-451a-b4b5-b002f087312c" (UID: "34469660-662f-451a-b4b5-b002f087312c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.052196 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "34469660-662f-451a-b4b5-b002f087312c" (UID: "34469660-662f-451a-b4b5-b002f087312c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.072450 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "34469660-662f-451a-b4b5-b002f087312c" (UID: "34469660-662f-451a-b4b5-b002f087312c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.078820 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.078849 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.078860 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.078869 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.078883 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.078890 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34469660-662f-451a-b4b5-b002f087312c-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.078898 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lgsg\" (UniqueName: \"kubernetes.io/projected/34469660-662f-451a-b4b5-b002f087312c-kube-api-access-5lgsg\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.078909 4978 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34469660-662f-451a-b4b5-b002f087312c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.080257 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.464356 4978 generic.go:334] "Generic (PLEG): container finished" podID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerID="e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a" exitCode=0 Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.464429 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bpd6r" event={"ID":"50c3dcff-1b40-4cbd-a680-11a44a18c26f","Type":"ContainerDied","Data":"e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a"} Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.464739 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bpd6r" event={"ID":"50c3dcff-1b40-4cbd-a680-11a44a18c26f","Type":"ContainerStarted","Data":"4ec8dcfff11b1f6a3ae6b73c799941e5d61f86c00d6b82df4d9e8db88f0222dd"} Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.466727 4978 generic.go:334] "Generic (PLEG): container finished" podID="34469660-662f-451a-b4b5-b002f087312c" containerID="6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8" exitCode=0 Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.466751 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34469660-662f-451a-b4b5-b002f087312c","Type":"ContainerDied","Data":"6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8"} Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.466767 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34469660-662f-451a-b4b5-b002f087312c","Type":"ContainerDied","Data":"f2dd6c2bcff587a18a465be692c4052f76ae15138ae9a98ee11e40aa4db8b3ca"} Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.466784 4978 scope.go:117] "RemoveContainer" containerID="6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.466823 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.495317 4978 scope.go:117] "RemoveContainer" containerID="26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.511424 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.522398 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.526082 4978 scope.go:117] "RemoveContainer" containerID="6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8" Feb 25 08:45:18 crc kubenswrapper[4978]: E0225 08:45:18.526759 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8\": container with ID starting with 6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8 not found: ID does not exist" containerID="6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.526818 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8"} err="failed to get container status \"6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8\": rpc error: code = NotFound desc = could not find container \"6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8\": container with ID starting with 6fbb8714fcd26e3ee99ef572bc5fa6e3af9230f013c19e5c02683762ce9202d8 not found: ID does not exist" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.526854 4978 scope.go:117] "RemoveContainer" containerID="26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408" Feb 25 08:45:18 crc kubenswrapper[4978]: E0225 08:45:18.527232 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408\": container with ID starting with 26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408 not found: ID does not exist" containerID="26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.527282 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408"} err="failed to get container status \"26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408\": rpc error: code = NotFound desc = could not find container \"26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408\": container with ID starting with 26736ca8b2d2371dba7d8cf433cd0ca4c126eb587e1e2fb1a7e200b32cc7f408 not found: ID does not exist" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.532849 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:45:18 crc kubenswrapper[4978]: E0225 08:45:18.533415 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34469660-662f-451a-b4b5-b002f087312c" containerName="cinder-api-log" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.533442 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="34469660-662f-451a-b4b5-b002f087312c" containerName="cinder-api-log" Feb 25 08:45:18 crc kubenswrapper[4978]: E0225 08:45:18.533474 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34469660-662f-451a-b4b5-b002f087312c" containerName="cinder-api" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.533488 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="34469660-662f-451a-b4b5-b002f087312c" containerName="cinder-api" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.533842 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="34469660-662f-451a-b4b5-b002f087312c" containerName="cinder-api-log" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.533877 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="34469660-662f-451a-b4b5-b002f087312c" containerName="cinder-api" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.535492 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.538110 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.538305 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.538434 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.542769 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.689768 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-public-tls-certs\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.689920 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/038563a2-70b7-4050-9183-a4af40007288-etc-machine-id\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.689943 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/038563a2-70b7-4050-9183-a4af40007288-logs\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.689991 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-config-data-custom\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.690007 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-scripts\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.690049 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.690075 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvxvp\" (UniqueName: \"kubernetes.io/projected/038563a2-70b7-4050-9183-a4af40007288-kube-api-access-mvxvp\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.690100 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-config-data\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.690188 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.791834 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.791917 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvxvp\" (UniqueName: \"kubernetes.io/projected/038563a2-70b7-4050-9183-a4af40007288-kube-api-access-mvxvp\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.791983 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-config-data\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.792053 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.792152 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-public-tls-certs\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.792295 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/038563a2-70b7-4050-9183-a4af40007288-etc-machine-id\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.792331 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/038563a2-70b7-4050-9183-a4af40007288-logs\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.792410 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-config-data-custom\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.792514 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-scripts\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.792529 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/038563a2-70b7-4050-9183-a4af40007288-etc-machine-id\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.793147 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/038563a2-70b7-4050-9183-a4af40007288-logs\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.800018 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.800103 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-scripts\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.803418 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.805167 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-config-data-custom\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.806685 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-config-data\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.817469 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/038563a2-70b7-4050-9183-a4af40007288-public-tls-certs\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.829313 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvxvp\" (UniqueName: \"kubernetes.io/projected/038563a2-70b7-4050-9183-a4af40007288-kube-api-access-mvxvp\") pod \"cinder-api-0\" (UID: \"038563a2-70b7-4050-9183-a4af40007288\") " pod="openstack/cinder-api-0" Feb 25 08:45:18 crc kubenswrapper[4978]: I0225 08:45:18.855027 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 25 08:45:19 crc kubenswrapper[4978]: I0225 08:45:19.345819 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34469660-662f-451a-b4b5-b002f087312c" path="/var/lib/kubelet/pods/34469660-662f-451a-b4b5-b002f087312c/volumes" Feb 25 08:45:19 crc kubenswrapper[4978]: I0225 08:45:19.476254 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bpd6r" event={"ID":"50c3dcff-1b40-4cbd-a680-11a44a18c26f","Type":"ContainerStarted","Data":"a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319"} Feb 25 08:45:19 crc kubenswrapper[4978]: W0225 08:45:19.490018 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod038563a2_70b7_4050_9183_a4af40007288.slice/crio-e9a5e780365c9039c6562144c41f0ff8497f8d3626d41986f7b225dd3b7d3e17 WatchSource:0}: Error finding container e9a5e780365c9039c6562144c41f0ff8497f8d3626d41986f7b225dd3b7d3e17: Status 404 returned error can't find the container with id e9a5e780365c9039c6562144c41f0ff8497f8d3626d41986f7b225dd3b7d3e17 Feb 25 08:45:19 crc kubenswrapper[4978]: I0225 08:45:19.491627 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 25 08:45:20 crc kubenswrapper[4978]: I0225 08:45:20.492610 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"038563a2-70b7-4050-9183-a4af40007288","Type":"ContainerStarted","Data":"9f6be8fbc07aa0028255f76b9c87fc2529152513d8fee5d128b0368a71cfe038"} Feb 25 08:45:20 crc kubenswrapper[4978]: I0225 08:45:20.492958 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"038563a2-70b7-4050-9183-a4af40007288","Type":"ContainerStarted","Data":"e9a5e780365c9039c6562144c41f0ff8497f8d3626d41986f7b225dd3b7d3e17"} Feb 25 08:45:21 crc kubenswrapper[4978]: I0225 08:45:21.505592 4978 generic.go:334] "Generic (PLEG): container finished" podID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerID="a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319" exitCode=0 Feb 25 08:45:21 crc kubenswrapper[4978]: I0225 08:45:21.505671 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bpd6r" event={"ID":"50c3dcff-1b40-4cbd-a680-11a44a18c26f","Type":"ContainerDied","Data":"a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319"} Feb 25 08:45:21 crc kubenswrapper[4978]: I0225 08:45:21.510170 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"038563a2-70b7-4050-9183-a4af40007288","Type":"ContainerStarted","Data":"5521342e1867567227f6aff39932f5fca8de108288db4f779727f4e909ad75af"} Feb 25 08:45:21 crc kubenswrapper[4978]: I0225 08:45:21.510354 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 25 08:45:21 crc kubenswrapper[4978]: I0225 08:45:21.574585 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.574561072 podStartE2EDuration="3.574561072s" podCreationTimestamp="2026-02-25 08:45:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:45:21.569184026 +0000 UTC m=+7215.008440565" watchObservedRunningTime="2026-02-25 08:45:21.574561072 +0000 UTC m=+7215.013817561" Feb 25 08:45:22 crc kubenswrapper[4978]: I0225 08:45:22.527494 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bpd6r" event={"ID":"50c3dcff-1b40-4cbd-a680-11a44a18c26f","Type":"ContainerStarted","Data":"e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55"} Feb 25 08:45:22 crc kubenswrapper[4978]: I0225 08:45:22.567991 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bpd6r" podStartSLOduration=3.05843782 podStartE2EDuration="6.567954698s" podCreationTimestamp="2026-02-25 08:45:16 +0000 UTC" firstStartedPulling="2026-02-25 08:45:18.467038919 +0000 UTC m=+7211.906295378" lastFinishedPulling="2026-02-25 08:45:21.976555787 +0000 UTC m=+7215.415812256" observedRunningTime="2026-02-25 08:45:22.551991054 +0000 UTC m=+7215.991247593" watchObservedRunningTime="2026-02-25 08:45:22.567954698 +0000 UTC m=+7216.007211197" Feb 25 08:45:23 crc kubenswrapper[4978]: I0225 08:45:23.281557 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 25 08:45:23 crc kubenswrapper[4978]: I0225 08:45:23.352113 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 08:45:23 crc kubenswrapper[4978]: I0225 08:45:23.533823 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" containerName="cinder-scheduler" containerID="cri-o://f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf" gracePeriod=30 Feb 25 08:45:23 crc kubenswrapper[4978]: I0225 08:45:23.533889 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" containerName="probe" containerID="cri-o://68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2" gracePeriod=30 Feb 25 08:45:24 crc kubenswrapper[4978]: I0225 08:45:24.548292 4978 generic.go:334] "Generic (PLEG): container finished" podID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" containerID="68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2" exitCode=0 Feb 25 08:45:24 crc kubenswrapper[4978]: I0225 08:45:24.548333 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"123dbda5-f0ac-465e-bd80-cbf8ede8722b","Type":"ContainerDied","Data":"68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2"} Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.543587 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.557350 4978 generic.go:334] "Generic (PLEG): container finished" podID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" containerID="f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf" exitCode=0 Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.557398 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"123dbda5-f0ac-465e-bd80-cbf8ede8722b","Type":"ContainerDied","Data":"f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf"} Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.557450 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"123dbda5-f0ac-465e-bd80-cbf8ede8722b","Type":"ContainerDied","Data":"7bb093dc4dec1d1193cc30fc0889f87c94f4750a9eb6556a10f3fe35d02db6b0"} Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.557447 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.557520 4978 scope.go:117] "RemoveContainer" containerID="68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.586030 4978 scope.go:117] "RemoveContainer" containerID="f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.605879 4978 scope.go:117] "RemoveContainer" containerID="68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2" Feb 25 08:45:25 crc kubenswrapper[4978]: E0225 08:45:25.606402 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2\": container with ID starting with 68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2 not found: ID does not exist" containerID="68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.606447 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2"} err="failed to get container status \"68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2\": rpc error: code = NotFound desc = could not find container \"68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2\": container with ID starting with 68c38dd3a32f4d9e403b41ba53cf9a1ba252f8a16f4dd512692dfd87f3a980a2 not found: ID does not exist" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.606474 4978 scope.go:117] "RemoveContainer" containerID="f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf" Feb 25 08:45:25 crc kubenswrapper[4978]: E0225 08:45:25.606939 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf\": container with ID starting with f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf not found: ID does not exist" containerID="f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.606995 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf"} err="failed to get container status \"f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf\": rpc error: code = NotFound desc = could not find container \"f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf\": container with ID starting with f27297e0ca4e922ec6152f1a2952ba002b6c0ede9b5fea6fdf7a022a13c83dbf not found: ID does not exist" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.658752 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data\") pod \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.658802 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6tt2\" (UniqueName: \"kubernetes.io/projected/123dbda5-f0ac-465e-bd80-cbf8ede8722b-kube-api-access-s6tt2\") pod \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.658945 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/123dbda5-f0ac-465e-bd80-cbf8ede8722b-etc-machine-id\") pod \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.658972 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data-custom\") pod \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.659014 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-scripts\") pod \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.659068 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-combined-ca-bundle\") pod \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\" (UID: \"123dbda5-f0ac-465e-bd80-cbf8ede8722b\") " Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.659069 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/123dbda5-f0ac-465e-bd80-cbf8ede8722b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "123dbda5-f0ac-465e-bd80-cbf8ede8722b" (UID: "123dbda5-f0ac-465e-bd80-cbf8ede8722b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.659454 4978 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/123dbda5-f0ac-465e-bd80-cbf8ede8722b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.665636 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/123dbda5-f0ac-465e-bd80-cbf8ede8722b-kube-api-access-s6tt2" (OuterVolumeSpecName: "kube-api-access-s6tt2") pod "123dbda5-f0ac-465e-bd80-cbf8ede8722b" (UID: "123dbda5-f0ac-465e-bd80-cbf8ede8722b"). InnerVolumeSpecName "kube-api-access-s6tt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.665663 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "123dbda5-f0ac-465e-bd80-cbf8ede8722b" (UID: "123dbda5-f0ac-465e-bd80-cbf8ede8722b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.666455 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-scripts" (OuterVolumeSpecName: "scripts") pod "123dbda5-f0ac-465e-bd80-cbf8ede8722b" (UID: "123dbda5-f0ac-465e-bd80-cbf8ede8722b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.725624 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "123dbda5-f0ac-465e-bd80-cbf8ede8722b" (UID: "123dbda5-f0ac-465e-bd80-cbf8ede8722b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.761536 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.761569 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.761579 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.761588 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6tt2\" (UniqueName: \"kubernetes.io/projected/123dbda5-f0ac-465e-bd80-cbf8ede8722b-kube-api-access-s6tt2\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.764282 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data" (OuterVolumeSpecName: "config-data") pod "123dbda5-f0ac-465e-bd80-cbf8ede8722b" (UID: "123dbda5-f0ac-465e-bd80-cbf8ede8722b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.864523 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/123dbda5-f0ac-465e-bd80-cbf8ede8722b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.906599 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.912337 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.931211 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 08:45:25 crc kubenswrapper[4978]: E0225 08:45:25.931559 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" containerName="probe" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.931574 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" containerName="probe" Feb 25 08:45:25 crc kubenswrapper[4978]: E0225 08:45:25.931588 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" containerName="cinder-scheduler" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.931595 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" containerName="cinder-scheduler" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.931756 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" containerName="probe" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.931786 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" containerName="cinder-scheduler" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.933113 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.935509 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 25 08:45:25 crc kubenswrapper[4978]: I0225 08:45:25.945590 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.066944 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.067293 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpssr\" (UniqueName: \"kubernetes.io/projected/2a2818b2-3da2-45cb-b9b1-d27341a11934-kube-api-access-rpssr\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.067338 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-scripts\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.067359 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.067414 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-config-data\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.067447 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a2818b2-3da2-45cb-b9b1-d27341a11934-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.169146 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.169202 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpssr\" (UniqueName: \"kubernetes.io/projected/2a2818b2-3da2-45cb-b9b1-d27341a11934-kube-api-access-rpssr\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.169251 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-scripts\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.169276 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.169315 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-config-data\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.169348 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a2818b2-3da2-45cb-b9b1-d27341a11934-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.169555 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a2818b2-3da2-45cb-b9b1-d27341a11934-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.174521 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.174687 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.174908 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-config-data\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.175356 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a2818b2-3da2-45cb-b9b1-d27341a11934-scripts\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.196303 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpssr\" (UniqueName: \"kubernetes.io/projected/2a2818b2-3da2-45cb-b9b1-d27341a11934-kube-api-access-rpssr\") pod \"cinder-scheduler-0\" (UID: \"2a2818b2-3da2-45cb-b9b1-d27341a11934\") " pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.260288 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 25 08:45:26 crc kubenswrapper[4978]: I0225 08:45:26.755222 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 25 08:45:27 crc kubenswrapper[4978]: I0225 08:45:27.097438 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:27 crc kubenswrapper[4978]: I0225 08:45:27.099479 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:27 crc kubenswrapper[4978]: I0225 08:45:27.338123 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="123dbda5-f0ac-465e-bd80-cbf8ede8722b" path="/var/lib/kubelet/pods/123dbda5-f0ac-465e-bd80-cbf8ede8722b/volumes" Feb 25 08:45:27 crc kubenswrapper[4978]: I0225 08:45:27.590454 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2a2818b2-3da2-45cb-b9b1-d27341a11934","Type":"ContainerStarted","Data":"5ad1d1193c7a2d0037b2a0ba6792e4c2e1a91e3d3ea44e01bd0b1ca603b2bd29"} Feb 25 08:45:27 crc kubenswrapper[4978]: I0225 08:45:27.590498 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2a2818b2-3da2-45cb-b9b1-d27341a11934","Type":"ContainerStarted","Data":"88dda1a0b44a3076134921c926fa671f715aff76cf090e3a9ff5cbe21c34fd27"} Feb 25 08:45:28 crc kubenswrapper[4978]: I0225 08:45:28.161445 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bpd6r" podUID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerName="registry-server" probeResult="failure" output=< Feb 25 08:45:28 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 08:45:28 crc kubenswrapper[4978]: > Feb 25 08:45:28 crc kubenswrapper[4978]: I0225 08:45:28.342112 4978 scope.go:117] "RemoveContainer" containerID="21c86d0f01079552721b78ea7c677a42c0056cb0501c68e66ecab2e8deabee6f" Feb 25 08:45:28 crc kubenswrapper[4978]: I0225 08:45:28.401038 4978 scope.go:117] "RemoveContainer" containerID="8aa1ba4fff420c5b41239f8e1cdcc3abf5bd6e845b1a71ad19b7fae8a1514784" Feb 25 08:45:28 crc kubenswrapper[4978]: I0225 08:45:28.607488 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2a2818b2-3da2-45cb-b9b1-d27341a11934","Type":"ContainerStarted","Data":"5868f3b831bd4801d78987b1e427007c313828c469ca6493d33266bb1780823d"} Feb 25 08:45:28 crc kubenswrapper[4978]: I0225 08:45:28.638747 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.638719127 podStartE2EDuration="3.638719127s" podCreationTimestamp="2026-02-25 08:45:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:45:28.629068267 +0000 UTC m=+7222.068324746" watchObservedRunningTime="2026-02-25 08:45:28.638719127 +0000 UTC m=+7222.077975596" Feb 25 08:45:30 crc kubenswrapper[4978]: I0225 08:45:30.665170 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 25 08:45:31 crc kubenswrapper[4978]: I0225 08:45:31.261380 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 25 08:45:32 crc kubenswrapper[4978]: I0225 08:45:32.328584 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:45:32 crc kubenswrapper[4978]: E0225 08:45:32.329026 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:45:36 crc kubenswrapper[4978]: I0225 08:45:36.460011 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 25 08:45:37 crc kubenswrapper[4978]: I0225 08:45:37.157809 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:37 crc kubenswrapper[4978]: I0225 08:45:37.212884 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:37 crc kubenswrapper[4978]: I0225 08:45:37.406681 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bpd6r"] Feb 25 08:45:38 crc kubenswrapper[4978]: I0225 08:45:38.700726 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bpd6r" podUID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerName="registry-server" containerID="cri-o://e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55" gracePeriod=2 Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.158490 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.335263 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-utilities\") pod \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.335412 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jltjc\" (UniqueName: \"kubernetes.io/projected/50c3dcff-1b40-4cbd-a680-11a44a18c26f-kube-api-access-jltjc\") pod \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.335452 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-catalog-content\") pod \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\" (UID: \"50c3dcff-1b40-4cbd-a680-11a44a18c26f\") " Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.336821 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-utilities" (OuterVolumeSpecName: "utilities") pod "50c3dcff-1b40-4cbd-a680-11a44a18c26f" (UID: "50c3dcff-1b40-4cbd-a680-11a44a18c26f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.343721 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50c3dcff-1b40-4cbd-a680-11a44a18c26f-kube-api-access-jltjc" (OuterVolumeSpecName: "kube-api-access-jltjc") pod "50c3dcff-1b40-4cbd-a680-11a44a18c26f" (UID: "50c3dcff-1b40-4cbd-a680-11a44a18c26f"). InnerVolumeSpecName "kube-api-access-jltjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.437736 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.438909 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jltjc\" (UniqueName: \"kubernetes.io/projected/50c3dcff-1b40-4cbd-a680-11a44a18c26f-kube-api-access-jltjc\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.467444 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50c3dcff-1b40-4cbd-a680-11a44a18c26f" (UID: "50c3dcff-1b40-4cbd-a680-11a44a18c26f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.541178 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c3dcff-1b40-4cbd-a680-11a44a18c26f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.713544 4978 generic.go:334] "Generic (PLEG): container finished" podID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerID="e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55" exitCode=0 Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.713600 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bpd6r" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.713633 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bpd6r" event={"ID":"50c3dcff-1b40-4cbd-a680-11a44a18c26f","Type":"ContainerDied","Data":"e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55"} Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.713995 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bpd6r" event={"ID":"50c3dcff-1b40-4cbd-a680-11a44a18c26f","Type":"ContainerDied","Data":"4ec8dcfff11b1f6a3ae6b73c799941e5d61f86c00d6b82df4d9e8db88f0222dd"} Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.714030 4978 scope.go:117] "RemoveContainer" containerID="e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.744526 4978 scope.go:117] "RemoveContainer" containerID="a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.767193 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bpd6r"] Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.775128 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bpd6r"] Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.785447 4978 scope.go:117] "RemoveContainer" containerID="e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.822365 4978 scope.go:117] "RemoveContainer" containerID="e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55" Feb 25 08:45:39 crc kubenswrapper[4978]: E0225 08:45:39.822942 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55\": container with ID starting with e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55 not found: ID does not exist" containerID="e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.823006 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55"} err="failed to get container status \"e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55\": rpc error: code = NotFound desc = could not find container \"e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55\": container with ID starting with e82c729fbacf761856bdb3eb3518c43d971192b5c0b876b7d0e730e4293c6c55 not found: ID does not exist" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.823042 4978 scope.go:117] "RemoveContainer" containerID="a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319" Feb 25 08:45:39 crc kubenswrapper[4978]: E0225 08:45:39.823655 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319\": container with ID starting with a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319 not found: ID does not exist" containerID="a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.823695 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319"} err="failed to get container status \"a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319\": rpc error: code = NotFound desc = could not find container \"a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319\": container with ID starting with a19078a65188441d65d4709e342f197017cec9ac97de8f696812243c4463e319 not found: ID does not exist" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.823729 4978 scope.go:117] "RemoveContainer" containerID="e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a" Feb 25 08:45:39 crc kubenswrapper[4978]: E0225 08:45:39.824075 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a\": container with ID starting with e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a not found: ID does not exist" containerID="e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a" Feb 25 08:45:39 crc kubenswrapper[4978]: I0225 08:45:39.824109 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a"} err="failed to get container status \"e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a\": rpc error: code = NotFound desc = could not find container \"e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a\": container with ID starting with e6d051e61cf6d87e384e852f3b628b9e0fc6ecaa5f306a8dd80c9057e007f93a not found: ID does not exist" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.012251 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-rkqhs"] Feb 25 08:45:40 crc kubenswrapper[4978]: E0225 08:45:40.012903 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerName="extract-content" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.012947 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerName="extract-content" Feb 25 08:45:40 crc kubenswrapper[4978]: E0225 08:45:40.012992 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerName="extract-utilities" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.013006 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerName="extract-utilities" Feb 25 08:45:40 crc kubenswrapper[4978]: E0225 08:45:40.013050 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerName="registry-server" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.013063 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerName="registry-server" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.013356 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" containerName="registry-server" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.014409 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rkqhs" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.032357 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rkqhs"] Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.108651 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b014-account-create-update-m2qtz"] Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.109746 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b014-account-create-update-m2qtz" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.113729 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.126629 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b014-account-create-update-m2qtz"] Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.156689 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a891b63-df50-4203-b1e6-c7b86991fec3-operator-scripts\") pod \"glance-db-create-rkqhs\" (UID: \"9a891b63-df50-4203-b1e6-c7b86991fec3\") " pod="openstack/glance-db-create-rkqhs" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.156752 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8vw9\" (UniqueName: \"kubernetes.io/projected/9a891b63-df50-4203-b1e6-c7b86991fec3-kube-api-access-n8vw9\") pod \"glance-db-create-rkqhs\" (UID: \"9a891b63-df50-4203-b1e6-c7b86991fec3\") " pod="openstack/glance-db-create-rkqhs" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.258639 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a891b63-df50-4203-b1e6-c7b86991fec3-operator-scripts\") pod \"glance-db-create-rkqhs\" (UID: \"9a891b63-df50-4203-b1e6-c7b86991fec3\") " pod="openstack/glance-db-create-rkqhs" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.258702 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-operator-scripts\") pod \"glance-b014-account-create-update-m2qtz\" (UID: \"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2\") " pod="openstack/glance-b014-account-create-update-m2qtz" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.258735 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8vw9\" (UniqueName: \"kubernetes.io/projected/9a891b63-df50-4203-b1e6-c7b86991fec3-kube-api-access-n8vw9\") pod \"glance-db-create-rkqhs\" (UID: \"9a891b63-df50-4203-b1e6-c7b86991fec3\") " pod="openstack/glance-db-create-rkqhs" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.258757 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5pj7\" (UniqueName: \"kubernetes.io/projected/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-kube-api-access-p5pj7\") pod \"glance-b014-account-create-update-m2qtz\" (UID: \"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2\") " pod="openstack/glance-b014-account-create-update-m2qtz" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.259823 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a891b63-df50-4203-b1e6-c7b86991fec3-operator-scripts\") pod \"glance-db-create-rkqhs\" (UID: \"9a891b63-df50-4203-b1e6-c7b86991fec3\") " pod="openstack/glance-db-create-rkqhs" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.284241 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8vw9\" (UniqueName: \"kubernetes.io/projected/9a891b63-df50-4203-b1e6-c7b86991fec3-kube-api-access-n8vw9\") pod \"glance-db-create-rkqhs\" (UID: \"9a891b63-df50-4203-b1e6-c7b86991fec3\") " pod="openstack/glance-db-create-rkqhs" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.352050 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rkqhs" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.359916 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-operator-scripts\") pod \"glance-b014-account-create-update-m2qtz\" (UID: \"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2\") " pod="openstack/glance-b014-account-create-update-m2qtz" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.359988 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5pj7\" (UniqueName: \"kubernetes.io/projected/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-kube-api-access-p5pj7\") pod \"glance-b014-account-create-update-m2qtz\" (UID: \"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2\") " pod="openstack/glance-b014-account-create-update-m2qtz" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.361065 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-operator-scripts\") pod \"glance-b014-account-create-update-m2qtz\" (UID: \"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2\") " pod="openstack/glance-b014-account-create-update-m2qtz" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.376594 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5pj7\" (UniqueName: \"kubernetes.io/projected/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-kube-api-access-p5pj7\") pod \"glance-b014-account-create-update-m2qtz\" (UID: \"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2\") " pod="openstack/glance-b014-account-create-update-m2qtz" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.432018 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b014-account-create-update-m2qtz" Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.855326 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-rkqhs"] Feb 25 08:45:40 crc kubenswrapper[4978]: W0225 08:45:40.909737 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34c1a11b_9cf0_4f8a_beb0_14739a3b79c2.slice/crio-09dbe1f731c31a8663c67a704c37b65b9ae62eb6641b1c51ec06555feec49c15 WatchSource:0}: Error finding container 09dbe1f731c31a8663c67a704c37b65b9ae62eb6641b1c51ec06555feec49c15: Status 404 returned error can't find the container with id 09dbe1f731c31a8663c67a704c37b65b9ae62eb6641b1c51ec06555feec49c15 Feb 25 08:45:40 crc kubenswrapper[4978]: I0225 08:45:40.909942 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b014-account-create-update-m2qtz"] Feb 25 08:45:41 crc kubenswrapper[4978]: I0225 08:45:41.343096 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50c3dcff-1b40-4cbd-a680-11a44a18c26f" path="/var/lib/kubelet/pods/50c3dcff-1b40-4cbd-a680-11a44a18c26f/volumes" Feb 25 08:45:41 crc kubenswrapper[4978]: I0225 08:45:41.737612 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a891b63-df50-4203-b1e6-c7b86991fec3" containerID="d04b76696689d976c05d7705c0608c9032fe21d53f4f9a2089fa60ffec5b251a" exitCode=0 Feb 25 08:45:41 crc kubenswrapper[4978]: I0225 08:45:41.737700 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rkqhs" event={"ID":"9a891b63-df50-4203-b1e6-c7b86991fec3","Type":"ContainerDied","Data":"d04b76696689d976c05d7705c0608c9032fe21d53f4f9a2089fa60ffec5b251a"} Feb 25 08:45:41 crc kubenswrapper[4978]: I0225 08:45:41.738049 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rkqhs" event={"ID":"9a891b63-df50-4203-b1e6-c7b86991fec3","Type":"ContainerStarted","Data":"77020ec55843e8a4d7c7ebe4b778f79981b790e4a62890bdfb5d6e6190139d49"} Feb 25 08:45:41 crc kubenswrapper[4978]: I0225 08:45:41.740005 4978 generic.go:334] "Generic (PLEG): container finished" podID="34c1a11b-9cf0-4f8a-beb0-14739a3b79c2" containerID="2763203f1c861e0380f198a1b4d21b94d8355638ad6da04ff4e4a1c1a93ae4a3" exitCode=0 Feb 25 08:45:41 crc kubenswrapper[4978]: I0225 08:45:41.740057 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b014-account-create-update-m2qtz" event={"ID":"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2","Type":"ContainerDied","Data":"2763203f1c861e0380f198a1b4d21b94d8355638ad6da04ff4e4a1c1a93ae4a3"} Feb 25 08:45:41 crc kubenswrapper[4978]: I0225 08:45:41.740115 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b014-account-create-update-m2qtz" event={"ID":"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2","Type":"ContainerStarted","Data":"09dbe1f731c31a8663c67a704c37b65b9ae62eb6641b1c51ec06555feec49c15"} Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.352155 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b014-account-create-update-m2qtz" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.356624 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rkqhs" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.478363 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-operator-scripts\") pod \"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2\" (UID: \"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2\") " Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.478467 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a891b63-df50-4203-b1e6-c7b86991fec3-operator-scripts\") pod \"9a891b63-df50-4203-b1e6-c7b86991fec3\" (UID: \"9a891b63-df50-4203-b1e6-c7b86991fec3\") " Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.478491 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5pj7\" (UniqueName: \"kubernetes.io/projected/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-kube-api-access-p5pj7\") pod \"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2\" (UID: \"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2\") " Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.479193 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a891b63-df50-4203-b1e6-c7b86991fec3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9a891b63-df50-4203-b1e6-c7b86991fec3" (UID: "9a891b63-df50-4203-b1e6-c7b86991fec3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.479269 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "34c1a11b-9cf0-4f8a-beb0-14739a3b79c2" (UID: "34c1a11b-9cf0-4f8a-beb0-14739a3b79c2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.479334 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8vw9\" (UniqueName: \"kubernetes.io/projected/9a891b63-df50-4203-b1e6-c7b86991fec3-kube-api-access-n8vw9\") pod \"9a891b63-df50-4203-b1e6-c7b86991fec3\" (UID: \"9a891b63-df50-4203-b1e6-c7b86991fec3\") " Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.479798 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.479815 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a891b63-df50-4203-b1e6-c7b86991fec3-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.483700 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-kube-api-access-p5pj7" (OuterVolumeSpecName: "kube-api-access-p5pj7") pod "34c1a11b-9cf0-4f8a-beb0-14739a3b79c2" (UID: "34c1a11b-9cf0-4f8a-beb0-14739a3b79c2"). InnerVolumeSpecName "kube-api-access-p5pj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.483906 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a891b63-df50-4203-b1e6-c7b86991fec3-kube-api-access-n8vw9" (OuterVolumeSpecName: "kube-api-access-n8vw9") pod "9a891b63-df50-4203-b1e6-c7b86991fec3" (UID: "9a891b63-df50-4203-b1e6-c7b86991fec3"). InnerVolumeSpecName "kube-api-access-n8vw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.581456 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5pj7\" (UniqueName: \"kubernetes.io/projected/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2-kube-api-access-p5pj7\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.581510 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8vw9\" (UniqueName: \"kubernetes.io/projected/9a891b63-df50-4203-b1e6-c7b86991fec3-kube-api-access-n8vw9\") on node \"crc\" DevicePath \"\"" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.768030 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-rkqhs" event={"ID":"9a891b63-df50-4203-b1e6-c7b86991fec3","Type":"ContainerDied","Data":"77020ec55843e8a4d7c7ebe4b778f79981b790e4a62890bdfb5d6e6190139d49"} Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.768353 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77020ec55843e8a4d7c7ebe4b778f79981b790e4a62890bdfb5d6e6190139d49" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.768110 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-rkqhs" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.770617 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b014-account-create-update-m2qtz" event={"ID":"34c1a11b-9cf0-4f8a-beb0-14739a3b79c2","Type":"ContainerDied","Data":"09dbe1f731c31a8663c67a704c37b65b9ae62eb6641b1c51ec06555feec49c15"} Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.770646 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b014-account-create-update-m2qtz" Feb 25 08:45:43 crc kubenswrapper[4978]: I0225 08:45:43.770670 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09dbe1f731c31a8663c67a704c37b65b9ae62eb6641b1c51ec06555feec49c15" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.294958 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-sxp4q"] Feb 25 08:45:45 crc kubenswrapper[4978]: E0225 08:45:45.295438 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c1a11b-9cf0-4f8a-beb0-14739a3b79c2" containerName="mariadb-account-create-update" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.295460 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c1a11b-9cf0-4f8a-beb0-14739a3b79c2" containerName="mariadb-account-create-update" Feb 25 08:45:45 crc kubenswrapper[4978]: E0225 08:45:45.295514 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a891b63-df50-4203-b1e6-c7b86991fec3" containerName="mariadb-database-create" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.295527 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a891b63-df50-4203-b1e6-c7b86991fec3" containerName="mariadb-database-create" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.295792 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a891b63-df50-4203-b1e6-c7b86991fec3" containerName="mariadb-database-create" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.295823 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c1a11b-9cf0-4f8a-beb0-14739a3b79c2" containerName="mariadb-account-create-update" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.296742 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.300276 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.302689 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-p5b46" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.310815 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sxp4q"] Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.334777 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzk4g\" (UniqueName: \"kubernetes.io/projected/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-kube-api-access-wzk4g\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.334800 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.334888 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-combined-ca-bundle\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.334965 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-db-sync-config-data\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.335031 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-config-data\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: E0225 08:45:45.335075 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.436253 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-db-sync-config-data\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.436363 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-config-data\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.436435 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzk4g\" (UniqueName: \"kubernetes.io/projected/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-kube-api-access-wzk4g\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.436564 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-combined-ca-bundle\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.453754 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-config-data\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.453821 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-combined-ca-bundle\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.453755 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-db-sync-config-data\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.460039 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzk4g\" (UniqueName: \"kubernetes.io/projected/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-kube-api-access-wzk4g\") pod \"glance-db-sync-sxp4q\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:45 crc kubenswrapper[4978]: I0225 08:45:45.640640 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sxp4q" Feb 25 08:45:46 crc kubenswrapper[4978]: W0225 08:45:46.209509 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcee2ed39_9b9f_466d_b2b9_89e3b598e4e5.slice/crio-951314c3eef6671eb0f674ab80eeaa17befee426d54329771f14858360e943c1 WatchSource:0}: Error finding container 951314c3eef6671eb0f674ab80eeaa17befee426d54329771f14858360e943c1: Status 404 returned error can't find the container with id 951314c3eef6671eb0f674ab80eeaa17befee426d54329771f14858360e943c1 Feb 25 08:45:46 crc kubenswrapper[4978]: I0225 08:45:46.219532 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sxp4q"] Feb 25 08:45:46 crc kubenswrapper[4978]: I0225 08:45:46.797294 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sxp4q" event={"ID":"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5","Type":"ContainerStarted","Data":"951314c3eef6671eb0f674ab80eeaa17befee426d54329771f14858360e943c1"} Feb 25 08:45:58 crc kubenswrapper[4978]: I0225 08:45:58.327557 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:45:58 crc kubenswrapper[4978]: E0225 08:45:58.328250 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:46:00 crc kubenswrapper[4978]: I0225 08:46:00.141351 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533486-qzncn"] Feb 25 08:46:00 crc kubenswrapper[4978]: I0225 08:46:00.142975 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533486-qzncn" Feb 25 08:46:00 crc kubenswrapper[4978]: I0225 08:46:00.146196 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:46:00 crc kubenswrapper[4978]: I0225 08:46:00.146535 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:46:00 crc kubenswrapper[4978]: I0225 08:46:00.147191 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:46:00 crc kubenswrapper[4978]: I0225 08:46:00.156932 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533486-qzncn"] Feb 25 08:46:00 crc kubenswrapper[4978]: I0225 08:46:00.242522 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8vtp\" (UniqueName: \"kubernetes.io/projected/2378b284-beb2-444b-aaa9-05a96e14c8c1-kube-api-access-v8vtp\") pod \"auto-csr-approver-29533486-qzncn\" (UID: \"2378b284-beb2-444b-aaa9-05a96e14c8c1\") " pod="openshift-infra/auto-csr-approver-29533486-qzncn" Feb 25 08:46:00 crc kubenswrapper[4978]: I0225 08:46:00.344913 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8vtp\" (UniqueName: \"kubernetes.io/projected/2378b284-beb2-444b-aaa9-05a96e14c8c1-kube-api-access-v8vtp\") pod \"auto-csr-approver-29533486-qzncn\" (UID: \"2378b284-beb2-444b-aaa9-05a96e14c8c1\") " pod="openshift-infra/auto-csr-approver-29533486-qzncn" Feb 25 08:46:00 crc kubenswrapper[4978]: I0225 08:46:00.369950 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8vtp\" (UniqueName: \"kubernetes.io/projected/2378b284-beb2-444b-aaa9-05a96e14c8c1-kube-api-access-v8vtp\") pod \"auto-csr-approver-29533486-qzncn\" (UID: \"2378b284-beb2-444b-aaa9-05a96e14c8c1\") " pod="openshift-infra/auto-csr-approver-29533486-qzncn" Feb 25 08:46:00 crc kubenswrapper[4978]: I0225 08:46:00.476981 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533486-qzncn" Feb 25 08:46:06 crc kubenswrapper[4978]: I0225 08:46:06.667737 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533486-qzncn"] Feb 25 08:46:06 crc kubenswrapper[4978]: I0225 08:46:06.990149 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533486-qzncn" event={"ID":"2378b284-beb2-444b-aaa9-05a96e14c8c1","Type":"ContainerStarted","Data":"4db38ae5a56247f5657dead0b77a5f474470e3672ca94b4fb0d4ef0d2097c9f0"} Feb 25 08:46:06 crc kubenswrapper[4978]: I0225 08:46:06.991360 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sxp4q" event={"ID":"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5","Type":"ContainerStarted","Data":"2db2b82e8d8653b420ee35ddb3dc7d8e6967bc6044563849bb6e8b09eb8d2af0"} Feb 25 08:46:07 crc kubenswrapper[4978]: I0225 08:46:07.378231 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-sxp4q" podStartSLOduration=2.325546042 podStartE2EDuration="22.378204691s" podCreationTimestamp="2026-02-25 08:45:45 +0000 UTC" firstStartedPulling="2026-02-25 08:45:46.211826498 +0000 UTC m=+7239.651082957" lastFinishedPulling="2026-02-25 08:46:06.264485127 +0000 UTC m=+7259.703741606" observedRunningTime="2026-02-25 08:46:07.015601728 +0000 UTC m=+7260.454858197" watchObservedRunningTime="2026-02-25 08:46:07.378204691 +0000 UTC m=+7260.817461190" Feb 25 08:46:08 crc kubenswrapper[4978]: I0225 08:46:08.000464 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533486-qzncn" event={"ID":"2378b284-beb2-444b-aaa9-05a96e14c8c1","Type":"ContainerStarted","Data":"f57e19bdd8f4cc851d980cd420e2ac9670d5ff1fbed8c1e499050f11d05a0cda"} Feb 25 08:46:08 crc kubenswrapper[4978]: I0225 08:46:08.019890 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533486-qzncn" podStartSLOduration=7.232684152 podStartE2EDuration="8.01986866s" podCreationTimestamp="2026-02-25 08:46:00 +0000 UTC" firstStartedPulling="2026-02-25 08:46:06.669240726 +0000 UTC m=+7260.108497205" lastFinishedPulling="2026-02-25 08:46:07.456425244 +0000 UTC m=+7260.895681713" observedRunningTime="2026-02-25 08:46:08.013112911 +0000 UTC m=+7261.452369390" watchObservedRunningTime="2026-02-25 08:46:08.01986866 +0000 UTC m=+7261.459125139" Feb 25 08:46:09 crc kubenswrapper[4978]: I0225 08:46:09.025332 4978 generic.go:334] "Generic (PLEG): container finished" podID="2378b284-beb2-444b-aaa9-05a96e14c8c1" containerID="f57e19bdd8f4cc851d980cd420e2ac9670d5ff1fbed8c1e499050f11d05a0cda" exitCode=0 Feb 25 08:46:09 crc kubenswrapper[4978]: I0225 08:46:09.025417 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533486-qzncn" event={"ID":"2378b284-beb2-444b-aaa9-05a96e14c8c1","Type":"ContainerDied","Data":"f57e19bdd8f4cc851d980cd420e2ac9670d5ff1fbed8c1e499050f11d05a0cda"} Feb 25 08:46:10 crc kubenswrapper[4978]: I0225 08:46:10.038921 4978 generic.go:334] "Generic (PLEG): container finished" podID="cee2ed39-9b9f-466d-b2b9-89e3b598e4e5" containerID="2db2b82e8d8653b420ee35ddb3dc7d8e6967bc6044563849bb6e8b09eb8d2af0" exitCode=0 Feb 25 08:46:10 crc kubenswrapper[4978]: I0225 08:46:10.039027 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sxp4q" event={"ID":"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5","Type":"ContainerDied","Data":"2db2b82e8d8653b420ee35ddb3dc7d8e6967bc6044563849bb6e8b09eb8d2af0"} Feb 25 08:46:10 crc kubenswrapper[4978]: I0225 08:46:10.407252 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533486-qzncn" Feb 25 08:46:10 crc kubenswrapper[4978]: I0225 08:46:10.477479 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8vtp\" (UniqueName: \"kubernetes.io/projected/2378b284-beb2-444b-aaa9-05a96e14c8c1-kube-api-access-v8vtp\") pod \"2378b284-beb2-444b-aaa9-05a96e14c8c1\" (UID: \"2378b284-beb2-444b-aaa9-05a96e14c8c1\") " Feb 25 08:46:10 crc kubenswrapper[4978]: I0225 08:46:10.484597 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2378b284-beb2-444b-aaa9-05a96e14c8c1-kube-api-access-v8vtp" (OuterVolumeSpecName: "kube-api-access-v8vtp") pod "2378b284-beb2-444b-aaa9-05a96e14c8c1" (UID: "2378b284-beb2-444b-aaa9-05a96e14c8c1"). InnerVolumeSpecName "kube-api-access-v8vtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:46:10 crc kubenswrapper[4978]: I0225 08:46:10.580122 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8vtp\" (UniqueName: \"kubernetes.io/projected/2378b284-beb2-444b-aaa9-05a96e14c8c1-kube-api-access-v8vtp\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.052585 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533486-qzncn" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.052590 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533486-qzncn" event={"ID":"2378b284-beb2-444b-aaa9-05a96e14c8c1","Type":"ContainerDied","Data":"4db38ae5a56247f5657dead0b77a5f474470e3672ca94b4fb0d4ef0d2097c9f0"} Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.053161 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4db38ae5a56247f5657dead0b77a5f474470e3672ca94b4fb0d4ef0d2097c9f0" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.124988 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533480-txzm8"] Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.143210 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533480-txzm8"] Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.339353 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0498e889-eade-48ed-8b3f-7f7b2dcb9b00" path="/var/lib/kubelet/pods/0498e889-eade-48ed-8b3f-7f7b2dcb9b00/volumes" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.480474 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sxp4q" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.597123 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-combined-ca-bundle\") pod \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.597246 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzk4g\" (UniqueName: \"kubernetes.io/projected/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-kube-api-access-wzk4g\") pod \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.597328 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-db-sync-config-data\") pod \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.597435 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-config-data\") pod \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\" (UID: \"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5\") " Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.602825 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-kube-api-access-wzk4g" (OuterVolumeSpecName: "kube-api-access-wzk4g") pod "cee2ed39-9b9f-466d-b2b9-89e3b598e4e5" (UID: "cee2ed39-9b9f-466d-b2b9-89e3b598e4e5"). InnerVolumeSpecName "kube-api-access-wzk4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.605702 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cee2ed39-9b9f-466d-b2b9-89e3b598e4e5" (UID: "cee2ed39-9b9f-466d-b2b9-89e3b598e4e5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.642247 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cee2ed39-9b9f-466d-b2b9-89e3b598e4e5" (UID: "cee2ed39-9b9f-466d-b2b9-89e3b598e4e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.688109 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-config-data" (OuterVolumeSpecName: "config-data") pod "cee2ed39-9b9f-466d-b2b9-89e3b598e4e5" (UID: "cee2ed39-9b9f-466d-b2b9-89e3b598e4e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.700412 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.700457 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzk4g\" (UniqueName: \"kubernetes.io/projected/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-kube-api-access-wzk4g\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.700477 4978 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:11 crc kubenswrapper[4978]: I0225 08:46:11.700496 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.066228 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sxp4q" event={"ID":"cee2ed39-9b9f-466d-b2b9-89e3b598e4e5","Type":"ContainerDied","Data":"951314c3eef6671eb0f674ab80eeaa17befee426d54329771f14858360e943c1"} Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.066266 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="951314c3eef6671eb0f674ab80eeaa17befee426d54329771f14858360e943c1" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.066357 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sxp4q" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.556270 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:46:12 crc kubenswrapper[4978]: E0225 08:46:12.556640 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee2ed39-9b9f-466d-b2b9-89e3b598e4e5" containerName="glance-db-sync" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.556659 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee2ed39-9b9f-466d-b2b9-89e3b598e4e5" containerName="glance-db-sync" Feb 25 08:46:12 crc kubenswrapper[4978]: E0225 08:46:12.556687 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2378b284-beb2-444b-aaa9-05a96e14c8c1" containerName="oc" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.556694 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2378b284-beb2-444b-aaa9-05a96e14c8c1" containerName="oc" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.556851 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2378b284-beb2-444b-aaa9-05a96e14c8c1" containerName="oc" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.556872 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee2ed39-9b9f-466d-b2b9-89e3b598e4e5" containerName="glance-db-sync" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.557756 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.559854 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.560084 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-p5b46" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.578228 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.580043 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.586714 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6fd85b86df-857nz"] Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.588380 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.619273 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.619345 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-config-data\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.619395 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt99b\" (UniqueName: \"kubernetes.io/projected/2810ade8-67b0-4661-8df5-83da89cccddc-kube-api-access-mt99b\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.619444 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-scripts\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.619478 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.619525 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-logs\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.670604 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fd85b86df-857nz"] Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.683920 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.685340 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.689687 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725168 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725208 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9h2b\" (UniqueName: \"kubernetes.io/projected/61d85cc4-e12f-44e9-b4a4-c2030412451d-kube-api-access-m9h2b\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725247 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-config-data\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725275 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt99b\" (UniqueName: \"kubernetes.io/projected/2810ade8-67b0-4661-8df5-83da89cccddc-kube-api-access-mt99b\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725317 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-scripts\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725335 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-nb\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725357 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725400 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-sb\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725422 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-logs\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725440 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-dns-svc\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725457 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-config\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725770 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.725829 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.726289 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-logs\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.731005 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-scripts\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.744162 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.745160 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-config-data\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.751310 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt99b\" (UniqueName: \"kubernetes.io/projected/2810ade8-67b0-4661-8df5-83da89cccddc-kube-api-access-mt99b\") pod \"glance-default-external-api-0\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826550 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-logs\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826602 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-nb\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826624 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vclcq\" (UniqueName: \"kubernetes.io/projected/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-kube-api-access-vclcq\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826645 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826680 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-sb\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826708 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-dns-svc\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826727 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-config\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826754 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826768 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826806 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9h2b\" (UniqueName: \"kubernetes.io/projected/61d85cc4-e12f-44e9-b4a4-c2030412451d-kube-api-access-m9h2b\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.826823 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.827706 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-nb\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.828212 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-sb\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.828721 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-dns-svc\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.829117 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-config\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.845510 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9h2b\" (UniqueName: \"kubernetes.io/projected/61d85cc4-e12f-44e9-b4a4-c2030412451d-kube-api-access-m9h2b\") pod \"dnsmasq-dns-6fd85b86df-857nz\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.879564 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.907212 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.927797 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vclcq\" (UniqueName: \"kubernetes.io/projected/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-kube-api-access-vclcq\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.927840 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.927907 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.927926 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.927968 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.928037 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-logs\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.928434 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-logs\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.932644 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.934734 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.936296 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.940091 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:12 crc kubenswrapper[4978]: I0225 08:46:12.946833 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vclcq\" (UniqueName: \"kubernetes.io/projected/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-kube-api-access-vclcq\") pod \"glance-default-internal-api-0\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:13 crc kubenswrapper[4978]: I0225 08:46:13.020161 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:13 crc kubenswrapper[4978]: I0225 08:46:13.328663 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:46:13 crc kubenswrapper[4978]: E0225 08:46:13.329112 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:46:13 crc kubenswrapper[4978]: I0225 08:46:13.397385 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6fd85b86df-857nz"] Feb 25 08:46:13 crc kubenswrapper[4978]: I0225 08:46:13.601621 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:46:13 crc kubenswrapper[4978]: I0225 08:46:13.674029 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:46:13 crc kubenswrapper[4978]: W0225 08:46:13.707758 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a6112f8_f64d_4391_8b79_3b8f7f3eb555.slice/crio-064e09c05c7ebfab75c0223db9816738c9e3d3ec56a7ebf85f92b0ac4e23f77f WatchSource:0}: Error finding container 064e09c05c7ebfab75c0223db9816738c9e3d3ec56a7ebf85f92b0ac4e23f77f: Status 404 returned error can't find the container with id 064e09c05c7ebfab75c0223db9816738c9e3d3ec56a7ebf85f92b0ac4e23f77f Feb 25 08:46:13 crc kubenswrapper[4978]: I0225 08:46:13.744428 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:46:14 crc kubenswrapper[4978]: I0225 08:46:14.114049 4978 generic.go:334] "Generic (PLEG): container finished" podID="61d85cc4-e12f-44e9-b4a4-c2030412451d" containerID="8e30ab1588fa07a5bf999816f9e51784a30f27a7b84f8f5c341182a636f60fb6" exitCode=0 Feb 25 08:46:14 crc kubenswrapper[4978]: I0225 08:46:14.114617 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" event={"ID":"61d85cc4-e12f-44e9-b4a4-c2030412451d","Type":"ContainerDied","Data":"8e30ab1588fa07a5bf999816f9e51784a30f27a7b84f8f5c341182a636f60fb6"} Feb 25 08:46:14 crc kubenswrapper[4978]: I0225 08:46:14.114665 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" event={"ID":"61d85cc4-e12f-44e9-b4a4-c2030412451d","Type":"ContainerStarted","Data":"8994e9ea2001281f6f4caae13be2a83fda33bb35e192b806bf1e70a6956593b4"} Feb 25 08:46:14 crc kubenswrapper[4978]: I0225 08:46:14.133901 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a6112f8-f64d-4391-8b79-3b8f7f3eb555","Type":"ContainerStarted","Data":"064e09c05c7ebfab75c0223db9816738c9e3d3ec56a7ebf85f92b0ac4e23f77f"} Feb 25 08:46:14 crc kubenswrapper[4978]: I0225 08:46:14.140741 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2810ade8-67b0-4661-8df5-83da89cccddc","Type":"ContainerStarted","Data":"97822b5ff628565a2b7a9cf235339a1729fb77d1c7bc3c7a60de0c5c2beff4c8"} Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.158937 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2810ade8-67b0-4661-8df5-83da89cccddc","Type":"ContainerStarted","Data":"0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49"} Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.160040 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2810ade8-67b0-4661-8df5-83da89cccddc","Type":"ContainerStarted","Data":"8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856"} Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.159654 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2810ade8-67b0-4661-8df5-83da89cccddc" containerName="glance-log" containerID="cri-o://8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856" gracePeriod=30 Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.160102 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2810ade8-67b0-4661-8df5-83da89cccddc" containerName="glance-httpd" containerID="cri-o://0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49" gracePeriod=30 Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.164922 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" event={"ID":"61d85cc4-e12f-44e9-b4a4-c2030412451d","Type":"ContainerStarted","Data":"d23247fee0326055409949cfa5eca34a48f4f213a8438dcf20c1956647325a5a"} Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.165410 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.172626 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a6112f8-f64d-4391-8b79-3b8f7f3eb555","Type":"ContainerStarted","Data":"7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342"} Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.172677 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a6112f8-f64d-4391-8b79-3b8f7f3eb555","Type":"ContainerStarted","Data":"1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262"} Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.185354 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.185334294 podStartE2EDuration="3.185334294s" podCreationTimestamp="2026-02-25 08:46:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:46:15.177673686 +0000 UTC m=+7268.616930145" watchObservedRunningTime="2026-02-25 08:46:15.185334294 +0000 UTC m=+7268.624590753" Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.205931 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.206505 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" podStartSLOduration=3.206483728 podStartE2EDuration="3.206483728s" podCreationTimestamp="2026-02-25 08:46:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:46:15.202637309 +0000 UTC m=+7268.641893798" watchObservedRunningTime="2026-02-25 08:46:15.206483728 +0000 UTC m=+7268.645740187" Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.235908 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.235877168 podStartE2EDuration="3.235877168s" podCreationTimestamp="2026-02-25 08:46:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:46:15.226061465 +0000 UTC m=+7268.665317954" watchObservedRunningTime="2026-02-25 08:46:15.235877168 +0000 UTC m=+7268.675133627" Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.888107 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.987911 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-combined-ca-bundle\") pod \"2810ade8-67b0-4661-8df5-83da89cccddc\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.988222 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-httpd-run\") pod \"2810ade8-67b0-4661-8df5-83da89cccddc\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.988426 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-logs\") pod \"2810ade8-67b0-4661-8df5-83da89cccddc\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.988460 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt99b\" (UniqueName: \"kubernetes.io/projected/2810ade8-67b0-4661-8df5-83da89cccddc-kube-api-access-mt99b\") pod \"2810ade8-67b0-4661-8df5-83da89cccddc\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.988551 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-scripts\") pod \"2810ade8-67b0-4661-8df5-83da89cccddc\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.988602 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2810ade8-67b0-4661-8df5-83da89cccddc" (UID: "2810ade8-67b0-4661-8df5-83da89cccddc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.988668 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-config-data\") pod \"2810ade8-67b0-4661-8df5-83da89cccddc\" (UID: \"2810ade8-67b0-4661-8df5-83da89cccddc\") " Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.988751 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-logs" (OuterVolumeSpecName: "logs") pod "2810ade8-67b0-4661-8df5-83da89cccddc" (UID: "2810ade8-67b0-4661-8df5-83da89cccddc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.989120 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.989142 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2810ade8-67b0-4661-8df5-83da89cccddc-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.993945 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-scripts" (OuterVolumeSpecName: "scripts") pod "2810ade8-67b0-4661-8df5-83da89cccddc" (UID: "2810ade8-67b0-4661-8df5-83da89cccddc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:15 crc kubenswrapper[4978]: I0225 08:46:15.994019 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2810ade8-67b0-4661-8df5-83da89cccddc-kube-api-access-mt99b" (OuterVolumeSpecName: "kube-api-access-mt99b") pod "2810ade8-67b0-4661-8df5-83da89cccddc" (UID: "2810ade8-67b0-4661-8df5-83da89cccddc"). InnerVolumeSpecName "kube-api-access-mt99b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.024467 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2810ade8-67b0-4661-8df5-83da89cccddc" (UID: "2810ade8-67b0-4661-8df5-83da89cccddc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.059516 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-config-data" (OuterVolumeSpecName: "config-data") pod "2810ade8-67b0-4661-8df5-83da89cccddc" (UID: "2810ade8-67b0-4661-8df5-83da89cccddc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.090296 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt99b\" (UniqueName: \"kubernetes.io/projected/2810ade8-67b0-4661-8df5-83da89cccddc-kube-api-access-mt99b\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.090328 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.090337 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.090350 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2810ade8-67b0-4661-8df5-83da89cccddc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.183450 4978 generic.go:334] "Generic (PLEG): container finished" podID="2810ade8-67b0-4661-8df5-83da89cccddc" containerID="0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49" exitCode=0 Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.183500 4978 generic.go:334] "Generic (PLEG): container finished" podID="2810ade8-67b0-4661-8df5-83da89cccddc" containerID="8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856" exitCode=143 Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.183519 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.183534 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2810ade8-67b0-4661-8df5-83da89cccddc","Type":"ContainerDied","Data":"0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49"} Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.183571 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2810ade8-67b0-4661-8df5-83da89cccddc","Type":"ContainerDied","Data":"8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856"} Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.183582 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2810ade8-67b0-4661-8df5-83da89cccddc","Type":"ContainerDied","Data":"97822b5ff628565a2b7a9cf235339a1729fb77d1c7bc3c7a60de0c5c2beff4c8"} Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.183607 4978 scope.go:117] "RemoveContainer" containerID="0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.226838 4978 scope.go:117] "RemoveContainer" containerID="8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.227343 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.241389 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.250613 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:46:16 crc kubenswrapper[4978]: E0225 08:46:16.250997 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2810ade8-67b0-4661-8df5-83da89cccddc" containerName="glance-log" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.251013 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2810ade8-67b0-4661-8df5-83da89cccddc" containerName="glance-log" Feb 25 08:46:16 crc kubenswrapper[4978]: E0225 08:46:16.251029 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2810ade8-67b0-4661-8df5-83da89cccddc" containerName="glance-httpd" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.251035 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2810ade8-67b0-4661-8df5-83da89cccddc" containerName="glance-httpd" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.251189 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2810ade8-67b0-4661-8df5-83da89cccddc" containerName="glance-httpd" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.251205 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2810ade8-67b0-4661-8df5-83da89cccddc" containerName="glance-log" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.252551 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.254942 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.258503 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.259292 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.261545 4978 scope.go:117] "RemoveContainer" containerID="0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49" Feb 25 08:46:16 crc kubenswrapper[4978]: E0225 08:46:16.262132 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49\": container with ID starting with 0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49 not found: ID does not exist" containerID="0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.262162 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49"} err="failed to get container status \"0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49\": rpc error: code = NotFound desc = could not find container \"0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49\": container with ID starting with 0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49 not found: ID does not exist" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.262184 4978 scope.go:117] "RemoveContainer" containerID="8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856" Feb 25 08:46:16 crc kubenswrapper[4978]: E0225 08:46:16.262437 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856\": container with ID starting with 8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856 not found: ID does not exist" containerID="8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.262460 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856"} err="failed to get container status \"8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856\": rpc error: code = NotFound desc = could not find container \"8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856\": container with ID starting with 8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856 not found: ID does not exist" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.262478 4978 scope.go:117] "RemoveContainer" containerID="0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.262685 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49"} err="failed to get container status \"0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49\": rpc error: code = NotFound desc = could not find container \"0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49\": container with ID starting with 0b08e11f8fe06a1935c0bbc8f1f9f7fb2cfb09c7a6de50f401c42770f3395f49 not found: ID does not exist" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.262703 4978 scope.go:117] "RemoveContainer" containerID="8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.262942 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856"} err="failed to get container status \"8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856\": rpc error: code = NotFound desc = could not find container \"8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856\": container with ID starting with 8747be8a27a4eb5b29058d61a74433747f7a4aade3809175b5697f5eeac87856 not found: ID does not exist" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.294489 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.294526 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.294576 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.294632 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-logs\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.294700 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.295720 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r97gt\" (UniqueName: \"kubernetes.io/projected/6a38b26a-a122-4440-8887-528e8614c0e3-kube-api-access-r97gt\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.295785 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.396917 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.396970 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.397063 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.397093 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-logs\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.397185 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.397246 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r97gt\" (UniqueName: \"kubernetes.io/projected/6a38b26a-a122-4440-8887-528e8614c0e3-kube-api-access-r97gt\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.397320 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.398220 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.398264 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-logs\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.401326 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.401512 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.401951 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.402230 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.417380 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r97gt\" (UniqueName: \"kubernetes.io/projected/6a38b26a-a122-4440-8887-528e8614c0e3-kube-api-access-r97gt\") pod \"glance-default-external-api-0\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " pod="openstack/glance-default-external-api-0" Feb 25 08:46:16 crc kubenswrapper[4978]: I0225 08:46:16.613746 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.199113 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" containerName="glance-httpd" containerID="cri-o://7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342" gracePeriod=30 Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.199102 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" containerName="glance-log" containerID="cri-o://1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262" gracePeriod=30 Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.274232 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:46:17 crc kubenswrapper[4978]: W0225 08:46:17.276145 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a38b26a_a122_4440_8887_528e8614c0e3.slice/crio-cb0cc2cfd295cdb62c2a2b8e5ef9cdd1ac980280a0d12c184af89737f7306139 WatchSource:0}: Error finding container cb0cc2cfd295cdb62c2a2b8e5ef9cdd1ac980280a0d12c184af89737f7306139: Status 404 returned error can't find the container with id cb0cc2cfd295cdb62c2a2b8e5ef9cdd1ac980280a0d12c184af89737f7306139 Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.337567 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2810ade8-67b0-4661-8df5-83da89cccddc" path="/var/lib/kubelet/pods/2810ade8-67b0-4661-8df5-83da89cccddc/volumes" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.769099 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.820266 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-logs\") pod \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.820377 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vclcq\" (UniqueName: \"kubernetes.io/projected/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-kube-api-access-vclcq\") pod \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.820449 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-combined-ca-bundle\") pod \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.820485 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-scripts\") pod \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.820517 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-config-data\") pod \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.820579 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-httpd-run\") pod \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\" (UID: \"9a6112f8-f64d-4391-8b79-3b8f7f3eb555\") " Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.820784 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-logs" (OuterVolumeSpecName: "logs") pod "9a6112f8-f64d-4391-8b79-3b8f7f3eb555" (UID: "9a6112f8-f64d-4391-8b79-3b8f7f3eb555"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.820979 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9a6112f8-f64d-4391-8b79-3b8f7f3eb555" (UID: "9a6112f8-f64d-4391-8b79-3b8f7f3eb555"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.820991 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.825897 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-kube-api-access-vclcq" (OuterVolumeSpecName: "kube-api-access-vclcq") pod "9a6112f8-f64d-4391-8b79-3b8f7f3eb555" (UID: "9a6112f8-f64d-4391-8b79-3b8f7f3eb555"). InnerVolumeSpecName "kube-api-access-vclcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.829776 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-scripts" (OuterVolumeSpecName: "scripts") pod "9a6112f8-f64d-4391-8b79-3b8f7f3eb555" (UID: "9a6112f8-f64d-4391-8b79-3b8f7f3eb555"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.854208 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a6112f8-f64d-4391-8b79-3b8f7f3eb555" (UID: "9a6112f8-f64d-4391-8b79-3b8f7f3eb555"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.870009 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-config-data" (OuterVolumeSpecName: "config-data") pod "9a6112f8-f64d-4391-8b79-3b8f7f3eb555" (UID: "9a6112f8-f64d-4391-8b79-3b8f7f3eb555"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.922799 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vclcq\" (UniqueName: \"kubernetes.io/projected/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-kube-api-access-vclcq\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.922837 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.922853 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.922865 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:17 crc kubenswrapper[4978]: I0225 08:46:17.922877 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9a6112f8-f64d-4391-8b79-3b8f7f3eb555-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.214074 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" containerID="7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342" exitCode=0 Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.214579 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" containerID="1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262" exitCode=143 Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.214141 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.214166 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a6112f8-f64d-4391-8b79-3b8f7f3eb555","Type":"ContainerDied","Data":"7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342"} Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.214713 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a6112f8-f64d-4391-8b79-3b8f7f3eb555","Type":"ContainerDied","Data":"1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262"} Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.214739 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9a6112f8-f64d-4391-8b79-3b8f7f3eb555","Type":"ContainerDied","Data":"064e09c05c7ebfab75c0223db9816738c9e3d3ec56a7ebf85f92b0ac4e23f77f"} Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.214766 4978 scope.go:117] "RemoveContainer" containerID="7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.218288 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a38b26a-a122-4440-8887-528e8614c0e3","Type":"ContainerStarted","Data":"44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133"} Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.218320 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a38b26a-a122-4440-8887-528e8614c0e3","Type":"ContainerStarted","Data":"cb0cc2cfd295cdb62c2a2b8e5ef9cdd1ac980280a0d12c184af89737f7306139"} Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.249895 4978 scope.go:117] "RemoveContainer" containerID="1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.282503 4978 scope.go:117] "RemoveContainer" containerID="7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342" Feb 25 08:46:18 crc kubenswrapper[4978]: E0225 08:46:18.284991 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342\": container with ID starting with 7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342 not found: ID does not exist" containerID="7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.285029 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342"} err="failed to get container status \"7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342\": rpc error: code = NotFound desc = could not find container \"7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342\": container with ID starting with 7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342 not found: ID does not exist" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.285058 4978 scope.go:117] "RemoveContainer" containerID="1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262" Feb 25 08:46:18 crc kubenswrapper[4978]: E0225 08:46:18.286724 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262\": container with ID starting with 1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262 not found: ID does not exist" containerID="1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.286762 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262"} err="failed to get container status \"1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262\": rpc error: code = NotFound desc = could not find container \"1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262\": container with ID starting with 1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262 not found: ID does not exist" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.286805 4978 scope.go:117] "RemoveContainer" containerID="7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.287365 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.289649 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342"} err="failed to get container status \"7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342\": rpc error: code = NotFound desc = could not find container \"7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342\": container with ID starting with 7fbfc5999e98566eba7629193fa2804889a8272975860c1956d28166bbab9342 not found: ID does not exist" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.289723 4978 scope.go:117] "RemoveContainer" containerID="1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.290526 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262"} err="failed to get container status \"1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262\": rpc error: code = NotFound desc = could not find container \"1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262\": container with ID starting with 1c2d39ed8039df0609e28af35c4451ecbb90b5bc72fec58cdfca303bb454d262 not found: ID does not exist" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.299353 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.316506 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:46:18 crc kubenswrapper[4978]: E0225 08:46:18.316975 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" containerName="glance-httpd" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.317001 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" containerName="glance-httpd" Feb 25 08:46:18 crc kubenswrapper[4978]: E0225 08:46:18.317034 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" containerName="glance-log" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.317043 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" containerName="glance-log" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.317232 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" containerName="glance-log" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.317261 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" containerName="glance-httpd" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.318420 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.321023 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.321986 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.328524 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.329340 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.329390 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.329442 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.329470 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.329514 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.329582 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.329616 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fjd4\" (UniqueName: \"kubernetes.io/projected/6e0da9b3-9cc8-475d-8f9c-f5d228575271-kube-api-access-2fjd4\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.431866 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.431917 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.431977 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.432010 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.432048 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.432174 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.432221 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fjd4\" (UniqueName: \"kubernetes.io/projected/6e0da9b3-9cc8-475d-8f9c-f5d228575271-kube-api-access-2fjd4\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.432759 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.433199 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.439271 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.440481 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.441436 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.441870 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.450888 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fjd4\" (UniqueName: \"kubernetes.io/projected/6e0da9b3-9cc8-475d-8f9c-f5d228575271-kube-api-access-2fjd4\") pod \"glance-default-internal-api-0\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:46:18 crc kubenswrapper[4978]: I0225 08:46:18.634830 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:19 crc kubenswrapper[4978]: I0225 08:46:19.165974 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:46:19 crc kubenswrapper[4978]: I0225 08:46:19.235667 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e0da9b3-9cc8-475d-8f9c-f5d228575271","Type":"ContainerStarted","Data":"35ea630b460af7b9bc53081a57206072b5392b5d571b573b07241fa8ad7b5aa9"} Feb 25 08:46:19 crc kubenswrapper[4978]: I0225 08:46:19.239844 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a38b26a-a122-4440-8887-528e8614c0e3","Type":"ContainerStarted","Data":"6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c"} Feb 25 08:46:19 crc kubenswrapper[4978]: I0225 08:46:19.270449 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.270425482 podStartE2EDuration="3.270425482s" podCreationTimestamp="2026-02-25 08:46:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:46:19.2606638 +0000 UTC m=+7272.699920279" watchObservedRunningTime="2026-02-25 08:46:19.270425482 +0000 UTC m=+7272.709681971" Feb 25 08:46:19 crc kubenswrapper[4978]: I0225 08:46:19.339642 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a6112f8-f64d-4391-8b79-3b8f7f3eb555" path="/var/lib/kubelet/pods/9a6112f8-f64d-4391-8b79-3b8f7f3eb555/volumes" Feb 25 08:46:20 crc kubenswrapper[4978]: I0225 08:46:20.253516 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e0da9b3-9cc8-475d-8f9c-f5d228575271","Type":"ContainerStarted","Data":"73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5"} Feb 25 08:46:21 crc kubenswrapper[4978]: I0225 08:46:21.264796 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e0da9b3-9cc8-475d-8f9c-f5d228575271","Type":"ContainerStarted","Data":"d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42"} Feb 25 08:46:21 crc kubenswrapper[4978]: I0225 08:46:21.303013 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.302996103 podStartE2EDuration="3.302996103s" podCreationTimestamp="2026-02-25 08:46:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:46:21.290553808 +0000 UTC m=+7274.729810277" watchObservedRunningTime="2026-02-25 08:46:21.302996103 +0000 UTC m=+7274.742252572" Feb 25 08:46:22 crc kubenswrapper[4978]: I0225 08:46:22.909677 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.034412 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cbfc78d5f-7rc2t"] Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.034732 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" podUID="e04d46cc-c0c6-4157-8701-c3b18996ade3" containerName="dnsmasq-dns" containerID="cri-o://48b47601de0639b6ff473276bc6c39289033f3d6d6411fc14103fe670738b562" gracePeriod=10 Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.289530 4978 generic.go:334] "Generic (PLEG): container finished" podID="e04d46cc-c0c6-4157-8701-c3b18996ade3" containerID="48b47601de0639b6ff473276bc6c39289033f3d6d6411fc14103fe670738b562" exitCode=0 Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.289597 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" event={"ID":"e04d46cc-c0c6-4157-8701-c3b18996ade3","Type":"ContainerDied","Data":"48b47601de0639b6ff473276bc6c39289033f3d6d6411fc14103fe670738b562"} Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.485570 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.534987 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-nb\") pod \"e04d46cc-c0c6-4157-8701-c3b18996ade3\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.535130 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-config\") pod \"e04d46cc-c0c6-4157-8701-c3b18996ade3\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.535172 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-sb\") pod \"e04d46cc-c0c6-4157-8701-c3b18996ade3\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.535221 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-dns-svc\") pod \"e04d46cc-c0c6-4157-8701-c3b18996ade3\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.535418 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txlw2\" (UniqueName: \"kubernetes.io/projected/e04d46cc-c0c6-4157-8701-c3b18996ade3-kube-api-access-txlw2\") pod \"e04d46cc-c0c6-4157-8701-c3b18996ade3\" (UID: \"e04d46cc-c0c6-4157-8701-c3b18996ade3\") " Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.547491 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e04d46cc-c0c6-4157-8701-c3b18996ade3-kube-api-access-txlw2" (OuterVolumeSpecName: "kube-api-access-txlw2") pod "e04d46cc-c0c6-4157-8701-c3b18996ade3" (UID: "e04d46cc-c0c6-4157-8701-c3b18996ade3"). InnerVolumeSpecName "kube-api-access-txlw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.585944 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-config" (OuterVolumeSpecName: "config") pod "e04d46cc-c0c6-4157-8701-c3b18996ade3" (UID: "e04d46cc-c0c6-4157-8701-c3b18996ade3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.587865 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e04d46cc-c0c6-4157-8701-c3b18996ade3" (UID: "e04d46cc-c0c6-4157-8701-c3b18996ade3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.593465 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e04d46cc-c0c6-4157-8701-c3b18996ade3" (UID: "e04d46cc-c0c6-4157-8701-c3b18996ade3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.605269 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e04d46cc-c0c6-4157-8701-c3b18996ade3" (UID: "e04d46cc-c0c6-4157-8701-c3b18996ade3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.637158 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txlw2\" (UniqueName: \"kubernetes.io/projected/e04d46cc-c0c6-4157-8701-c3b18996ade3-kube-api-access-txlw2\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.637192 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.637206 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.637219 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:23 crc kubenswrapper[4978]: I0225 08:46:23.637233 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e04d46cc-c0c6-4157-8701-c3b18996ade3-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:24 crc kubenswrapper[4978]: I0225 08:46:24.302126 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" event={"ID":"e04d46cc-c0c6-4157-8701-c3b18996ade3","Type":"ContainerDied","Data":"005dccfcdd5e53f873a2320d2f82622388c8442b341d413fb2e5b7d62a077e9f"} Feb 25 08:46:24 crc kubenswrapper[4978]: I0225 08:46:24.302191 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cbfc78d5f-7rc2t" Feb 25 08:46:24 crc kubenswrapper[4978]: I0225 08:46:24.302506 4978 scope.go:117] "RemoveContainer" containerID="48b47601de0639b6ff473276bc6c39289033f3d6d6411fc14103fe670738b562" Feb 25 08:46:24 crc kubenswrapper[4978]: I0225 08:46:24.330007 4978 scope.go:117] "RemoveContainer" containerID="71d48a7c2e9a1e190bdf3bf1a6d5eecaa5bc940f175870eecb06001dfabe7eb1" Feb 25 08:46:24 crc kubenswrapper[4978]: I0225 08:46:24.361723 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cbfc78d5f-7rc2t"] Feb 25 08:46:24 crc kubenswrapper[4978]: I0225 08:46:24.379491 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cbfc78d5f-7rc2t"] Feb 25 08:46:25 crc kubenswrapper[4978]: I0225 08:46:25.346509 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e04d46cc-c0c6-4157-8701-c3b18996ade3" path="/var/lib/kubelet/pods/e04d46cc-c0c6-4157-8701-c3b18996ade3/volumes" Feb 25 08:46:26 crc kubenswrapper[4978]: I0225 08:46:26.614266 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 08:46:26 crc kubenswrapper[4978]: I0225 08:46:26.617251 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 08:46:26 crc kubenswrapper[4978]: I0225 08:46:26.664958 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 08:46:26 crc kubenswrapper[4978]: I0225 08:46:26.694608 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 08:46:27 crc kubenswrapper[4978]: I0225 08:46:27.351932 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 08:46:27 crc kubenswrapper[4978]: I0225 08:46:27.352288 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 08:46:28 crc kubenswrapper[4978]: I0225 08:46:28.328196 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:46:28 crc kubenswrapper[4978]: E0225 08:46:28.328706 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:46:28 crc kubenswrapper[4978]: I0225 08:46:28.503663 4978 scope.go:117] "RemoveContainer" containerID="e89fc75a2be6143a30a68d5f6e274dc12ab6555b340c3ef7e9f74887319d1034" Feb 25 08:46:28 crc kubenswrapper[4978]: I0225 08:46:28.643345 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:28 crc kubenswrapper[4978]: I0225 08:46:28.643735 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:28 crc kubenswrapper[4978]: I0225 08:46:28.695819 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:28 crc kubenswrapper[4978]: I0225 08:46:28.702702 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:29 crc kubenswrapper[4978]: I0225 08:46:29.153220 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 08:46:29 crc kubenswrapper[4978]: I0225 08:46:29.295474 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 08:46:29 crc kubenswrapper[4978]: I0225 08:46:29.371576 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:29 crc kubenswrapper[4978]: I0225 08:46:29.371612 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:31 crc kubenswrapper[4978]: I0225 08:46:31.203480 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:31 crc kubenswrapper[4978]: I0225 08:46:31.268147 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.050690 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-smtfd"] Feb 25 08:46:38 crc kubenswrapper[4978]: E0225 08:46:38.056651 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e04d46cc-c0c6-4157-8701-c3b18996ade3" containerName="init" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.056724 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e04d46cc-c0c6-4157-8701-c3b18996ade3" containerName="init" Feb 25 08:46:38 crc kubenswrapper[4978]: E0225 08:46:38.056786 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e04d46cc-c0c6-4157-8701-c3b18996ade3" containerName="dnsmasq-dns" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.056805 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e04d46cc-c0c6-4157-8701-c3b18996ade3" containerName="dnsmasq-dns" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.057489 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e04d46cc-c0c6-4157-8701-c3b18996ade3" containerName="dnsmasq-dns" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.059048 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-smtfd" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.065125 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-smtfd"] Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.150519 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-acdf-account-create-update-wfk78"] Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.151708 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-acdf-account-create-update-wfk78" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.154432 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.161453 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-acdf-account-create-update-wfk78"] Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.179567 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0cb40be-d13e-4d76-a41f-ac97096d1d44-operator-scripts\") pod \"placement-db-create-smtfd\" (UID: \"e0cb40be-d13e-4d76-a41f-ac97096d1d44\") " pod="openstack/placement-db-create-smtfd" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.179661 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jmj9\" (UniqueName: \"kubernetes.io/projected/e0cb40be-d13e-4d76-a41f-ac97096d1d44-kube-api-access-2jmj9\") pod \"placement-db-create-smtfd\" (UID: \"e0cb40be-d13e-4d76-a41f-ac97096d1d44\") " pod="openstack/placement-db-create-smtfd" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.281699 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0cb40be-d13e-4d76-a41f-ac97096d1d44-operator-scripts\") pod \"placement-db-create-smtfd\" (UID: \"e0cb40be-d13e-4d76-a41f-ac97096d1d44\") " pod="openstack/placement-db-create-smtfd" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.282079 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-operator-scripts\") pod \"placement-acdf-account-create-update-wfk78\" (UID: \"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f\") " pod="openstack/placement-acdf-account-create-update-wfk78" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.282203 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jmj9\" (UniqueName: \"kubernetes.io/projected/e0cb40be-d13e-4d76-a41f-ac97096d1d44-kube-api-access-2jmj9\") pod \"placement-db-create-smtfd\" (UID: \"e0cb40be-d13e-4d76-a41f-ac97096d1d44\") " pod="openstack/placement-db-create-smtfd" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.282393 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4z5c\" (UniqueName: \"kubernetes.io/projected/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-kube-api-access-x4z5c\") pod \"placement-acdf-account-create-update-wfk78\" (UID: \"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f\") " pod="openstack/placement-acdf-account-create-update-wfk78" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.283088 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0cb40be-d13e-4d76-a41f-ac97096d1d44-operator-scripts\") pod \"placement-db-create-smtfd\" (UID: \"e0cb40be-d13e-4d76-a41f-ac97096d1d44\") " pod="openstack/placement-db-create-smtfd" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.309091 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jmj9\" (UniqueName: \"kubernetes.io/projected/e0cb40be-d13e-4d76-a41f-ac97096d1d44-kube-api-access-2jmj9\") pod \"placement-db-create-smtfd\" (UID: \"e0cb40be-d13e-4d76-a41f-ac97096d1d44\") " pod="openstack/placement-db-create-smtfd" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.384431 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-operator-scripts\") pod \"placement-acdf-account-create-update-wfk78\" (UID: \"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f\") " pod="openstack/placement-acdf-account-create-update-wfk78" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.384617 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4z5c\" (UniqueName: \"kubernetes.io/projected/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-kube-api-access-x4z5c\") pod \"placement-acdf-account-create-update-wfk78\" (UID: \"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f\") " pod="openstack/placement-acdf-account-create-update-wfk78" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.385332 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-operator-scripts\") pod \"placement-acdf-account-create-update-wfk78\" (UID: \"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f\") " pod="openstack/placement-acdf-account-create-update-wfk78" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.388841 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-smtfd" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.415044 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4z5c\" (UniqueName: \"kubernetes.io/projected/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-kube-api-access-x4z5c\") pod \"placement-acdf-account-create-update-wfk78\" (UID: \"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f\") " pod="openstack/placement-acdf-account-create-update-wfk78" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.467487 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-acdf-account-create-update-wfk78" Feb 25 08:46:38 crc kubenswrapper[4978]: I0225 08:46:38.905725 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-smtfd"] Feb 25 08:46:39 crc kubenswrapper[4978]: W0225 08:46:39.041099 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7c760da_bbaf_4b92_8b90_5b1b99ed9d7f.slice/crio-8e507d4383eb69b9024856a1c62c078b1c0f62fa8793d32cef8b53e6a88d7992 WatchSource:0}: Error finding container 8e507d4383eb69b9024856a1c62c078b1c0f62fa8793d32cef8b53e6a88d7992: Status 404 returned error can't find the container with id 8e507d4383eb69b9024856a1c62c078b1c0f62fa8793d32cef8b53e6a88d7992 Feb 25 08:46:39 crc kubenswrapper[4978]: I0225 08:46:39.050791 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-acdf-account-create-update-wfk78"] Feb 25 08:46:39 crc kubenswrapper[4978]: I0225 08:46:39.466337 4978 generic.go:334] "Generic (PLEG): container finished" podID="e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f" containerID="389ff171ff2aab135beef50dd73fbd983bd61c38428fc1c39575fdcc7c58769b" exitCode=0 Feb 25 08:46:39 crc kubenswrapper[4978]: I0225 08:46:39.466437 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-acdf-account-create-update-wfk78" event={"ID":"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f","Type":"ContainerDied","Data":"389ff171ff2aab135beef50dd73fbd983bd61c38428fc1c39575fdcc7c58769b"} Feb 25 08:46:39 crc kubenswrapper[4978]: I0225 08:46:39.466722 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-acdf-account-create-update-wfk78" event={"ID":"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f","Type":"ContainerStarted","Data":"8e507d4383eb69b9024856a1c62c078b1c0f62fa8793d32cef8b53e6a88d7992"} Feb 25 08:46:39 crc kubenswrapper[4978]: I0225 08:46:39.469153 4978 generic.go:334] "Generic (PLEG): container finished" podID="e0cb40be-d13e-4d76-a41f-ac97096d1d44" containerID="3a0280a607543b4588fc74f488860c3ac84a29d25c45ef37d747296342df7091" exitCode=0 Feb 25 08:46:39 crc kubenswrapper[4978]: I0225 08:46:39.469217 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-smtfd" event={"ID":"e0cb40be-d13e-4d76-a41f-ac97096d1d44","Type":"ContainerDied","Data":"3a0280a607543b4588fc74f488860c3ac84a29d25c45ef37d747296342df7091"} Feb 25 08:46:39 crc kubenswrapper[4978]: I0225 08:46:39.469293 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-smtfd" event={"ID":"e0cb40be-d13e-4d76-a41f-ac97096d1d44","Type":"ContainerStarted","Data":"bb6371f46cff23a9bda8a8c5a0b96d693b25f771319e61f180b40bf27d70a2d5"} Feb 25 08:46:40 crc kubenswrapper[4978]: I0225 08:46:40.934559 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-smtfd" Feb 25 08:46:40 crc kubenswrapper[4978]: I0225 08:46:40.939384 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-acdf-account-create-update-wfk78" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.037174 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0cb40be-d13e-4d76-a41f-ac97096d1d44-operator-scripts\") pod \"e0cb40be-d13e-4d76-a41f-ac97096d1d44\" (UID: \"e0cb40be-d13e-4d76-a41f-ac97096d1d44\") " Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.037395 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jmj9\" (UniqueName: \"kubernetes.io/projected/e0cb40be-d13e-4d76-a41f-ac97096d1d44-kube-api-access-2jmj9\") pod \"e0cb40be-d13e-4d76-a41f-ac97096d1d44\" (UID: \"e0cb40be-d13e-4d76-a41f-ac97096d1d44\") " Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.037488 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-operator-scripts\") pod \"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f\" (UID: \"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f\") " Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.037512 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4z5c\" (UniqueName: \"kubernetes.io/projected/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-kube-api-access-x4z5c\") pod \"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f\" (UID: \"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f\") " Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.038136 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0cb40be-d13e-4d76-a41f-ac97096d1d44-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e0cb40be-d13e-4d76-a41f-ac97096d1d44" (UID: "e0cb40be-d13e-4d76-a41f-ac97096d1d44"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.038141 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f" (UID: "e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.043606 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-kube-api-access-x4z5c" (OuterVolumeSpecName: "kube-api-access-x4z5c") pod "e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f" (UID: "e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f"). InnerVolumeSpecName "kube-api-access-x4z5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.044313 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0cb40be-d13e-4d76-a41f-ac97096d1d44-kube-api-access-2jmj9" (OuterVolumeSpecName: "kube-api-access-2jmj9") pod "e0cb40be-d13e-4d76-a41f-ac97096d1d44" (UID: "e0cb40be-d13e-4d76-a41f-ac97096d1d44"). InnerVolumeSpecName "kube-api-access-2jmj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.139299 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.139338 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4z5c\" (UniqueName: \"kubernetes.io/projected/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f-kube-api-access-x4z5c\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.139356 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0cb40be-d13e-4d76-a41f-ac97096d1d44-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.139389 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jmj9\" (UniqueName: \"kubernetes.io/projected/e0cb40be-d13e-4d76-a41f-ac97096d1d44-kube-api-access-2jmj9\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.498004 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-acdf-account-create-update-wfk78" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.497992 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-acdf-account-create-update-wfk78" event={"ID":"e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f","Type":"ContainerDied","Data":"8e507d4383eb69b9024856a1c62c078b1c0f62fa8793d32cef8b53e6a88d7992"} Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.498441 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e507d4383eb69b9024856a1c62c078b1c0f62fa8793d32cef8b53e6a88d7992" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.501031 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-smtfd" event={"ID":"e0cb40be-d13e-4d76-a41f-ac97096d1d44","Type":"ContainerDied","Data":"bb6371f46cff23a9bda8a8c5a0b96d693b25f771319e61f180b40bf27d70a2d5"} Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.501078 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb6371f46cff23a9bda8a8c5a0b96d693b25f771319e61f180b40bf27d70a2d5" Feb 25 08:46:41 crc kubenswrapper[4978]: I0225 08:46:41.501091 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-smtfd" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.328697 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:46:43 crc kubenswrapper[4978]: E0225 08:46:43.329130 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.567659 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-778bcd845c-fw9q9"] Feb 25 08:46:43 crc kubenswrapper[4978]: E0225 08:46:43.568093 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cb40be-d13e-4d76-a41f-ac97096d1d44" containerName="mariadb-database-create" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.568109 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cb40be-d13e-4d76-a41f-ac97096d1d44" containerName="mariadb-database-create" Feb 25 08:46:43 crc kubenswrapper[4978]: E0225 08:46:43.568136 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f" containerName="mariadb-account-create-update" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.568143 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f" containerName="mariadb-account-create-update" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.568318 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f" containerName="mariadb-account-create-update" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.568340 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0cb40be-d13e-4d76-a41f-ac97096d1d44" containerName="mariadb-database-create" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.569266 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.584965 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-dhmxq"] Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.590311 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.593703 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.594236 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-2xtzl" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.594395 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.619730 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dhmxq"] Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.636446 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-778bcd845c-fw9q9"] Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.694237 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztjf6\" (UniqueName: \"kubernetes.io/projected/b985867e-5709-4c7d-8c1d-facadb04db8a-kube-api-access-ztjf6\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.694286 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-nb\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.694315 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d009f647-5ba2-4877-b775-47ecd94b3e6f-logs\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.694393 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-config\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.694455 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc4rn\" (UniqueName: \"kubernetes.io/projected/d009f647-5ba2-4877-b775-47ecd94b3e6f-kube-api-access-mc4rn\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.694545 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-combined-ca-bundle\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.694583 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-scripts\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.694789 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-dns-svc\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.694894 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-sb\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.694924 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-config-data\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.798477 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-dns-svc\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.799870 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-dns-svc\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.799936 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-sb\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.799978 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-config-data\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.800848 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-sb\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.801320 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztjf6\" (UniqueName: \"kubernetes.io/projected/b985867e-5709-4c7d-8c1d-facadb04db8a-kube-api-access-ztjf6\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.801379 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-nb\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.801426 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d009f647-5ba2-4877-b775-47ecd94b3e6f-logs\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.801557 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-config\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.801710 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc4rn\" (UniqueName: \"kubernetes.io/projected/d009f647-5ba2-4877-b775-47ecd94b3e6f-kube-api-access-mc4rn\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.801752 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-combined-ca-bundle\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.801815 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-scripts\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.802090 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d009f647-5ba2-4877-b775-47ecd94b3e6f-logs\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.802189 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-nb\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.802378 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-config\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.809944 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-combined-ca-bundle\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.810725 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-scripts\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.810795 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-config-data\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.819297 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc4rn\" (UniqueName: \"kubernetes.io/projected/d009f647-5ba2-4877-b775-47ecd94b3e6f-kube-api-access-mc4rn\") pod \"placement-db-sync-dhmxq\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.822017 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztjf6\" (UniqueName: \"kubernetes.io/projected/b985867e-5709-4c7d-8c1d-facadb04db8a-kube-api-access-ztjf6\") pod \"dnsmasq-dns-778bcd845c-fw9q9\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.899204 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:43 crc kubenswrapper[4978]: I0225 08:46:43.942962 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:44 crc kubenswrapper[4978]: W0225 08:46:44.410588 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb985867e_5709_4c7d_8c1d_facadb04db8a.slice/crio-7972d043d2ab1afc98d40f405b1d7f910b8cb5a143785807a462a7fc96383b02 WatchSource:0}: Error finding container 7972d043d2ab1afc98d40f405b1d7f910b8cb5a143785807a462a7fc96383b02: Status 404 returned error can't find the container with id 7972d043d2ab1afc98d40f405b1d7f910b8cb5a143785807a462a7fc96383b02 Feb 25 08:46:44 crc kubenswrapper[4978]: I0225 08:46:44.413504 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-778bcd845c-fw9q9"] Feb 25 08:46:44 crc kubenswrapper[4978]: I0225 08:46:44.483636 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dhmxq"] Feb 25 08:46:44 crc kubenswrapper[4978]: W0225 08:46:44.486617 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd009f647_5ba2_4877_b775_47ecd94b3e6f.slice/crio-dc025267b8dcf3ddf5c260e2f9f3438e74cc5e4ff10e3b33ce2bdc718058d345 WatchSource:0}: Error finding container dc025267b8dcf3ddf5c260e2f9f3438e74cc5e4ff10e3b33ce2bdc718058d345: Status 404 returned error can't find the container with id dc025267b8dcf3ddf5c260e2f9f3438e74cc5e4ff10e3b33ce2bdc718058d345 Feb 25 08:46:44 crc kubenswrapper[4978]: I0225 08:46:44.551217 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dhmxq" event={"ID":"d009f647-5ba2-4877-b775-47ecd94b3e6f","Type":"ContainerStarted","Data":"dc025267b8dcf3ddf5c260e2f9f3438e74cc5e4ff10e3b33ce2bdc718058d345"} Feb 25 08:46:44 crc kubenswrapper[4978]: I0225 08:46:44.552391 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" event={"ID":"b985867e-5709-4c7d-8c1d-facadb04db8a","Type":"ContainerStarted","Data":"7972d043d2ab1afc98d40f405b1d7f910b8cb5a143785807a462a7fc96383b02"} Feb 25 08:46:45 crc kubenswrapper[4978]: I0225 08:46:45.562664 4978 generic.go:334] "Generic (PLEG): container finished" podID="b985867e-5709-4c7d-8c1d-facadb04db8a" containerID="13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4" exitCode=0 Feb 25 08:46:45 crc kubenswrapper[4978]: I0225 08:46:45.562852 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" event={"ID":"b985867e-5709-4c7d-8c1d-facadb04db8a","Type":"ContainerDied","Data":"13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4"} Feb 25 08:46:46 crc kubenswrapper[4978]: I0225 08:46:46.588260 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" event={"ID":"b985867e-5709-4c7d-8c1d-facadb04db8a","Type":"ContainerStarted","Data":"5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734"} Feb 25 08:46:46 crc kubenswrapper[4978]: I0225 08:46:46.589775 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:46 crc kubenswrapper[4978]: I0225 08:46:46.626036 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" podStartSLOduration=3.626015003 podStartE2EDuration="3.626015003s" podCreationTimestamp="2026-02-25 08:46:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:46:46.617381196 +0000 UTC m=+7300.056637655" watchObservedRunningTime="2026-02-25 08:46:46.626015003 +0000 UTC m=+7300.065271462" Feb 25 08:46:48 crc kubenswrapper[4978]: I0225 08:46:48.607271 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dhmxq" event={"ID":"d009f647-5ba2-4877-b775-47ecd94b3e6f","Type":"ContainerStarted","Data":"faa24b9256a8f0a12e21502a1ca50a38220d731f994ae9b0d8a028a6765261de"} Feb 25 08:46:48 crc kubenswrapper[4978]: I0225 08:46:48.639425 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-dhmxq" podStartSLOduration=2.164568946 podStartE2EDuration="5.63939991s" podCreationTimestamp="2026-02-25 08:46:43 +0000 UTC" firstStartedPulling="2026-02-25 08:46:44.488641045 +0000 UTC m=+7297.927897504" lastFinishedPulling="2026-02-25 08:46:47.963471989 +0000 UTC m=+7301.402728468" observedRunningTime="2026-02-25 08:46:48.628708719 +0000 UTC m=+7302.067965208" watchObservedRunningTime="2026-02-25 08:46:48.63939991 +0000 UTC m=+7302.078656409" Feb 25 08:46:49 crc kubenswrapper[4978]: I0225 08:46:49.620325 4978 generic.go:334] "Generic (PLEG): container finished" podID="d009f647-5ba2-4877-b775-47ecd94b3e6f" containerID="faa24b9256a8f0a12e21502a1ca50a38220d731f994ae9b0d8a028a6765261de" exitCode=0 Feb 25 08:46:49 crc kubenswrapper[4978]: I0225 08:46:49.620447 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dhmxq" event={"ID":"d009f647-5ba2-4877-b775-47ecd94b3e6f","Type":"ContainerDied","Data":"faa24b9256a8f0a12e21502a1ca50a38220d731f994ae9b0d8a028a6765261de"} Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.102578 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.130880 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-scripts\") pod \"d009f647-5ba2-4877-b775-47ecd94b3e6f\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.130956 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-config-data\") pod \"d009f647-5ba2-4877-b775-47ecd94b3e6f\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.130986 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc4rn\" (UniqueName: \"kubernetes.io/projected/d009f647-5ba2-4877-b775-47ecd94b3e6f-kube-api-access-mc4rn\") pod \"d009f647-5ba2-4877-b775-47ecd94b3e6f\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.131015 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-combined-ca-bundle\") pod \"d009f647-5ba2-4877-b775-47ecd94b3e6f\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.131052 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d009f647-5ba2-4877-b775-47ecd94b3e6f-logs\") pod \"d009f647-5ba2-4877-b775-47ecd94b3e6f\" (UID: \"d009f647-5ba2-4877-b775-47ecd94b3e6f\") " Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.131946 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d009f647-5ba2-4877-b775-47ecd94b3e6f-logs" (OuterVolumeSpecName: "logs") pod "d009f647-5ba2-4877-b775-47ecd94b3e6f" (UID: "d009f647-5ba2-4877-b775-47ecd94b3e6f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.139755 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-scripts" (OuterVolumeSpecName: "scripts") pod "d009f647-5ba2-4877-b775-47ecd94b3e6f" (UID: "d009f647-5ba2-4877-b775-47ecd94b3e6f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.141673 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d009f647-5ba2-4877-b775-47ecd94b3e6f-kube-api-access-mc4rn" (OuterVolumeSpecName: "kube-api-access-mc4rn") pod "d009f647-5ba2-4877-b775-47ecd94b3e6f" (UID: "d009f647-5ba2-4877-b775-47ecd94b3e6f"). InnerVolumeSpecName "kube-api-access-mc4rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.195908 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d009f647-5ba2-4877-b775-47ecd94b3e6f" (UID: "d009f647-5ba2-4877-b775-47ecd94b3e6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.201641 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-config-data" (OuterVolumeSpecName: "config-data") pod "d009f647-5ba2-4877-b775-47ecd94b3e6f" (UID: "d009f647-5ba2-4877-b775-47ecd94b3e6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.234995 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc4rn\" (UniqueName: \"kubernetes.io/projected/d009f647-5ba2-4877-b775-47ecd94b3e6f-kube-api-access-mc4rn\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.235214 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.235345 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d009f647-5ba2-4877-b775-47ecd94b3e6f-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.235497 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.235642 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d009f647-5ba2-4877-b775-47ecd94b3e6f-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.649745 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dhmxq" event={"ID":"d009f647-5ba2-4877-b775-47ecd94b3e6f","Type":"ContainerDied","Data":"dc025267b8dcf3ddf5c260e2f9f3438e74cc5e4ff10e3b33ce2bdc718058d345"} Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.650195 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc025267b8dcf3ddf5c260e2f9f3438e74cc5e4ff10e3b33ce2bdc718058d345" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.649828 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dhmxq" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.768672 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-894b87bcd-94p22"] Feb 25 08:46:51 crc kubenswrapper[4978]: E0225 08:46:51.769275 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d009f647-5ba2-4877-b775-47ecd94b3e6f" containerName="placement-db-sync" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.769306 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d009f647-5ba2-4877-b775-47ecd94b3e6f" containerName="placement-db-sync" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.769713 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d009f647-5ba2-4877-b775-47ecd94b3e6f" containerName="placement-db-sync" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.771402 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.775408 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.775883 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-2xtzl" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.776238 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.776600 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.777077 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.807264 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-894b87bcd-94p22"] Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.847538 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-scripts\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.847601 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-config-data\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.847643 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-public-tls-certs\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.847718 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-internal-tls-certs\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.847757 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-combined-ca-bundle\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.847781 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlzzz\" (UniqueName: \"kubernetes.io/projected/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-kube-api-access-qlzzz\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.847808 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-logs\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.949654 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-scripts\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.950286 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-config-data\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.950323 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-public-tls-certs\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.950392 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-internal-tls-certs\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.950420 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-combined-ca-bundle\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.950440 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlzzz\" (UniqueName: \"kubernetes.io/projected/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-kube-api-access-qlzzz\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.950459 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-logs\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.950830 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-logs\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.953027 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-scripts\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.954245 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-combined-ca-bundle\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.954747 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-public-tls-certs\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.954891 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-internal-tls-certs\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.956778 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-config-data\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:51 crc kubenswrapper[4978]: I0225 08:46:51.972827 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlzzz\" (UniqueName: \"kubernetes.io/projected/d17346e6-d110-4b4e-bdc2-afd1eeea0b6e-kube-api-access-qlzzz\") pod \"placement-894b87bcd-94p22\" (UID: \"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e\") " pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:52 crc kubenswrapper[4978]: I0225 08:46:52.121829 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:52 crc kubenswrapper[4978]: W0225 08:46:52.612253 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd17346e6_d110_4b4e_bdc2_afd1eeea0b6e.slice/crio-cdc7d37a25c1c02ede7246eb86e1058b49eadc6829cff3ae43fd342e3f1b56fd WatchSource:0}: Error finding container cdc7d37a25c1c02ede7246eb86e1058b49eadc6829cff3ae43fd342e3f1b56fd: Status 404 returned error can't find the container with id cdc7d37a25c1c02ede7246eb86e1058b49eadc6829cff3ae43fd342e3f1b56fd Feb 25 08:46:52 crc kubenswrapper[4978]: I0225 08:46:52.614431 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-894b87bcd-94p22"] Feb 25 08:46:52 crc kubenswrapper[4978]: I0225 08:46:52.660430 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-894b87bcd-94p22" event={"ID":"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e","Type":"ContainerStarted","Data":"cdc7d37a25c1c02ede7246eb86e1058b49eadc6829cff3ae43fd342e3f1b56fd"} Feb 25 08:46:53 crc kubenswrapper[4978]: I0225 08:46:53.675408 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-894b87bcd-94p22" event={"ID":"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e","Type":"ContainerStarted","Data":"f64aa368000473072b12471696a81d4f542da4299b43ac4721337f385f20013f"} Feb 25 08:46:53 crc kubenswrapper[4978]: I0225 08:46:53.675836 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-894b87bcd-94p22" event={"ID":"d17346e6-d110-4b4e-bdc2-afd1eeea0b6e","Type":"ContainerStarted","Data":"8610c9572b062f83ad6fc2877c01b9321f2a307e165c15b1c495da0183c155ae"} Feb 25 08:46:53 crc kubenswrapper[4978]: I0225 08:46:53.675872 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:53 crc kubenswrapper[4978]: I0225 08:46:53.675921 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-894b87bcd-94p22" Feb 25 08:46:53 crc kubenswrapper[4978]: I0225 08:46:53.708773 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-894b87bcd-94p22" podStartSLOduration=2.708738142 podStartE2EDuration="2.708738142s" podCreationTimestamp="2026-02-25 08:46:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:46:53.708242167 +0000 UTC m=+7307.147498686" watchObservedRunningTime="2026-02-25 08:46:53.708738142 +0000 UTC m=+7307.147994671" Feb 25 08:46:53 crc kubenswrapper[4978]: I0225 08:46:53.901639 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:46:53 crc kubenswrapper[4978]: I0225 08:46:53.984298 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fd85b86df-857nz"] Feb 25 08:46:53 crc kubenswrapper[4978]: I0225 08:46:53.984805 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" podUID="61d85cc4-e12f-44e9-b4a4-c2030412451d" containerName="dnsmasq-dns" containerID="cri-o://d23247fee0326055409949cfa5eca34a48f4f213a8438dcf20c1956647325a5a" gracePeriod=10 Feb 25 08:46:54 crc kubenswrapper[4978]: E0225 08:46:54.165621 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61d85cc4_e12f_44e9_b4a4_c2030412451d.slice/crio-d23247fee0326055409949cfa5eca34a48f4f213a8438dcf20c1956647325a5a.scope\": RecentStats: unable to find data in memory cache]" Feb 25 08:46:54 crc kubenswrapper[4978]: I0225 08:46:54.688290 4978 generic.go:334] "Generic (PLEG): container finished" podID="61d85cc4-e12f-44e9-b4a4-c2030412451d" containerID="d23247fee0326055409949cfa5eca34a48f4f213a8438dcf20c1956647325a5a" exitCode=0 Feb 25 08:46:54 crc kubenswrapper[4978]: I0225 08:46:54.688363 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" event={"ID":"61d85cc4-e12f-44e9-b4a4-c2030412451d","Type":"ContainerDied","Data":"d23247fee0326055409949cfa5eca34a48f4f213a8438dcf20c1956647325a5a"} Feb 25 08:46:54 crc kubenswrapper[4978]: I0225 08:46:54.988190 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.108224 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-config\") pod \"61d85cc4-e12f-44e9-b4a4-c2030412451d\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.108305 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9h2b\" (UniqueName: \"kubernetes.io/projected/61d85cc4-e12f-44e9-b4a4-c2030412451d-kube-api-access-m9h2b\") pod \"61d85cc4-e12f-44e9-b4a4-c2030412451d\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.108347 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-sb\") pod \"61d85cc4-e12f-44e9-b4a4-c2030412451d\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.108394 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-dns-svc\") pod \"61d85cc4-e12f-44e9-b4a4-c2030412451d\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.108428 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-nb\") pod \"61d85cc4-e12f-44e9-b4a4-c2030412451d\" (UID: \"61d85cc4-e12f-44e9-b4a4-c2030412451d\") " Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.113635 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61d85cc4-e12f-44e9-b4a4-c2030412451d-kube-api-access-m9h2b" (OuterVolumeSpecName: "kube-api-access-m9h2b") pod "61d85cc4-e12f-44e9-b4a4-c2030412451d" (UID: "61d85cc4-e12f-44e9-b4a4-c2030412451d"). InnerVolumeSpecName "kube-api-access-m9h2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.150066 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "61d85cc4-e12f-44e9-b4a4-c2030412451d" (UID: "61d85cc4-e12f-44e9-b4a4-c2030412451d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.151701 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "61d85cc4-e12f-44e9-b4a4-c2030412451d" (UID: "61d85cc4-e12f-44e9-b4a4-c2030412451d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.154895 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "61d85cc4-e12f-44e9-b4a4-c2030412451d" (UID: "61d85cc4-e12f-44e9-b4a4-c2030412451d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.156475 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-config" (OuterVolumeSpecName: "config") pod "61d85cc4-e12f-44e9-b4a4-c2030412451d" (UID: "61d85cc4-e12f-44e9-b4a4-c2030412451d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.223747 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.223807 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9h2b\" (UniqueName: \"kubernetes.io/projected/61d85cc4-e12f-44e9-b4a4-c2030412451d-kube-api-access-m9h2b\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.223820 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.223838 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.223849 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61d85cc4-e12f-44e9-b4a4-c2030412451d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.703718 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" event={"ID":"61d85cc4-e12f-44e9-b4a4-c2030412451d","Type":"ContainerDied","Data":"8994e9ea2001281f6f4caae13be2a83fda33bb35e192b806bf1e70a6956593b4"} Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.703788 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6fd85b86df-857nz" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.703857 4978 scope.go:117] "RemoveContainer" containerID="d23247fee0326055409949cfa5eca34a48f4f213a8438dcf20c1956647325a5a" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.744746 4978 scope.go:117] "RemoveContainer" containerID="8e30ab1588fa07a5bf999816f9e51784a30f27a7b84f8f5c341182a636f60fb6" Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.751634 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6fd85b86df-857nz"] Feb 25 08:46:55 crc kubenswrapper[4978]: I0225 08:46:55.762686 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6fd85b86df-857nz"] Feb 25 08:46:57 crc kubenswrapper[4978]: I0225 08:46:57.346724 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61d85cc4-e12f-44e9-b4a4-c2030412451d" path="/var/lib/kubelet/pods/61d85cc4-e12f-44e9-b4a4-c2030412451d/volumes" Feb 25 08:46:58 crc kubenswrapper[4978]: I0225 08:46:58.328877 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:46:58 crc kubenswrapper[4978]: E0225 08:46:58.329358 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:47:11 crc kubenswrapper[4978]: I0225 08:47:11.327953 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:47:11 crc kubenswrapper[4978]: E0225 08:47:11.328958 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:47:23 crc kubenswrapper[4978]: I0225 08:47:23.182938 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-894b87bcd-94p22" Feb 25 08:47:23 crc kubenswrapper[4978]: I0225 08:47:23.225459 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-894b87bcd-94p22" Feb 25 08:47:25 crc kubenswrapper[4978]: I0225 08:47:25.328506 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:47:25 crc kubenswrapper[4978]: E0225 08:47:25.329041 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.067729 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-djsgk"] Feb 25 08:47:34 crc kubenswrapper[4978]: E0225 08:47:34.068980 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d85cc4-e12f-44e9-b4a4-c2030412451d" containerName="dnsmasq-dns" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.069017 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d85cc4-e12f-44e9-b4a4-c2030412451d" containerName="dnsmasq-dns" Feb 25 08:47:34 crc kubenswrapper[4978]: E0225 08:47:34.069069 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61d85cc4-e12f-44e9-b4a4-c2030412451d" containerName="init" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.069084 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="61d85cc4-e12f-44e9-b4a4-c2030412451d" containerName="init" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.069545 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="61d85cc4-e12f-44e9-b4a4-c2030412451d" containerName="dnsmasq-dns" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.072596 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.089339 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-djsgk"] Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.101953 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65nzp\" (UniqueName: \"kubernetes.io/projected/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-kube-api-access-65nzp\") pod \"certified-operators-djsgk\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.104094 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-utilities\") pod \"certified-operators-djsgk\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.104693 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-catalog-content\") pod \"certified-operators-djsgk\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.206589 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65nzp\" (UniqueName: \"kubernetes.io/projected/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-kube-api-access-65nzp\") pod \"certified-operators-djsgk\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.206631 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-utilities\") pod \"certified-operators-djsgk\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.206718 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-catalog-content\") pod \"certified-operators-djsgk\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.207179 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-catalog-content\") pod \"certified-operators-djsgk\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.207663 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-utilities\") pod \"certified-operators-djsgk\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.226758 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65nzp\" (UniqueName: \"kubernetes.io/projected/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-kube-api-access-65nzp\") pod \"certified-operators-djsgk\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.414805 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:34 crc kubenswrapper[4978]: I0225 08:47:34.899100 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-djsgk"] Feb 25 08:47:35 crc kubenswrapper[4978]: I0225 08:47:35.149090 4978 generic.go:334] "Generic (PLEG): container finished" podID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerID="747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1" exitCode=0 Feb 25 08:47:35 crc kubenswrapper[4978]: I0225 08:47:35.149180 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsgk" event={"ID":"5d8978a9-cd99-489e-b82d-ea3bb98eb72f","Type":"ContainerDied","Data":"747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1"} Feb 25 08:47:35 crc kubenswrapper[4978]: I0225 08:47:35.149439 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsgk" event={"ID":"5d8978a9-cd99-489e-b82d-ea3bb98eb72f","Type":"ContainerStarted","Data":"b25a4e387b58c1288ad31042b7c8e6edb74e3562b8736af4463a80add090852d"} Feb 25 08:47:35 crc kubenswrapper[4978]: E0225 08:47:35.244098 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d8978a9_cd99_489e_b82d_ea3bb98eb72f.slice/crio-747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d8978a9_cd99_489e_b82d_ea3bb98eb72f.slice/crio-conmon-747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1.scope\": RecentStats: unable to find data in memory cache]" Feb 25 08:47:37 crc kubenswrapper[4978]: I0225 08:47:37.170308 4978 generic.go:334] "Generic (PLEG): container finished" podID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerID="54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8" exitCode=0 Feb 25 08:47:37 crc kubenswrapper[4978]: I0225 08:47:37.170413 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsgk" event={"ID":"5d8978a9-cd99-489e-b82d-ea3bb98eb72f","Type":"ContainerDied","Data":"54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8"} Feb 25 08:47:37 crc kubenswrapper[4978]: I0225 08:47:37.341426 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:47:37 crc kubenswrapper[4978]: E0225 08:47:37.341795 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.184713 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsgk" event={"ID":"5d8978a9-cd99-489e-b82d-ea3bb98eb72f","Type":"ContainerStarted","Data":"a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4"} Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.213546 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-djsgk" podStartSLOduration=1.551671042 podStartE2EDuration="4.213524919s" podCreationTimestamp="2026-02-25 08:47:34 +0000 UTC" firstStartedPulling="2026-02-25 08:47:35.152048532 +0000 UTC m=+7348.591304991" lastFinishedPulling="2026-02-25 08:47:37.813902369 +0000 UTC m=+7351.253158868" observedRunningTime="2026-02-25 08:47:38.209176845 +0000 UTC m=+7351.648433354" watchObservedRunningTime="2026-02-25 08:47:38.213524919 +0000 UTC m=+7351.652781398" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.436144 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bcb8d"] Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.438525 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.447986 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bcb8d"] Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.486616 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-utilities\") pod \"redhat-marketplace-bcb8d\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.486722 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-catalog-content\") pod \"redhat-marketplace-bcb8d\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.486803 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84knk\" (UniqueName: \"kubernetes.io/projected/64897ae4-94ce-4733-a084-011ce63b8719-kube-api-access-84knk\") pod \"redhat-marketplace-bcb8d\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.588690 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-utilities\") pod \"redhat-marketplace-bcb8d\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.588758 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-catalog-content\") pod \"redhat-marketplace-bcb8d\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.588837 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84knk\" (UniqueName: \"kubernetes.io/projected/64897ae4-94ce-4733-a084-011ce63b8719-kube-api-access-84knk\") pod \"redhat-marketplace-bcb8d\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.589286 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-utilities\") pod \"redhat-marketplace-bcb8d\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.589612 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-catalog-content\") pod \"redhat-marketplace-bcb8d\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.610774 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84knk\" (UniqueName: \"kubernetes.io/projected/64897ae4-94ce-4733-a084-011ce63b8719-kube-api-access-84knk\") pod \"redhat-marketplace-bcb8d\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:38 crc kubenswrapper[4978]: I0225 08:47:38.770404 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:39 crc kubenswrapper[4978]: I0225 08:47:39.231766 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bcb8d"] Feb 25 08:47:40 crc kubenswrapper[4978]: I0225 08:47:40.203152 4978 generic.go:334] "Generic (PLEG): container finished" podID="64897ae4-94ce-4733-a084-011ce63b8719" containerID="64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0" exitCode=0 Feb 25 08:47:40 crc kubenswrapper[4978]: I0225 08:47:40.203200 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bcb8d" event={"ID":"64897ae4-94ce-4733-a084-011ce63b8719","Type":"ContainerDied","Data":"64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0"} Feb 25 08:47:40 crc kubenswrapper[4978]: I0225 08:47:40.203233 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bcb8d" event={"ID":"64897ae4-94ce-4733-a084-011ce63b8719","Type":"ContainerStarted","Data":"03021d803b60c0462796cd5a87bf5033cc482d2dd4465c016acb02cd4c9f4586"} Feb 25 08:47:42 crc kubenswrapper[4978]: I0225 08:47:42.232897 4978 generic.go:334] "Generic (PLEG): container finished" podID="64897ae4-94ce-4733-a084-011ce63b8719" containerID="243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22" exitCode=0 Feb 25 08:47:42 crc kubenswrapper[4978]: I0225 08:47:42.232998 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bcb8d" event={"ID":"64897ae4-94ce-4733-a084-011ce63b8719","Type":"ContainerDied","Data":"243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22"} Feb 25 08:47:44 crc kubenswrapper[4978]: I0225 08:47:44.258187 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bcb8d" event={"ID":"64897ae4-94ce-4733-a084-011ce63b8719","Type":"ContainerStarted","Data":"1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238"} Feb 25 08:47:44 crc kubenswrapper[4978]: I0225 08:47:44.293995 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bcb8d" podStartSLOduration=3.29608776 podStartE2EDuration="6.293975187s" podCreationTimestamp="2026-02-25 08:47:38 +0000 UTC" firstStartedPulling="2026-02-25 08:47:40.205111241 +0000 UTC m=+7353.644367700" lastFinishedPulling="2026-02-25 08:47:43.202998668 +0000 UTC m=+7356.642255127" observedRunningTime="2026-02-25 08:47:44.286787905 +0000 UTC m=+7357.726044434" watchObservedRunningTime="2026-02-25 08:47:44.293975187 +0000 UTC m=+7357.733231646" Feb 25 08:47:44 crc kubenswrapper[4978]: I0225 08:47:44.415066 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:44 crc kubenswrapper[4978]: I0225 08:47:44.415327 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:44 crc kubenswrapper[4978]: I0225 08:47:44.476893 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:45 crc kubenswrapper[4978]: I0225 08:47:45.360874 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:45 crc kubenswrapper[4978]: I0225 08:47:45.617789 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-djsgk"] Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.292413 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-djsgk" podUID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerName="registry-server" containerID="cri-o://a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4" gracePeriod=2 Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.779745 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.840289 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-ww92q"] Feb 25 08:47:47 crc kubenswrapper[4978]: E0225 08:47:47.840601 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerName="extract-content" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.840617 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerName="extract-content" Feb 25 08:47:47 crc kubenswrapper[4978]: E0225 08:47:47.840634 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerName="extract-utilities" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.840641 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerName="extract-utilities" Feb 25 08:47:47 crc kubenswrapper[4978]: E0225 08:47:47.840663 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerName="registry-server" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.840669 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerName="registry-server" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.840821 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerName="registry-server" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.841602 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ww92q" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.856788 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ww92q"] Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.883127 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-utilities\") pod \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.883192 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-catalog-content\") pod \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.883219 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65nzp\" (UniqueName: \"kubernetes.io/projected/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-kube-api-access-65nzp\") pod \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\" (UID: \"5d8978a9-cd99-489e-b82d-ea3bb98eb72f\") " Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.883975 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-utilities" (OuterVolumeSpecName: "utilities") pod "5d8978a9-cd99-489e-b82d-ea3bb98eb72f" (UID: "5d8978a9-cd99-489e-b82d-ea3bb98eb72f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.899576 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-kube-api-access-65nzp" (OuterVolumeSpecName: "kube-api-access-65nzp") pod "5d8978a9-cd99-489e-b82d-ea3bb98eb72f" (UID: "5d8978a9-cd99-489e-b82d-ea3bb98eb72f"). InnerVolumeSpecName "kube-api-access-65nzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.937497 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-d46hl"] Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.938610 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-d46hl" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.953397 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-d46hl"] Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.960270 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d8978a9-cd99-489e-b82d-ea3bb98eb72f" (UID: "5d8978a9-cd99-489e-b82d-ea3bb98eb72f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.964251 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8d1c-account-create-update-zhr8r"] Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.965436 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8d1c-account-create-update-zhr8r" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.970694 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.972161 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8d1c-account-create-update-zhr8r"] Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.988017 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2m98\" (UniqueName: \"kubernetes.io/projected/4365133f-7067-4dd7-9ef3-77d8c1312e89-kube-api-access-r2m98\") pod \"nova-api-db-create-ww92q\" (UID: \"4365133f-7067-4dd7-9ef3-77d8c1312e89\") " pod="openstack/nova-api-db-create-ww92q" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.988082 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fhd4\" (UniqueName: \"kubernetes.io/projected/4dadb5f9-c6c6-43e0-98a2-917dd5748656-kube-api-access-9fhd4\") pod \"nova-cell0-db-create-d46hl\" (UID: \"4dadb5f9-c6c6-43e0-98a2-917dd5748656\") " pod="openstack/nova-cell0-db-create-d46hl" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.988178 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4365133f-7067-4dd7-9ef3-77d8c1312e89-operator-scripts\") pod \"nova-api-db-create-ww92q\" (UID: \"4365133f-7067-4dd7-9ef3-77d8c1312e89\") " pod="openstack/nova-api-db-create-ww92q" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.988233 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q7vp\" (UniqueName: \"kubernetes.io/projected/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-kube-api-access-6q7vp\") pod \"nova-api-8d1c-account-create-update-zhr8r\" (UID: \"a7cf88a7-6d20-4909-a39a-08fbe51bd45a\") " pod="openstack/nova-api-8d1c-account-create-update-zhr8r" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.988274 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dadb5f9-c6c6-43e0-98a2-917dd5748656-operator-scripts\") pod \"nova-cell0-db-create-d46hl\" (UID: \"4dadb5f9-c6c6-43e0-98a2-917dd5748656\") " pod="openstack/nova-cell0-db-create-d46hl" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.988353 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-operator-scripts\") pod \"nova-api-8d1c-account-create-update-zhr8r\" (UID: \"a7cf88a7-6d20-4909-a39a-08fbe51bd45a\") " pod="openstack/nova-api-8d1c-account-create-update-zhr8r" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.988438 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.988460 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:47 crc kubenswrapper[4978]: I0225 08:47:47.988476 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65nzp\" (UniqueName: \"kubernetes.io/projected/5d8978a9-cd99-489e-b82d-ea3bb98eb72f-kube-api-access-65nzp\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.030116 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-r7rr7"] Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.031316 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r7rr7" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.037899 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-r7rr7"] Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.090878 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smscr\" (UniqueName: \"kubernetes.io/projected/105c1f03-d5a4-43d3-93d7-ce9682adddab-kube-api-access-smscr\") pod \"nova-cell1-db-create-r7rr7\" (UID: \"105c1f03-d5a4-43d3-93d7-ce9682adddab\") " pod="openstack/nova-cell1-db-create-r7rr7" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.091151 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/105c1f03-d5a4-43d3-93d7-ce9682adddab-operator-scripts\") pod \"nova-cell1-db-create-r7rr7\" (UID: \"105c1f03-d5a4-43d3-93d7-ce9682adddab\") " pod="openstack/nova-cell1-db-create-r7rr7" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.091181 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-operator-scripts\") pod \"nova-api-8d1c-account-create-update-zhr8r\" (UID: \"a7cf88a7-6d20-4909-a39a-08fbe51bd45a\") " pod="openstack/nova-api-8d1c-account-create-update-zhr8r" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.091203 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2m98\" (UniqueName: \"kubernetes.io/projected/4365133f-7067-4dd7-9ef3-77d8c1312e89-kube-api-access-r2m98\") pod \"nova-api-db-create-ww92q\" (UID: \"4365133f-7067-4dd7-9ef3-77d8c1312e89\") " pod="openstack/nova-api-db-create-ww92q" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.091223 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fhd4\" (UniqueName: \"kubernetes.io/projected/4dadb5f9-c6c6-43e0-98a2-917dd5748656-kube-api-access-9fhd4\") pod \"nova-cell0-db-create-d46hl\" (UID: \"4dadb5f9-c6c6-43e0-98a2-917dd5748656\") " pod="openstack/nova-cell0-db-create-d46hl" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.091286 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4365133f-7067-4dd7-9ef3-77d8c1312e89-operator-scripts\") pod \"nova-api-db-create-ww92q\" (UID: \"4365133f-7067-4dd7-9ef3-77d8c1312e89\") " pod="openstack/nova-api-db-create-ww92q" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.091325 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q7vp\" (UniqueName: \"kubernetes.io/projected/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-kube-api-access-6q7vp\") pod \"nova-api-8d1c-account-create-update-zhr8r\" (UID: \"a7cf88a7-6d20-4909-a39a-08fbe51bd45a\") " pod="openstack/nova-api-8d1c-account-create-update-zhr8r" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.091349 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dadb5f9-c6c6-43e0-98a2-917dd5748656-operator-scripts\") pod \"nova-cell0-db-create-d46hl\" (UID: \"4dadb5f9-c6c6-43e0-98a2-917dd5748656\") " pod="openstack/nova-cell0-db-create-d46hl" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.093838 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-operator-scripts\") pod \"nova-api-8d1c-account-create-update-zhr8r\" (UID: \"a7cf88a7-6d20-4909-a39a-08fbe51bd45a\") " pod="openstack/nova-api-8d1c-account-create-update-zhr8r" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.094054 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dadb5f9-c6c6-43e0-98a2-917dd5748656-operator-scripts\") pod \"nova-cell0-db-create-d46hl\" (UID: \"4dadb5f9-c6c6-43e0-98a2-917dd5748656\") " pod="openstack/nova-cell0-db-create-d46hl" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.094418 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4365133f-7067-4dd7-9ef3-77d8c1312e89-operator-scripts\") pod \"nova-api-db-create-ww92q\" (UID: \"4365133f-7067-4dd7-9ef3-77d8c1312e89\") " pod="openstack/nova-api-db-create-ww92q" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.108433 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fhd4\" (UniqueName: \"kubernetes.io/projected/4dadb5f9-c6c6-43e0-98a2-917dd5748656-kube-api-access-9fhd4\") pod \"nova-cell0-db-create-d46hl\" (UID: \"4dadb5f9-c6c6-43e0-98a2-917dd5748656\") " pod="openstack/nova-cell0-db-create-d46hl" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.113823 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2m98\" (UniqueName: \"kubernetes.io/projected/4365133f-7067-4dd7-9ef3-77d8c1312e89-kube-api-access-r2m98\") pod \"nova-api-db-create-ww92q\" (UID: \"4365133f-7067-4dd7-9ef3-77d8c1312e89\") " pod="openstack/nova-api-db-create-ww92q" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.132889 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q7vp\" (UniqueName: \"kubernetes.io/projected/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-kube-api-access-6q7vp\") pod \"nova-api-8d1c-account-create-update-zhr8r\" (UID: \"a7cf88a7-6d20-4909-a39a-08fbe51bd45a\") " pod="openstack/nova-api-8d1c-account-create-update-zhr8r" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.148897 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-bf46-account-create-update-v8jtg"] Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.150190 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.157399 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-bf46-account-create-update-v8jtg"] Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.157592 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.157709 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ww92q" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.192450 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/105c1f03-d5a4-43d3-93d7-ce9682adddab-operator-scripts\") pod \"nova-cell1-db-create-r7rr7\" (UID: \"105c1f03-d5a4-43d3-93d7-ce9682adddab\") " pod="openstack/nova-cell1-db-create-r7rr7" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.192606 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smscr\" (UniqueName: \"kubernetes.io/projected/105c1f03-d5a4-43d3-93d7-ce9682adddab-kube-api-access-smscr\") pod \"nova-cell1-db-create-r7rr7\" (UID: \"105c1f03-d5a4-43d3-93d7-ce9682adddab\") " pod="openstack/nova-cell1-db-create-r7rr7" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.193448 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/105c1f03-d5a4-43d3-93d7-ce9682adddab-operator-scripts\") pod \"nova-cell1-db-create-r7rr7\" (UID: \"105c1f03-d5a4-43d3-93d7-ce9682adddab\") " pod="openstack/nova-cell1-db-create-r7rr7" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.209837 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smscr\" (UniqueName: \"kubernetes.io/projected/105c1f03-d5a4-43d3-93d7-ce9682adddab-kube-api-access-smscr\") pod \"nova-cell1-db-create-r7rr7\" (UID: \"105c1f03-d5a4-43d3-93d7-ce9682adddab\") " pod="openstack/nova-cell1-db-create-r7rr7" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.255464 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-d46hl" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.283150 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8d1c-account-create-update-zhr8r" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.294834 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14ebdf50-6a9b-486a-81b7-9d1738e47101-operator-scripts\") pod \"nova-cell0-bf46-account-create-update-v8jtg\" (UID: \"14ebdf50-6a9b-486a-81b7-9d1738e47101\") " pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.294921 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnslw\" (UniqueName: \"kubernetes.io/projected/14ebdf50-6a9b-486a-81b7-9d1738e47101-kube-api-access-gnslw\") pod \"nova-cell0-bf46-account-create-update-v8jtg\" (UID: \"14ebdf50-6a9b-486a-81b7-9d1738e47101\") " pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.301165 4978 generic.go:334] "Generic (PLEG): container finished" podID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" containerID="a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4" exitCode=0 Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.301212 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsgk" event={"ID":"5d8978a9-cd99-489e-b82d-ea3bb98eb72f","Type":"ContainerDied","Data":"a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4"} Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.301245 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djsgk" event={"ID":"5d8978a9-cd99-489e-b82d-ea3bb98eb72f","Type":"ContainerDied","Data":"b25a4e387b58c1288ad31042b7c8e6edb74e3562b8736af4463a80add090852d"} Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.301263 4978 scope.go:117] "RemoveContainer" containerID="a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.301427 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-djsgk" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.351132 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-djsgk"] Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.359762 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ce19-account-create-update-99rrk"] Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.360916 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ce19-account-create-update-99rrk" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.363653 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.376205 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r7rr7" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.376772 4978 scope.go:117] "RemoveContainer" containerID="54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.376936 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-djsgk"] Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.381638 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ce19-account-create-update-99rrk"] Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.396266 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnslw\" (UniqueName: \"kubernetes.io/projected/14ebdf50-6a9b-486a-81b7-9d1738e47101-kube-api-access-gnslw\") pod \"nova-cell0-bf46-account-create-update-v8jtg\" (UID: \"14ebdf50-6a9b-486a-81b7-9d1738e47101\") " pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.396435 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14ebdf50-6a9b-486a-81b7-9d1738e47101-operator-scripts\") pod \"nova-cell0-bf46-account-create-update-v8jtg\" (UID: \"14ebdf50-6a9b-486a-81b7-9d1738e47101\") " pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.402022 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14ebdf50-6a9b-486a-81b7-9d1738e47101-operator-scripts\") pod \"nova-cell0-bf46-account-create-update-v8jtg\" (UID: \"14ebdf50-6a9b-486a-81b7-9d1738e47101\") " pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.413502 4978 scope.go:117] "RemoveContainer" containerID="747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.413938 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnslw\" (UniqueName: \"kubernetes.io/projected/14ebdf50-6a9b-486a-81b7-9d1738e47101-kube-api-access-gnslw\") pod \"nova-cell0-bf46-account-create-update-v8jtg\" (UID: \"14ebdf50-6a9b-486a-81b7-9d1738e47101\") " pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.480762 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.499068 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktjlt\" (UniqueName: \"kubernetes.io/projected/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-kube-api-access-ktjlt\") pod \"nova-cell1-ce19-account-create-update-99rrk\" (UID: \"0f36fcfd-2834-4e66-ad11-16d96d2cbf02\") " pod="openstack/nova-cell1-ce19-account-create-update-99rrk" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.499125 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-operator-scripts\") pod \"nova-cell1-ce19-account-create-update-99rrk\" (UID: \"0f36fcfd-2834-4e66-ad11-16d96d2cbf02\") " pod="openstack/nova-cell1-ce19-account-create-update-99rrk" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.514378 4978 scope.go:117] "RemoveContainer" containerID="a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4" Feb 25 08:47:48 crc kubenswrapper[4978]: E0225 08:47:48.514897 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4\": container with ID starting with a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4 not found: ID does not exist" containerID="a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.514975 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4"} err="failed to get container status \"a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4\": rpc error: code = NotFound desc = could not find container \"a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4\": container with ID starting with a2676bb4106ed13b996327761a2a5363411c270646bc89d0062f5769ee7344d4 not found: ID does not exist" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.514995 4978 scope.go:117] "RemoveContainer" containerID="54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8" Feb 25 08:47:48 crc kubenswrapper[4978]: E0225 08:47:48.515400 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8\": container with ID starting with 54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8 not found: ID does not exist" containerID="54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.515418 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8"} err="failed to get container status \"54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8\": rpc error: code = NotFound desc = could not find container \"54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8\": container with ID starting with 54f8c79e5707e18bd3ebead2e90abc8bc516588d515f3811a09c8939a46471e8 not found: ID does not exist" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.515430 4978 scope.go:117] "RemoveContainer" containerID="747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1" Feb 25 08:47:48 crc kubenswrapper[4978]: E0225 08:47:48.515656 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1\": container with ID starting with 747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1 not found: ID does not exist" containerID="747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.515677 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1"} err="failed to get container status \"747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1\": rpc error: code = NotFound desc = could not find container \"747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1\": container with ID starting with 747cb50940149a2cb72e501681938d249940860f21f811aded9503cefb501ef1 not found: ID does not exist" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.600680 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktjlt\" (UniqueName: \"kubernetes.io/projected/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-kube-api-access-ktjlt\") pod \"nova-cell1-ce19-account-create-update-99rrk\" (UID: \"0f36fcfd-2834-4e66-ad11-16d96d2cbf02\") " pod="openstack/nova-cell1-ce19-account-create-update-99rrk" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.600860 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-operator-scripts\") pod \"nova-cell1-ce19-account-create-update-99rrk\" (UID: \"0f36fcfd-2834-4e66-ad11-16d96d2cbf02\") " pod="openstack/nova-cell1-ce19-account-create-update-99rrk" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.601893 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-operator-scripts\") pod \"nova-cell1-ce19-account-create-update-99rrk\" (UID: \"0f36fcfd-2834-4e66-ad11-16d96d2cbf02\") " pod="openstack/nova-cell1-ce19-account-create-update-99rrk" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.608876 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ww92q"] Feb 25 08:47:48 crc kubenswrapper[4978]: W0225 08:47:48.613647 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4365133f_7067_4dd7_9ef3_77d8c1312e89.slice/crio-a04f7970fda2c27964c7d459e8c2a886ad64d5a460d0b3660e381f009c7663b0 WatchSource:0}: Error finding container a04f7970fda2c27964c7d459e8c2a886ad64d5a460d0b3660e381f009c7663b0: Status 404 returned error can't find the container with id a04f7970fda2c27964c7d459e8c2a886ad64d5a460d0b3660e381f009c7663b0 Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.618827 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktjlt\" (UniqueName: \"kubernetes.io/projected/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-kube-api-access-ktjlt\") pod \"nova-cell1-ce19-account-create-update-99rrk\" (UID: \"0f36fcfd-2834-4e66-ad11-16d96d2cbf02\") " pod="openstack/nova-cell1-ce19-account-create-update-99rrk" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.670041 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8d1c-account-create-update-zhr8r"] Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.696670 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ce19-account-create-update-99rrk" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.748159 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-d46hl"] Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.761740 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-r7rr7"] Feb 25 08:47:48 crc kubenswrapper[4978]: W0225 08:47:48.768401 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4dadb5f9_c6c6_43e0_98a2_917dd5748656.slice/crio-f8bc209eabc9efc2b0291ec5d682310baaa64f40231f22515ee3fbb394c5f4e0 WatchSource:0}: Error finding container f8bc209eabc9efc2b0291ec5d682310baaa64f40231f22515ee3fbb394c5f4e0: Status 404 returned error can't find the container with id f8bc209eabc9efc2b0291ec5d682310baaa64f40231f22515ee3fbb394c5f4e0 Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.770839 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.771407 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.832861 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:48 crc kubenswrapper[4978]: I0225 08:47:48.998550 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-bf46-account-create-update-v8jtg"] Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.126449 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ce19-account-create-update-99rrk"] Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.310419 4978 generic.go:334] "Generic (PLEG): container finished" podID="a7cf88a7-6d20-4909-a39a-08fbe51bd45a" containerID="f1c0033a1fabc69444f5f5fe3bfecb54fd07596c276d57302c7597782c6b1ed1" exitCode=0 Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.310517 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8d1c-account-create-update-zhr8r" event={"ID":"a7cf88a7-6d20-4909-a39a-08fbe51bd45a","Type":"ContainerDied","Data":"f1c0033a1fabc69444f5f5fe3bfecb54fd07596c276d57302c7597782c6b1ed1"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.310575 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8d1c-account-create-update-zhr8r" event={"ID":"a7cf88a7-6d20-4909-a39a-08fbe51bd45a","Type":"ContainerStarted","Data":"fa193ce8ff71bed284ffe792978c6a4705280dcebf165f73d70749eb7409e3a8"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.311948 4978 generic.go:334] "Generic (PLEG): container finished" podID="105c1f03-d5a4-43d3-93d7-ce9682adddab" containerID="6e2cc42ac90096e89d9fa3b15d415de4a4496a34981b7bd378b8763c917cc9d1" exitCode=0 Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.311979 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r7rr7" event={"ID":"105c1f03-d5a4-43d3-93d7-ce9682adddab","Type":"ContainerDied","Data":"6e2cc42ac90096e89d9fa3b15d415de4a4496a34981b7bd378b8763c917cc9d1"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.312014 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r7rr7" event={"ID":"105c1f03-d5a4-43d3-93d7-ce9682adddab","Type":"ContainerStarted","Data":"4e424eeea6e28cbbf17a2dab444592fe89d91f1e042637b301a0532ff6160540"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.313299 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" event={"ID":"14ebdf50-6a9b-486a-81b7-9d1738e47101","Type":"ContainerStarted","Data":"6d859e57b682f1cb9b5daffd35493c5a5862453b86aad2e0c87e4c11ffafaa1b"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.313342 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" event={"ID":"14ebdf50-6a9b-486a-81b7-9d1738e47101","Type":"ContainerStarted","Data":"bae45c3e011bd9ae2c5d4dc4e4fe49c6bfd965b1fa5befa365f8681854472254"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.314192 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ce19-account-create-update-99rrk" event={"ID":"0f36fcfd-2834-4e66-ad11-16d96d2cbf02","Type":"ContainerStarted","Data":"fd852b3e90c2a2531c90313479e749dc30b5f109552f51f5d31c7be0ecae6911"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.315361 4978 generic.go:334] "Generic (PLEG): container finished" podID="4dadb5f9-c6c6-43e0-98a2-917dd5748656" containerID="fbb0605b18bbb774138802fe465d1ceb037c141f087e0f5a2d3699959eb4dfb4" exitCode=0 Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.315442 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-d46hl" event={"ID":"4dadb5f9-c6c6-43e0-98a2-917dd5748656","Type":"ContainerDied","Data":"fbb0605b18bbb774138802fe465d1ceb037c141f087e0f5a2d3699959eb4dfb4"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.315477 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-d46hl" event={"ID":"4dadb5f9-c6c6-43e0-98a2-917dd5748656","Type":"ContainerStarted","Data":"f8bc209eabc9efc2b0291ec5d682310baaa64f40231f22515ee3fbb394c5f4e0"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.316961 4978 generic.go:334] "Generic (PLEG): container finished" podID="4365133f-7067-4dd7-9ef3-77d8c1312e89" containerID="8a86ab3bcf2f60da67cdebd1d4fadad5f4755e71ea59fc896f3955ecfe5b556c" exitCode=0 Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.317026 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ww92q" event={"ID":"4365133f-7067-4dd7-9ef3-77d8c1312e89","Type":"ContainerDied","Data":"8a86ab3bcf2f60da67cdebd1d4fadad5f4755e71ea59fc896f3955ecfe5b556c"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.317048 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ww92q" event={"ID":"4365133f-7067-4dd7-9ef3-77d8c1312e89","Type":"ContainerStarted","Data":"a04f7970fda2c27964c7d459e8c2a886ad64d5a460d0b3660e381f009c7663b0"} Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.342479 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d8978a9-cd99-489e-b82d-ea3bb98eb72f" path="/var/lib/kubelet/pods/5d8978a9-cd99-489e-b82d-ea3bb98eb72f/volumes" Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.356781 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" podStartSLOduration=1.356766407 podStartE2EDuration="1.356766407s" podCreationTimestamp="2026-02-25 08:47:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:47:49.354012172 +0000 UTC m=+7362.793268631" watchObservedRunningTime="2026-02-25 08:47:49.356766407 +0000 UTC m=+7362.796022866" Feb 25 08:47:49 crc kubenswrapper[4978]: I0225 08:47:49.378341 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.328792 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:47:50 crc kubenswrapper[4978]: E0225 08:47:50.329809 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.334656 4978 generic.go:334] "Generic (PLEG): container finished" podID="14ebdf50-6a9b-486a-81b7-9d1738e47101" containerID="6d859e57b682f1cb9b5daffd35493c5a5862453b86aad2e0c87e4c11ffafaa1b" exitCode=0 Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.334728 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" event={"ID":"14ebdf50-6a9b-486a-81b7-9d1738e47101","Type":"ContainerDied","Data":"6d859e57b682f1cb9b5daffd35493c5a5862453b86aad2e0c87e4c11ffafaa1b"} Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.337057 4978 generic.go:334] "Generic (PLEG): container finished" podID="0f36fcfd-2834-4e66-ad11-16d96d2cbf02" containerID="c28cf767dd6d6ef31292b5a97fd23d78ef85205f6de7a28d720aec88b635f002" exitCode=0 Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.337116 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ce19-account-create-update-99rrk" event={"ID":"0f36fcfd-2834-4e66-ad11-16d96d2cbf02","Type":"ContainerDied","Data":"c28cf767dd6d6ef31292b5a97fd23d78ef85205f6de7a28d720aec88b635f002"} Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.811312 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8d1c-account-create-update-zhr8r" Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.951526 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q7vp\" (UniqueName: \"kubernetes.io/projected/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-kube-api-access-6q7vp\") pod \"a7cf88a7-6d20-4909-a39a-08fbe51bd45a\" (UID: \"a7cf88a7-6d20-4909-a39a-08fbe51bd45a\") " Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.951616 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-operator-scripts\") pod \"a7cf88a7-6d20-4909-a39a-08fbe51bd45a\" (UID: \"a7cf88a7-6d20-4909-a39a-08fbe51bd45a\") " Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.951956 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a7cf88a7-6d20-4909-a39a-08fbe51bd45a" (UID: "a7cf88a7-6d20-4909-a39a-08fbe51bd45a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.952066 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:50 crc kubenswrapper[4978]: I0225 08:47:50.956706 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-kube-api-access-6q7vp" (OuterVolumeSpecName: "kube-api-access-6q7vp") pod "a7cf88a7-6d20-4909-a39a-08fbe51bd45a" (UID: "a7cf88a7-6d20-4909-a39a-08fbe51bd45a"). InnerVolumeSpecName "kube-api-access-6q7vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.009678 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-d46hl" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.015655 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bcb8d"] Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.019638 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ww92q" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.032050 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r7rr7" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.054478 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q7vp\" (UniqueName: \"kubernetes.io/projected/a7cf88a7-6d20-4909-a39a-08fbe51bd45a-kube-api-access-6q7vp\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.155804 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4365133f-7067-4dd7-9ef3-77d8c1312e89-operator-scripts\") pod \"4365133f-7067-4dd7-9ef3-77d8c1312e89\" (UID: \"4365133f-7067-4dd7-9ef3-77d8c1312e89\") " Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.155972 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/105c1f03-d5a4-43d3-93d7-ce9682adddab-operator-scripts\") pod \"105c1f03-d5a4-43d3-93d7-ce9682adddab\" (UID: \"105c1f03-d5a4-43d3-93d7-ce9682adddab\") " Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.156012 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dadb5f9-c6c6-43e0-98a2-917dd5748656-operator-scripts\") pod \"4dadb5f9-c6c6-43e0-98a2-917dd5748656\" (UID: \"4dadb5f9-c6c6-43e0-98a2-917dd5748656\") " Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.156042 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smscr\" (UniqueName: \"kubernetes.io/projected/105c1f03-d5a4-43d3-93d7-ce9682adddab-kube-api-access-smscr\") pod \"105c1f03-d5a4-43d3-93d7-ce9682adddab\" (UID: \"105c1f03-d5a4-43d3-93d7-ce9682adddab\") " Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.156076 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2m98\" (UniqueName: \"kubernetes.io/projected/4365133f-7067-4dd7-9ef3-77d8c1312e89-kube-api-access-r2m98\") pod \"4365133f-7067-4dd7-9ef3-77d8c1312e89\" (UID: \"4365133f-7067-4dd7-9ef3-77d8c1312e89\") " Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.156109 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fhd4\" (UniqueName: \"kubernetes.io/projected/4dadb5f9-c6c6-43e0-98a2-917dd5748656-kube-api-access-9fhd4\") pod \"4dadb5f9-c6c6-43e0-98a2-917dd5748656\" (UID: \"4dadb5f9-c6c6-43e0-98a2-917dd5748656\") " Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.156819 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dadb5f9-c6c6-43e0-98a2-917dd5748656-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4dadb5f9-c6c6-43e0-98a2-917dd5748656" (UID: "4dadb5f9-c6c6-43e0-98a2-917dd5748656"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.157180 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4365133f-7067-4dd7-9ef3-77d8c1312e89-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4365133f-7067-4dd7-9ef3-77d8c1312e89" (UID: "4365133f-7067-4dd7-9ef3-77d8c1312e89"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.157235 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/105c1f03-d5a4-43d3-93d7-ce9682adddab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "105c1f03-d5a4-43d3-93d7-ce9682adddab" (UID: "105c1f03-d5a4-43d3-93d7-ce9682adddab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.160590 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/105c1f03-d5a4-43d3-93d7-ce9682adddab-kube-api-access-smscr" (OuterVolumeSpecName: "kube-api-access-smscr") pod "105c1f03-d5a4-43d3-93d7-ce9682adddab" (UID: "105c1f03-d5a4-43d3-93d7-ce9682adddab"). InnerVolumeSpecName "kube-api-access-smscr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.161036 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dadb5f9-c6c6-43e0-98a2-917dd5748656-kube-api-access-9fhd4" (OuterVolumeSpecName: "kube-api-access-9fhd4") pod "4dadb5f9-c6c6-43e0-98a2-917dd5748656" (UID: "4dadb5f9-c6c6-43e0-98a2-917dd5748656"). InnerVolumeSpecName "kube-api-access-9fhd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.162291 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4365133f-7067-4dd7-9ef3-77d8c1312e89-kube-api-access-r2m98" (OuterVolumeSpecName: "kube-api-access-r2m98") pod "4365133f-7067-4dd7-9ef3-77d8c1312e89" (UID: "4365133f-7067-4dd7-9ef3-77d8c1312e89"). InnerVolumeSpecName "kube-api-access-r2m98". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.258828 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smscr\" (UniqueName: \"kubernetes.io/projected/105c1f03-d5a4-43d3-93d7-ce9682adddab-kube-api-access-smscr\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.258874 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2m98\" (UniqueName: \"kubernetes.io/projected/4365133f-7067-4dd7-9ef3-77d8c1312e89-kube-api-access-r2m98\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.258893 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fhd4\" (UniqueName: \"kubernetes.io/projected/4dadb5f9-c6c6-43e0-98a2-917dd5748656-kube-api-access-9fhd4\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.258911 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4365133f-7067-4dd7-9ef3-77d8c1312e89-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.258931 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/105c1f03-d5a4-43d3-93d7-ce9682adddab-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.258949 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dadb5f9-c6c6-43e0-98a2-917dd5748656-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.348210 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ww92q" event={"ID":"4365133f-7067-4dd7-9ef3-77d8c1312e89","Type":"ContainerDied","Data":"a04f7970fda2c27964c7d459e8c2a886ad64d5a460d0b3660e381f009c7663b0"} Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.348245 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ww92q" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.348270 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a04f7970fda2c27964c7d459e8c2a886ad64d5a460d0b3660e381f009c7663b0" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.352358 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8d1c-account-create-update-zhr8r" event={"ID":"a7cf88a7-6d20-4909-a39a-08fbe51bd45a","Type":"ContainerDied","Data":"fa193ce8ff71bed284ffe792978c6a4705280dcebf165f73d70749eb7409e3a8"} Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.352421 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8d1c-account-create-update-zhr8r" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.352420 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa193ce8ff71bed284ffe792978c6a4705280dcebf165f73d70749eb7409e3a8" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.353924 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-r7rr7" event={"ID":"105c1f03-d5a4-43d3-93d7-ce9682adddab","Type":"ContainerDied","Data":"4e424eeea6e28cbbf17a2dab444592fe89d91f1e042637b301a0532ff6160540"} Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.353974 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e424eeea6e28cbbf17a2dab444592fe89d91f1e042637b301a0532ff6160540" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.354059 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-r7rr7" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.363466 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-d46hl" event={"ID":"4dadb5f9-c6c6-43e0-98a2-917dd5748656","Type":"ContainerDied","Data":"f8bc209eabc9efc2b0291ec5d682310baaa64f40231f22515ee3fbb394c5f4e0"} Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.363514 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8bc209eabc9efc2b0291ec5d682310baaa64f40231f22515ee3fbb394c5f4e0" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.363960 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-d46hl" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.690685 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.775059 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14ebdf50-6a9b-486a-81b7-9d1738e47101-operator-scripts\") pod \"14ebdf50-6a9b-486a-81b7-9d1738e47101\" (UID: \"14ebdf50-6a9b-486a-81b7-9d1738e47101\") " Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.775167 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnslw\" (UniqueName: \"kubernetes.io/projected/14ebdf50-6a9b-486a-81b7-9d1738e47101-kube-api-access-gnslw\") pod \"14ebdf50-6a9b-486a-81b7-9d1738e47101\" (UID: \"14ebdf50-6a9b-486a-81b7-9d1738e47101\") " Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.776803 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14ebdf50-6a9b-486a-81b7-9d1738e47101-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "14ebdf50-6a9b-486a-81b7-9d1738e47101" (UID: "14ebdf50-6a9b-486a-81b7-9d1738e47101"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.780621 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ebdf50-6a9b-486a-81b7-9d1738e47101-kube-api-access-gnslw" (OuterVolumeSpecName: "kube-api-access-gnslw") pod "14ebdf50-6a9b-486a-81b7-9d1738e47101" (UID: "14ebdf50-6a9b-486a-81b7-9d1738e47101"). InnerVolumeSpecName "kube-api-access-gnslw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.850305 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ce19-account-create-update-99rrk" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.877450 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnslw\" (UniqueName: \"kubernetes.io/projected/14ebdf50-6a9b-486a-81b7-9d1738e47101-kube-api-access-gnslw\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.877491 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14ebdf50-6a9b-486a-81b7-9d1738e47101-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.978418 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-operator-scripts\") pod \"0f36fcfd-2834-4e66-ad11-16d96d2cbf02\" (UID: \"0f36fcfd-2834-4e66-ad11-16d96d2cbf02\") " Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.978764 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktjlt\" (UniqueName: \"kubernetes.io/projected/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-kube-api-access-ktjlt\") pod \"0f36fcfd-2834-4e66-ad11-16d96d2cbf02\" (UID: \"0f36fcfd-2834-4e66-ad11-16d96d2cbf02\") " Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.979173 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f36fcfd-2834-4e66-ad11-16d96d2cbf02" (UID: "0f36fcfd-2834-4e66-ad11-16d96d2cbf02"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.979492 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:51 crc kubenswrapper[4978]: I0225 08:47:51.982063 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-kube-api-access-ktjlt" (OuterVolumeSpecName: "kube-api-access-ktjlt") pod "0f36fcfd-2834-4e66-ad11-16d96d2cbf02" (UID: "0f36fcfd-2834-4e66-ad11-16d96d2cbf02"). InnerVolumeSpecName "kube-api-access-ktjlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:47:52 crc kubenswrapper[4978]: I0225 08:47:52.082025 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktjlt\" (UniqueName: \"kubernetes.io/projected/0f36fcfd-2834-4e66-ad11-16d96d2cbf02-kube-api-access-ktjlt\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:52 crc kubenswrapper[4978]: I0225 08:47:52.371838 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" event={"ID":"14ebdf50-6a9b-486a-81b7-9d1738e47101","Type":"ContainerDied","Data":"bae45c3e011bd9ae2c5d4dc4e4fe49c6bfd965b1fa5befa365f8681854472254"} Feb 25 08:47:52 crc kubenswrapper[4978]: I0225 08:47:52.371876 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bae45c3e011bd9ae2c5d4dc4e4fe49c6bfd965b1fa5befa365f8681854472254" Feb 25 08:47:52 crc kubenswrapper[4978]: I0225 08:47:52.371896 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-bf46-account-create-update-v8jtg" Feb 25 08:47:52 crc kubenswrapper[4978]: I0225 08:47:52.373679 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ce19-account-create-update-99rrk" event={"ID":"0f36fcfd-2834-4e66-ad11-16d96d2cbf02","Type":"ContainerDied","Data":"fd852b3e90c2a2531c90313479e749dc30b5f109552f51f5d31c7be0ecae6911"} Feb 25 08:47:52 crc kubenswrapper[4978]: I0225 08:47:52.373742 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd852b3e90c2a2531c90313479e749dc30b5f109552f51f5d31c7be0ecae6911" Feb 25 08:47:52 crc kubenswrapper[4978]: I0225 08:47:52.373700 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bcb8d" podUID="64897ae4-94ce-4733-a084-011ce63b8719" containerName="registry-server" containerID="cri-o://1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238" gracePeriod=2 Feb 25 08:47:52 crc kubenswrapper[4978]: I0225 08:47:52.373957 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ce19-account-create-update-99rrk" Feb 25 08:47:52 crc kubenswrapper[4978]: I0225 08:47:52.936672 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.003240 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-catalog-content\") pod \"64897ae4-94ce-4733-a084-011ce63b8719\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.003337 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-utilities\") pod \"64897ae4-94ce-4733-a084-011ce63b8719\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.003510 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84knk\" (UniqueName: \"kubernetes.io/projected/64897ae4-94ce-4733-a084-011ce63b8719-kube-api-access-84knk\") pod \"64897ae4-94ce-4733-a084-011ce63b8719\" (UID: \"64897ae4-94ce-4733-a084-011ce63b8719\") " Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.004833 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-utilities" (OuterVolumeSpecName: "utilities") pod "64897ae4-94ce-4733-a084-011ce63b8719" (UID: "64897ae4-94ce-4733-a084-011ce63b8719"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.009449 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64897ae4-94ce-4733-a084-011ce63b8719-kube-api-access-84knk" (OuterVolumeSpecName: "kube-api-access-84knk") pod "64897ae4-94ce-4733-a084-011ce63b8719" (UID: "64897ae4-94ce-4733-a084-011ce63b8719"). InnerVolumeSpecName "kube-api-access-84knk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.037894 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64897ae4-94ce-4733-a084-011ce63b8719" (UID: "64897ae4-94ce-4733-a084-011ce63b8719"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.105281 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.105305 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64897ae4-94ce-4733-a084-011ce63b8719-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.105315 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84knk\" (UniqueName: \"kubernetes.io/projected/64897ae4-94ce-4733-a084-011ce63b8719-kube-api-access-84knk\") on node \"crc\" DevicePath \"\"" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.344904 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wg2x6"] Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.346338 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64897ae4-94ce-4733-a084-011ce63b8719" containerName="registry-server" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346380 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="64897ae4-94ce-4733-a084-011ce63b8719" containerName="registry-server" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.346399 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f36fcfd-2834-4e66-ad11-16d96d2cbf02" containerName="mariadb-account-create-update" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346407 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f36fcfd-2834-4e66-ad11-16d96d2cbf02" containerName="mariadb-account-create-update" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.346425 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105c1f03-d5a4-43d3-93d7-ce9682adddab" containerName="mariadb-database-create" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346433 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="105c1f03-d5a4-43d3-93d7-ce9682adddab" containerName="mariadb-database-create" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.346450 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4365133f-7067-4dd7-9ef3-77d8c1312e89" containerName="mariadb-database-create" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346458 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4365133f-7067-4dd7-9ef3-77d8c1312e89" containerName="mariadb-database-create" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.346472 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7cf88a7-6d20-4909-a39a-08fbe51bd45a" containerName="mariadb-account-create-update" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346480 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7cf88a7-6d20-4909-a39a-08fbe51bd45a" containerName="mariadb-account-create-update" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.346507 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dadb5f9-c6c6-43e0-98a2-917dd5748656" containerName="mariadb-database-create" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346514 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dadb5f9-c6c6-43e0-98a2-917dd5748656" containerName="mariadb-database-create" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.346534 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64897ae4-94ce-4733-a084-011ce63b8719" containerName="extract-utilities" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346542 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="64897ae4-94ce-4733-a084-011ce63b8719" containerName="extract-utilities" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.346589 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64897ae4-94ce-4733-a084-011ce63b8719" containerName="extract-content" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346596 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="64897ae4-94ce-4733-a084-011ce63b8719" containerName="extract-content" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.346610 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ebdf50-6a9b-486a-81b7-9d1738e47101" containerName="mariadb-account-create-update" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346618 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ebdf50-6a9b-486a-81b7-9d1738e47101" containerName="mariadb-account-create-update" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346820 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dadb5f9-c6c6-43e0-98a2-917dd5748656" containerName="mariadb-database-create" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346844 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f36fcfd-2834-4e66-ad11-16d96d2cbf02" containerName="mariadb-account-create-update" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346855 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4365133f-7067-4dd7-9ef3-77d8c1312e89" containerName="mariadb-database-create" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346871 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="105c1f03-d5a4-43d3-93d7-ce9682adddab" containerName="mariadb-database-create" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346888 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ebdf50-6a9b-486a-81b7-9d1738e47101" containerName="mariadb-account-create-update" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346898 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="64897ae4-94ce-4733-a084-011ce63b8719" containerName="registry-server" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.346909 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7cf88a7-6d20-4909-a39a-08fbe51bd45a" containerName="mariadb-account-create-update" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.347801 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.349452 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.350846 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9ld7q" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.351844 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.359807 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wg2x6"] Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.393296 4978 generic.go:334] "Generic (PLEG): container finished" podID="64897ae4-94ce-4733-a084-011ce63b8719" containerID="1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238" exitCode=0 Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.393338 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bcb8d" event={"ID":"64897ae4-94ce-4733-a084-011ce63b8719","Type":"ContainerDied","Data":"1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238"} Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.393391 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bcb8d" event={"ID":"64897ae4-94ce-4733-a084-011ce63b8719","Type":"ContainerDied","Data":"03021d803b60c0462796cd5a87bf5033cc482d2dd4465c016acb02cd4c9f4586"} Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.393414 4978 scope.go:117] "RemoveContainer" containerID="1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.393466 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bcb8d" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.412168 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-config-data\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.412463 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-scripts\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.412698 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhnj9\" (UniqueName: \"kubernetes.io/projected/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-kube-api-access-qhnj9\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.412987 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.416352 4978 scope.go:117] "RemoveContainer" containerID="243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.424897 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bcb8d"] Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.447589 4978 scope.go:117] "RemoveContainer" containerID="64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.450195 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bcb8d"] Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.498537 4978 scope.go:117] "RemoveContainer" containerID="1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.498997 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238\": container with ID starting with 1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238 not found: ID does not exist" containerID="1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.499034 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238"} err="failed to get container status \"1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238\": rpc error: code = NotFound desc = could not find container \"1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238\": container with ID starting with 1d9c5d16c21eb8399e51e1d92a79f86797dc6e9642c439d065ff21aa2a0f9238 not found: ID does not exist" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.499060 4978 scope.go:117] "RemoveContainer" containerID="243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.499436 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22\": container with ID starting with 243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22 not found: ID does not exist" containerID="243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.499533 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22"} err="failed to get container status \"243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22\": rpc error: code = NotFound desc = could not find container \"243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22\": container with ID starting with 243ce04935b11ac184b248f9e85cba859b39eb5d5c2b4a17fbafcabc04983b22 not found: ID does not exist" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.499619 4978 scope.go:117] "RemoveContainer" containerID="64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0" Feb 25 08:47:53 crc kubenswrapper[4978]: E0225 08:47:53.500023 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0\": container with ID starting with 64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0 not found: ID does not exist" containerID="64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.500142 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0"} err="failed to get container status \"64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0\": rpc error: code = NotFound desc = could not find container \"64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0\": container with ID starting with 64ac268918379c8ef9547a7d4e5fdaeb0801dcb20fae8584bc37ac993d8135d0 not found: ID does not exist" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.514426 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhnj9\" (UniqueName: \"kubernetes.io/projected/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-kube-api-access-qhnj9\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.514608 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.514776 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-config-data\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.514873 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-scripts\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.519489 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.519616 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-config-data\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.520050 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-scripts\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.530435 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhnj9\" (UniqueName: \"kubernetes.io/projected/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-kube-api-access-qhnj9\") pod \"nova-cell0-conductor-db-sync-wg2x6\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:53 crc kubenswrapper[4978]: I0225 08:47:53.672945 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:47:54 crc kubenswrapper[4978]: I0225 08:47:54.212436 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wg2x6"] Feb 25 08:47:54 crc kubenswrapper[4978]: W0225 08:47:54.230045 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c9e435e_8baf_4b51_8bf3_e43fc7dccf37.slice/crio-8159beeaff9af9f6e624dde221abb487dfa374b8e8df3afc0222059a587535f8 WatchSource:0}: Error finding container 8159beeaff9af9f6e624dde221abb487dfa374b8e8df3afc0222059a587535f8: Status 404 returned error can't find the container with id 8159beeaff9af9f6e624dde221abb487dfa374b8e8df3afc0222059a587535f8 Feb 25 08:47:54 crc kubenswrapper[4978]: I0225 08:47:54.404881 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wg2x6" event={"ID":"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37","Type":"ContainerStarted","Data":"8159beeaff9af9f6e624dde221abb487dfa374b8e8df3afc0222059a587535f8"} Feb 25 08:47:55 crc kubenswrapper[4978]: I0225 08:47:55.339796 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64897ae4-94ce-4733-a084-011ce63b8719" path="/var/lib/kubelet/pods/64897ae4-94ce-4733-a084-011ce63b8719/volumes" Feb 25 08:48:00 crc kubenswrapper[4978]: I0225 08:48:00.155099 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533488-pfcfp"] Feb 25 08:48:00 crc kubenswrapper[4978]: I0225 08:48:00.158030 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533488-pfcfp" Feb 25 08:48:00 crc kubenswrapper[4978]: I0225 08:48:00.161862 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:48:00 crc kubenswrapper[4978]: I0225 08:48:00.162044 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:48:00 crc kubenswrapper[4978]: I0225 08:48:00.162703 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:48:00 crc kubenswrapper[4978]: I0225 08:48:00.169829 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533488-pfcfp"] Feb 25 08:48:00 crc kubenswrapper[4978]: I0225 08:48:00.288112 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vswnn\" (UniqueName: \"kubernetes.io/projected/8cb4a363-4e1d-4b95-bacd-ab271457785f-kube-api-access-vswnn\") pod \"auto-csr-approver-29533488-pfcfp\" (UID: \"8cb4a363-4e1d-4b95-bacd-ab271457785f\") " pod="openshift-infra/auto-csr-approver-29533488-pfcfp" Feb 25 08:48:00 crc kubenswrapper[4978]: I0225 08:48:00.390591 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vswnn\" (UniqueName: \"kubernetes.io/projected/8cb4a363-4e1d-4b95-bacd-ab271457785f-kube-api-access-vswnn\") pod \"auto-csr-approver-29533488-pfcfp\" (UID: \"8cb4a363-4e1d-4b95-bacd-ab271457785f\") " pod="openshift-infra/auto-csr-approver-29533488-pfcfp" Feb 25 08:48:00 crc kubenswrapper[4978]: I0225 08:48:00.416004 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vswnn\" (UniqueName: \"kubernetes.io/projected/8cb4a363-4e1d-4b95-bacd-ab271457785f-kube-api-access-vswnn\") pod \"auto-csr-approver-29533488-pfcfp\" (UID: \"8cb4a363-4e1d-4b95-bacd-ab271457785f\") " pod="openshift-infra/auto-csr-approver-29533488-pfcfp" Feb 25 08:48:00 crc kubenswrapper[4978]: I0225 08:48:00.492201 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533488-pfcfp" Feb 25 08:48:03 crc kubenswrapper[4978]: W0225 08:48:03.482682 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cb4a363_4e1d_4b95_bacd_ab271457785f.slice/crio-12d8440c702fb0cb9ebde79e7ec6775765a97689a7c884e6aec270ea5b2693ee WatchSource:0}: Error finding container 12d8440c702fb0cb9ebde79e7ec6775765a97689a7c884e6aec270ea5b2693ee: Status 404 returned error can't find the container with id 12d8440c702fb0cb9ebde79e7ec6775765a97689a7c884e6aec270ea5b2693ee Feb 25 08:48:03 crc kubenswrapper[4978]: I0225 08:48:03.482968 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533488-pfcfp"] Feb 25 08:48:03 crc kubenswrapper[4978]: I0225 08:48:03.495064 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wg2x6" event={"ID":"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37","Type":"ContainerStarted","Data":"702d53de66165222bd7c2d97cd679ada7c060aed8645c986e09b3f658c29a92e"} Feb 25 08:48:03 crc kubenswrapper[4978]: I0225 08:48:03.517498 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-wg2x6" podStartSLOduration=1.703398261 podStartE2EDuration="10.517482048s" podCreationTimestamp="2026-02-25 08:47:53 +0000 UTC" firstStartedPulling="2026-02-25 08:47:54.239420576 +0000 UTC m=+7367.678677045" lastFinishedPulling="2026-02-25 08:48:03.053504363 +0000 UTC m=+7376.492760832" observedRunningTime="2026-02-25 08:48:03.516555969 +0000 UTC m=+7376.955812488" watchObservedRunningTime="2026-02-25 08:48:03.517482048 +0000 UTC m=+7376.956738507" Feb 25 08:48:04 crc kubenswrapper[4978]: I0225 08:48:04.515792 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533488-pfcfp" event={"ID":"8cb4a363-4e1d-4b95-bacd-ab271457785f","Type":"ContainerStarted","Data":"12d8440c702fb0cb9ebde79e7ec6775765a97689a7c884e6aec270ea5b2693ee"} Feb 25 08:48:05 crc kubenswrapper[4978]: I0225 08:48:05.328532 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:48:05 crc kubenswrapper[4978]: E0225 08:48:05.329644 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:48:05 crc kubenswrapper[4978]: I0225 08:48:05.527614 4978 generic.go:334] "Generic (PLEG): container finished" podID="8cb4a363-4e1d-4b95-bacd-ab271457785f" containerID="6cc0f8b5f7058efef4006d3e810c23c9b3121b26c989cf6949a622f33dfb40a7" exitCode=0 Feb 25 08:48:05 crc kubenswrapper[4978]: I0225 08:48:05.527694 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533488-pfcfp" event={"ID":"8cb4a363-4e1d-4b95-bacd-ab271457785f","Type":"ContainerDied","Data":"6cc0f8b5f7058efef4006d3e810c23c9b3121b26c989cf6949a622f33dfb40a7"} Feb 25 08:48:06 crc kubenswrapper[4978]: I0225 08:48:06.971896 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533488-pfcfp" Feb 25 08:48:07 crc kubenswrapper[4978]: I0225 08:48:07.021572 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vswnn\" (UniqueName: \"kubernetes.io/projected/8cb4a363-4e1d-4b95-bacd-ab271457785f-kube-api-access-vswnn\") pod \"8cb4a363-4e1d-4b95-bacd-ab271457785f\" (UID: \"8cb4a363-4e1d-4b95-bacd-ab271457785f\") " Feb 25 08:48:07 crc kubenswrapper[4978]: I0225 08:48:07.029625 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cb4a363-4e1d-4b95-bacd-ab271457785f-kube-api-access-vswnn" (OuterVolumeSpecName: "kube-api-access-vswnn") pod "8cb4a363-4e1d-4b95-bacd-ab271457785f" (UID: "8cb4a363-4e1d-4b95-bacd-ab271457785f"). InnerVolumeSpecName "kube-api-access-vswnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:48:07 crc kubenswrapper[4978]: I0225 08:48:07.123357 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vswnn\" (UniqueName: \"kubernetes.io/projected/8cb4a363-4e1d-4b95-bacd-ab271457785f-kube-api-access-vswnn\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:07 crc kubenswrapper[4978]: I0225 08:48:07.555891 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533488-pfcfp" event={"ID":"8cb4a363-4e1d-4b95-bacd-ab271457785f","Type":"ContainerDied","Data":"12d8440c702fb0cb9ebde79e7ec6775765a97689a7c884e6aec270ea5b2693ee"} Feb 25 08:48:07 crc kubenswrapper[4978]: I0225 08:48:07.555933 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12d8440c702fb0cb9ebde79e7ec6775765a97689a7c884e6aec270ea5b2693ee" Feb 25 08:48:07 crc kubenswrapper[4978]: I0225 08:48:07.555977 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533488-pfcfp" Feb 25 08:48:08 crc kubenswrapper[4978]: I0225 08:48:08.057169 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533482-wvthz"] Feb 25 08:48:08 crc kubenswrapper[4978]: I0225 08:48:08.072323 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533482-wvthz"] Feb 25 08:48:08 crc kubenswrapper[4978]: I0225 08:48:08.570928 4978 generic.go:334] "Generic (PLEG): container finished" podID="2c9e435e-8baf-4b51-8bf3-e43fc7dccf37" containerID="702d53de66165222bd7c2d97cd679ada7c060aed8645c986e09b3f658c29a92e" exitCode=0 Feb 25 08:48:08 crc kubenswrapper[4978]: I0225 08:48:08.570999 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wg2x6" event={"ID":"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37","Type":"ContainerDied","Data":"702d53de66165222bd7c2d97cd679ada7c060aed8645c986e09b3f658c29a92e"} Feb 25 08:48:09 crc kubenswrapper[4978]: I0225 08:48:09.340448 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a" path="/var/lib/kubelet/pods/fcdffbf5-f70a-4dd9-acc9-2a93f5c9bf2a/volumes" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.012063 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.089435 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-scripts\") pod \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.089586 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhnj9\" (UniqueName: \"kubernetes.io/projected/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-kube-api-access-qhnj9\") pod \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.089670 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-combined-ca-bundle\") pod \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.089886 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-config-data\") pod \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\" (UID: \"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37\") " Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.095519 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-scripts" (OuterVolumeSpecName: "scripts") pod "2c9e435e-8baf-4b51-8bf3-e43fc7dccf37" (UID: "2c9e435e-8baf-4b51-8bf3-e43fc7dccf37"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.112002 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-kube-api-access-qhnj9" (OuterVolumeSpecName: "kube-api-access-qhnj9") pod "2c9e435e-8baf-4b51-8bf3-e43fc7dccf37" (UID: "2c9e435e-8baf-4b51-8bf3-e43fc7dccf37"). InnerVolumeSpecName "kube-api-access-qhnj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.118252 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-config-data" (OuterVolumeSpecName: "config-data") pod "2c9e435e-8baf-4b51-8bf3-e43fc7dccf37" (UID: "2c9e435e-8baf-4b51-8bf3-e43fc7dccf37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.119804 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c9e435e-8baf-4b51-8bf3-e43fc7dccf37" (UID: "2c9e435e-8baf-4b51-8bf3-e43fc7dccf37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.192879 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.193418 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.193440 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.193460 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhnj9\" (UniqueName: \"kubernetes.io/projected/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37-kube-api-access-qhnj9\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.600854 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wg2x6" event={"ID":"2c9e435e-8baf-4b51-8bf3-e43fc7dccf37","Type":"ContainerDied","Data":"8159beeaff9af9f6e624dde221abb487dfa374b8e8df3afc0222059a587535f8"} Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.600920 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8159beeaff9af9f6e624dde221abb487dfa374b8e8df3afc0222059a587535f8" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.601030 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wg2x6" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.709578 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 08:48:10 crc kubenswrapper[4978]: E0225 08:48:10.710059 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb4a363-4e1d-4b95-bacd-ab271457785f" containerName="oc" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.710094 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb4a363-4e1d-4b95-bacd-ab271457785f" containerName="oc" Feb 25 08:48:10 crc kubenswrapper[4978]: E0225 08:48:10.710113 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9e435e-8baf-4b51-8bf3-e43fc7dccf37" containerName="nova-cell0-conductor-db-sync" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.710121 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9e435e-8baf-4b51-8bf3-e43fc7dccf37" containerName="nova-cell0-conductor-db-sync" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.710781 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c9e435e-8baf-4b51-8bf3-e43fc7dccf37" containerName="nova-cell0-conductor-db-sync" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.710819 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cb4a363-4e1d-4b95-bacd-ab271457785f" containerName="oc" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.711404 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.713507 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.742200 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9ld7q" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.760102 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.811444 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.811549 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.811577 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn5zp\" (UniqueName: \"kubernetes.io/projected/1282752c-57e3-4e24-aa3f-cecdea868016-kube-api-access-vn5zp\") pod \"nova-cell0-conductor-0\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.914237 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.914317 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn5zp\" (UniqueName: \"kubernetes.io/projected/1282752c-57e3-4e24-aa3f-cecdea868016-kube-api-access-vn5zp\") pod \"nova-cell0-conductor-0\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.914593 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.928227 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.929080 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:10 crc kubenswrapper[4978]: I0225 08:48:10.931461 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn5zp\" (UniqueName: \"kubernetes.io/projected/1282752c-57e3-4e24-aa3f-cecdea868016-kube-api-access-vn5zp\") pod \"nova-cell0-conductor-0\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:11 crc kubenswrapper[4978]: I0225 08:48:11.077031 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:11 crc kubenswrapper[4978]: I0225 08:48:11.595887 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 08:48:11 crc kubenswrapper[4978]: I0225 08:48:11.618481 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1282752c-57e3-4e24-aa3f-cecdea868016","Type":"ContainerStarted","Data":"178d8e1065554bc7c5ad34b946b36eba3b039a2726d6341870a42d43c3a5e5b8"} Feb 25 08:48:12 crc kubenswrapper[4978]: I0225 08:48:12.635795 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1282752c-57e3-4e24-aa3f-cecdea868016","Type":"ContainerStarted","Data":"a96209f29de2fef3610427b4fa5713572b82f84289fa56e7de27488ba09e7301"} Feb 25 08:48:12 crc kubenswrapper[4978]: I0225 08:48:12.636849 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:12 crc kubenswrapper[4978]: I0225 08:48:12.674833 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.674811441 podStartE2EDuration="2.674811441s" podCreationTimestamp="2026-02-25 08:48:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:48:12.671200858 +0000 UTC m=+7386.110457327" watchObservedRunningTime="2026-02-25 08:48:12.674811441 +0000 UTC m=+7386.114067910" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.111514 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.647766 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-6lvjg"] Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.649969 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.652296 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.653228 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.658269 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-6lvjg"] Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.738957 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5dfh\" (UniqueName: \"kubernetes.io/projected/55021e7b-713c-4697-a80b-49e0a23cff45-kube-api-access-z5dfh\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.739048 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.739075 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-config-data\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.739115 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-scripts\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.840039 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-scripts\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.840161 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5dfh\" (UniqueName: \"kubernetes.io/projected/55021e7b-713c-4697-a80b-49e0a23cff45-kube-api-access-z5dfh\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.840212 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.840228 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-config-data\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.846907 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-config-data\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.847329 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-scripts\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.855447 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.855547 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.857985 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.863948 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.864464 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.865856 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.867566 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.878603 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.879337 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5dfh\" (UniqueName: \"kubernetes.io/projected/55021e7b-713c-4697-a80b-49e0a23cff45-kube-api-access-z5dfh\") pod \"nova-cell0-cell-mapping-6lvjg\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.895596 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.911173 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.912347 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.918805 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.941958 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fnr7\" (UniqueName: \"kubernetes.io/projected/598d57b6-5348-4ff3-b5ba-b4df010ef25f-kube-api-access-4fnr7\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.941992 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/598d57b6-5348-4ff3-b5ba-b4df010ef25f-logs\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.942024 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-config-data\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.942120 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.942155 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.942201 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d9nk\" (UniqueName: \"kubernetes.io/projected/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-kube-api-access-6d9nk\") pod \"nova-cell1-novncproxy-0\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.942219 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.942249 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.942276 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-config-data\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.942305 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97lzg\" (UniqueName: \"kubernetes.io/projected/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-kube-api-access-97lzg\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.942322 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-logs\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.958817 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.992987 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-677484c4fc-lxtjs"] Feb 25 08:48:16 crc kubenswrapper[4978]: I0225 08:48:16.995000 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.005272 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.006490 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.007036 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.008334 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.030178 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-677484c4fc-lxtjs"] Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.033072 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.044842 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.044895 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.044920 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-config-data\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.044941 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-config\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.044961 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-config-data\") pod \"nova-scheduler-0\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.044980 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97lzg\" (UniqueName: \"kubernetes.io/projected/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-kube-api-access-97lzg\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.044995 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-logs\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045020 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-sb\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045058 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-nb\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045076 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhwvz\" (UniqueName: \"kubernetes.io/projected/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-kube-api-access-nhwvz\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045116 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-dns-svc\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045137 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fnr7\" (UniqueName: \"kubernetes.io/projected/598d57b6-5348-4ff3-b5ba-b4df010ef25f-kube-api-access-4fnr7\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045155 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjs6g\" (UniqueName: \"kubernetes.io/projected/386f01a1-7e55-498a-adda-977e1ada8872-kube-api-access-mjs6g\") pod \"nova-scheduler-0\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045171 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/598d57b6-5348-4ff3-b5ba-b4df010ef25f-logs\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045189 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-config-data\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045206 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045228 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045247 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.045279 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d9nk\" (UniqueName: \"kubernetes.io/projected/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-kube-api-access-6d9nk\") pod \"nova-cell1-novncproxy-0\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.047897 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/598d57b6-5348-4ff3-b5ba-b4df010ef25f-logs\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.053002 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-logs\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.057088 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.058382 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-config-data\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.063180 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.066931 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.067582 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-config-data\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.068661 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.077773 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d9nk\" (UniqueName: \"kubernetes.io/projected/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-kube-api-access-6d9nk\") pod \"nova-cell1-novncproxy-0\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.080700 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97lzg\" (UniqueName: \"kubernetes.io/projected/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-kube-api-access-97lzg\") pod \"nova-metadata-0\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " pod="openstack/nova-metadata-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.081762 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fnr7\" (UniqueName: \"kubernetes.io/projected/598d57b6-5348-4ff3-b5ba-b4df010ef25f-kube-api-access-4fnr7\") pod \"nova-api-0\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " pod="openstack/nova-api-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.147248 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.147379 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-config\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.147407 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-config-data\") pod \"nova-scheduler-0\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.147454 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-sb\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.147492 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-nb\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.147509 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhwvz\" (UniqueName: \"kubernetes.io/projected/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-kube-api-access-nhwvz\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.147548 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-dns-svc\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.147567 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjs6g\" (UniqueName: \"kubernetes.io/projected/386f01a1-7e55-498a-adda-977e1ada8872-kube-api-access-mjs6g\") pod \"nova-scheduler-0\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.149721 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-config\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.150608 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-nb\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.151122 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-sb\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.151911 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-dns-svc\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.158796 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.162946 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-config-data\") pod \"nova-scheduler-0\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.167350 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjs6g\" (UniqueName: \"kubernetes.io/projected/386f01a1-7e55-498a-adda-977e1ada8872-kube-api-access-mjs6g\") pod \"nova-scheduler-0\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.168268 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhwvz\" (UniqueName: \"kubernetes.io/projected/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-kube-api-access-nhwvz\") pod \"dnsmasq-dns-677484c4fc-lxtjs\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.249298 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.277050 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.285112 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.332748 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.383565 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.534730 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-6lvjg"] Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.584288 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-mfd2w"] Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.585927 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.590885 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.591160 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.605445 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-mfd2w"] Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.687247 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-scripts\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.687663 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trgpf\" (UniqueName: \"kubernetes.io/projected/5cf80876-ea3d-4885-90fa-cd977f45aee4-kube-api-access-trgpf\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.687881 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-config-data\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.688009 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.693019 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.697929 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6lvjg" event={"ID":"55021e7b-713c-4697-a80b-49e0a23cff45","Type":"ContainerStarted","Data":"da45351cc69bea9cab3e6cca9ef2e876f2639df065a5ceefaa289b7db1040442"} Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.790329 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.790478 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-scripts\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.790516 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trgpf\" (UniqueName: \"kubernetes.io/projected/5cf80876-ea3d-4885-90fa-cd977f45aee4-kube-api-access-trgpf\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.790568 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-config-data\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.796203 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-config-data\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.798536 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.800000 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-scripts\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.816961 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trgpf\" (UniqueName: \"kubernetes.io/projected/5cf80876-ea3d-4885-90fa-cd977f45aee4-kube-api-access-trgpf\") pod \"nova-cell1-conductor-db-sync-mfd2w\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.912956 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.927892 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-677484c4fc-lxtjs"] Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.938437 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 08:48:17 crc kubenswrapper[4978]: I0225 08:48:17.975568 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.062828 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:48:18 crc kubenswrapper[4978]: W0225 08:48:18.093615 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod386f01a1_7e55_498a_adda_977e1ada8872.slice/crio-1f8e69bf38c24d1e93b4a09573b8d3b9f74b2df859e1387cc95dff44f1394fcc WatchSource:0}: Error finding container 1f8e69bf38c24d1e93b4a09573b8d3b9f74b2df859e1387cc95dff44f1394fcc: Status 404 returned error can't find the container with id 1f8e69bf38c24d1e93b4a09573b8d3b9f74b2df859e1387cc95dff44f1394fcc Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.327807 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.420170 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-mfd2w"] Feb 25 08:48:18 crc kubenswrapper[4978]: W0225 08:48:18.424068 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cf80876_ea3d_4885_90fa_cd977f45aee4.slice/crio-572843bb28d6a4e7bb32206bda7eb5cff69dfc6968e1be28a5fc33a81fece70c WatchSource:0}: Error finding container 572843bb28d6a4e7bb32206bda7eb5cff69dfc6968e1be28a5fc33a81fece70c: Status 404 returned error can't find the container with id 572843bb28d6a4e7bb32206bda7eb5cff69dfc6968e1be28a5fc33a81fece70c Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.712310 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1ca54c-26b2-4635-9dd1-20b2eeab454b","Type":"ContainerStarted","Data":"52892fa67915d7338137b9c5375c2d1305ba042cdd92a102d700f5a30ab242ff"} Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.720717 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-mfd2w" event={"ID":"5cf80876-ea3d-4885-90fa-cd977f45aee4","Type":"ContainerStarted","Data":"751f9625f412705ac75b12905ac0ccb115dc26186b4e8a15eae405a4eefba4e8"} Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.720760 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-mfd2w" event={"ID":"5cf80876-ea3d-4885-90fa-cd977f45aee4","Type":"ContainerStarted","Data":"572843bb28d6a4e7bb32206bda7eb5cff69dfc6968e1be28a5fc33a81fece70c"} Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.725069 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6lvjg" event={"ID":"55021e7b-713c-4697-a80b-49e0a23cff45","Type":"ContainerStarted","Data":"a9a430476c450bdbcda78400c12c520b9944453552bd31acb86f41886183e9a2"} Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.729336 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"386f01a1-7e55-498a-adda-977e1ada8872","Type":"ContainerStarted","Data":"1f8e69bf38c24d1e93b4a09573b8d3b9f74b2df859e1387cc95dff44f1394fcc"} Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.735449 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"c572bfb7b460a06e6b3fd8562f5f892b5fdb169c89eb0df815351c3f11f58b78"} Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.736148 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-mfd2w" podStartSLOduration=1.736129895 podStartE2EDuration="1.736129895s" podCreationTimestamp="2026-02-25 08:48:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:48:18.731936775 +0000 UTC m=+7392.171193234" watchObservedRunningTime="2026-02-25 08:48:18.736129895 +0000 UTC m=+7392.175386354" Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.741336 4978 generic.go:334] "Generic (PLEG): container finished" podID="ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" containerID="9e36a0cfa136b732c763f3b9118af15510d98e98d67ea2917b0b99b05ab61ded" exitCode=0 Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.741459 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" event={"ID":"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8","Type":"ContainerDied","Data":"9e36a0cfa136b732c763f3b9118af15510d98e98d67ea2917b0b99b05ab61ded"} Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.741513 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" event={"ID":"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8","Type":"ContainerStarted","Data":"5947db7210ae49dca8dbcc6d90a3386aa8ea309e8c24c14842095de5d3876b0a"} Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.742705 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"598d57b6-5348-4ff3-b5ba-b4df010ef25f","Type":"ContainerStarted","Data":"3d7d28139723b25d70b7082fee577c7377865d3177acef8639cf9ce9d385f10b"} Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.750953 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"caf834ac-cc3b-4728-b4c7-bd19abd9a82a","Type":"ContainerStarted","Data":"fcd95ab490fba2c35b13104e3a02d27b1acbbbc83ba32387f007fa29e627c0e4"} Feb 25 08:48:18 crc kubenswrapper[4978]: I0225 08:48:18.790816 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-6lvjg" podStartSLOduration=2.790796289 podStartE2EDuration="2.790796289s" podCreationTimestamp="2026-02-25 08:48:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:48:18.753598216 +0000 UTC m=+7392.192854705" watchObservedRunningTime="2026-02-25 08:48:18.790796289 +0000 UTC m=+7392.230052748" Feb 25 08:48:20 crc kubenswrapper[4978]: I0225 08:48:20.784489 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" event={"ID":"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8","Type":"ContainerStarted","Data":"2856f9525243e46948863bde61e1a085af07ea2bf627e25e0b1b1a6ee119a68d"} Feb 25 08:48:20 crc kubenswrapper[4978]: I0225 08:48:20.784997 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:20 crc kubenswrapper[4978]: I0225 08:48:20.805216 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" podStartSLOduration=4.8051987369999996 podStartE2EDuration="4.805198737s" podCreationTimestamp="2026-02-25 08:48:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:48:20.802438331 +0000 UTC m=+7394.241694790" watchObservedRunningTime="2026-02-25 08:48:20.805198737 +0000 UTC m=+7394.244455196" Feb 25 08:48:21 crc kubenswrapper[4978]: I0225 08:48:21.352245 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 08:48:21 crc kubenswrapper[4978]: I0225 08:48:21.352286 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:21 crc kubenswrapper[4978]: I0225 08:48:21.803521 4978 generic.go:334] "Generic (PLEG): container finished" podID="5cf80876-ea3d-4885-90fa-cd977f45aee4" containerID="751f9625f412705ac75b12905ac0ccb115dc26186b4e8a15eae405a4eefba4e8" exitCode=0 Feb 25 08:48:21 crc kubenswrapper[4978]: I0225 08:48:21.804568 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-mfd2w" event={"ID":"5cf80876-ea3d-4885-90fa-cd977f45aee4","Type":"ContainerDied","Data":"751f9625f412705ac75b12905ac0ccb115dc26186b4e8a15eae405a4eefba4e8"} Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.814540 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1ca54c-26b2-4635-9dd1-20b2eeab454b","Type":"ContainerStarted","Data":"759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa"} Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.814802 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1ca54c-26b2-4635-9dd1-20b2eeab454b","Type":"ContainerStarted","Data":"92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10"} Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.814743 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" containerName="nova-metadata-metadata" containerID="cri-o://759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa" gracePeriod=30 Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.814658 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" containerName="nova-metadata-log" containerID="cri-o://92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10" gracePeriod=30 Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.816668 4978 generic.go:334] "Generic (PLEG): container finished" podID="55021e7b-713c-4697-a80b-49e0a23cff45" containerID="a9a430476c450bdbcda78400c12c520b9944453552bd31acb86f41886183e9a2" exitCode=0 Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.816696 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6lvjg" event={"ID":"55021e7b-713c-4697-a80b-49e0a23cff45","Type":"ContainerDied","Data":"a9a430476c450bdbcda78400c12c520b9944453552bd31acb86f41886183e9a2"} Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.818176 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"386f01a1-7e55-498a-adda-977e1ada8872","Type":"ContainerStarted","Data":"c8465ec81edabb8d55345d8e1a838840f4c03b5d88087767bd4016e883901e12"} Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.823868 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"598d57b6-5348-4ff3-b5ba-b4df010ef25f","Type":"ContainerStarted","Data":"b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee"} Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.823924 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"598d57b6-5348-4ff3-b5ba-b4df010ef25f","Type":"ContainerStarted","Data":"0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1"} Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.826105 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"caf834ac-cc3b-4728-b4c7-bd19abd9a82a","Type":"ContainerStarted","Data":"b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64"} Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.826268 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="caf834ac-cc3b-4728-b4c7-bd19abd9a82a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64" gracePeriod=30 Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.847829 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.803239665 podStartE2EDuration="6.847809379s" podCreationTimestamp="2026-02-25 08:48:16 +0000 UTC" firstStartedPulling="2026-02-25 08:48:17.981777955 +0000 UTC m=+7391.421034414" lastFinishedPulling="2026-02-25 08:48:22.026347659 +0000 UTC m=+7395.465604128" observedRunningTime="2026-02-25 08:48:22.835771746 +0000 UTC m=+7396.275028205" watchObservedRunningTime="2026-02-25 08:48:22.847809379 +0000 UTC m=+7396.287065838" Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.861053 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.932552811 podStartE2EDuration="6.861038189s" podCreationTimestamp="2026-02-25 08:48:16 +0000 UTC" firstStartedPulling="2026-02-25 08:48:18.097338565 +0000 UTC m=+7391.536595014" lastFinishedPulling="2026-02-25 08:48:22.025823933 +0000 UTC m=+7395.465080392" observedRunningTime="2026-02-25 08:48:22.856786367 +0000 UTC m=+7396.296042836" watchObservedRunningTime="2026-02-25 08:48:22.861038189 +0000 UTC m=+7396.300294648" Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.889624 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.832513731 podStartE2EDuration="6.889606234s" podCreationTimestamp="2026-02-25 08:48:16 +0000 UTC" firstStartedPulling="2026-02-25 08:48:17.967519653 +0000 UTC m=+7391.406776112" lastFinishedPulling="2026-02-25 08:48:22.024612156 +0000 UTC m=+7395.463868615" observedRunningTime="2026-02-25 08:48:22.878674325 +0000 UTC m=+7396.317930794" watchObservedRunningTime="2026-02-25 08:48:22.889606234 +0000 UTC m=+7396.328862693" Feb 25 08:48:22 crc kubenswrapper[4978]: I0225 08:48:22.905029 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.585530679 podStartE2EDuration="6.905013001s" podCreationTimestamp="2026-02-25 08:48:16 +0000 UTC" firstStartedPulling="2026-02-25 08:48:17.713647308 +0000 UTC m=+7391.152903767" lastFinishedPulling="2026-02-25 08:48:22.03312959 +0000 UTC m=+7395.472386089" observedRunningTime="2026-02-25 08:48:22.899148889 +0000 UTC m=+7396.338405348" watchObservedRunningTime="2026-02-25 08:48:22.905013001 +0000 UTC m=+7396.344269460" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.197059 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.312478 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-combined-ca-bundle\") pod \"5cf80876-ea3d-4885-90fa-cd977f45aee4\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.312621 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-config-data\") pod \"5cf80876-ea3d-4885-90fa-cd977f45aee4\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.312784 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trgpf\" (UniqueName: \"kubernetes.io/projected/5cf80876-ea3d-4885-90fa-cd977f45aee4-kube-api-access-trgpf\") pod \"5cf80876-ea3d-4885-90fa-cd977f45aee4\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.312825 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-scripts\") pod \"5cf80876-ea3d-4885-90fa-cd977f45aee4\" (UID: \"5cf80876-ea3d-4885-90fa-cd977f45aee4\") " Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.319485 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-scripts" (OuterVolumeSpecName: "scripts") pod "5cf80876-ea3d-4885-90fa-cd977f45aee4" (UID: "5cf80876-ea3d-4885-90fa-cd977f45aee4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.319652 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cf80876-ea3d-4885-90fa-cd977f45aee4-kube-api-access-trgpf" (OuterVolumeSpecName: "kube-api-access-trgpf") pod "5cf80876-ea3d-4885-90fa-cd977f45aee4" (UID: "5cf80876-ea3d-4885-90fa-cd977f45aee4"). InnerVolumeSpecName "kube-api-access-trgpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.342424 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cf80876-ea3d-4885-90fa-cd977f45aee4" (UID: "5cf80876-ea3d-4885-90fa-cd977f45aee4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.343923 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-config-data" (OuterVolumeSpecName: "config-data") pod "5cf80876-ea3d-4885-90fa-cd977f45aee4" (UID: "5cf80876-ea3d-4885-90fa-cd977f45aee4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.414995 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trgpf\" (UniqueName: \"kubernetes.io/projected/5cf80876-ea3d-4885-90fa-cd977f45aee4-kube-api-access-trgpf\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.415031 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.415045 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.415059 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf80876-ea3d-4885-90fa-cd977f45aee4-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.420397 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.516244 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-combined-ca-bundle\") pod \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.516539 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-logs\") pod \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.516664 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97lzg\" (UniqueName: \"kubernetes.io/projected/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-kube-api-access-97lzg\") pod \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.516786 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-config-data\") pod \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\" (UID: \"ec1ca54c-26b2-4635-9dd1-20b2eeab454b\") " Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.516968 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-logs" (OuterVolumeSpecName: "logs") pod "ec1ca54c-26b2-4635-9dd1-20b2eeab454b" (UID: "ec1ca54c-26b2-4635-9dd1-20b2eeab454b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.517580 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.521418 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-kube-api-access-97lzg" (OuterVolumeSpecName: "kube-api-access-97lzg") pod "ec1ca54c-26b2-4635-9dd1-20b2eeab454b" (UID: "ec1ca54c-26b2-4635-9dd1-20b2eeab454b"). InnerVolumeSpecName "kube-api-access-97lzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.540545 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec1ca54c-26b2-4635-9dd1-20b2eeab454b" (UID: "ec1ca54c-26b2-4635-9dd1-20b2eeab454b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.541867 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-config-data" (OuterVolumeSpecName: "config-data") pod "ec1ca54c-26b2-4635-9dd1-20b2eeab454b" (UID: "ec1ca54c-26b2-4635-9dd1-20b2eeab454b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.620175 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97lzg\" (UniqueName: \"kubernetes.io/projected/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-kube-api-access-97lzg\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.620226 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.620246 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec1ca54c-26b2-4635-9dd1-20b2eeab454b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.838951 4978 generic.go:334] "Generic (PLEG): container finished" podID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" containerID="759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa" exitCode=0 Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.840201 4978 generic.go:334] "Generic (PLEG): container finished" podID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" containerID="92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10" exitCode=143 Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.840459 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1ca54c-26b2-4635-9dd1-20b2eeab454b","Type":"ContainerDied","Data":"759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa"} Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.840627 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1ca54c-26b2-4635-9dd1-20b2eeab454b","Type":"ContainerDied","Data":"92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10"} Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.840763 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec1ca54c-26b2-4635-9dd1-20b2eeab454b","Type":"ContainerDied","Data":"52892fa67915d7338137b9c5375c2d1305ba042cdd92a102d700f5a30ab242ff"} Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.840898 4978 scope.go:117] "RemoveContainer" containerID="759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.841190 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.850671 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-mfd2w" event={"ID":"5cf80876-ea3d-4885-90fa-cd977f45aee4","Type":"ContainerDied","Data":"572843bb28d6a4e7bb32206bda7eb5cff69dfc6968e1be28a5fc33a81fece70c"} Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.850736 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="572843bb28d6a4e7bb32206bda7eb5cff69dfc6968e1be28a5fc33a81fece70c" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.850771 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-mfd2w" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.879628 4978 scope.go:117] "RemoveContainer" containerID="92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.906018 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.921523 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.930667 4978 scope.go:117] "RemoveContainer" containerID="759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa" Feb 25 08:48:23 crc kubenswrapper[4978]: E0225 08:48:23.931081 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa\": container with ID starting with 759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa not found: ID does not exist" containerID="759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.931107 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa"} err="failed to get container status \"759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa\": rpc error: code = NotFound desc = could not find container \"759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa\": container with ID starting with 759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa not found: ID does not exist" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.931125 4978 scope.go:117] "RemoveContainer" containerID="92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10" Feb 25 08:48:23 crc kubenswrapper[4978]: E0225 08:48:23.933258 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10\": container with ID starting with 92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10 not found: ID does not exist" containerID="92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.933286 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10"} err="failed to get container status \"92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10\": rpc error: code = NotFound desc = could not find container \"92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10\": container with ID starting with 92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10 not found: ID does not exist" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.933303 4978 scope.go:117] "RemoveContainer" containerID="759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.934805 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa"} err="failed to get container status \"759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa\": rpc error: code = NotFound desc = could not find container \"759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa\": container with ID starting with 759b8812af79b2bd7dab10ce86f4ddf0c861cc103bd3e53c0d796ff0f52960aa not found: ID does not exist" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.934827 4978 scope.go:117] "RemoveContainer" containerID="92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.938440 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10"} err="failed to get container status \"92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10\": rpc error: code = NotFound desc = could not find container \"92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10\": container with ID starting with 92c0d919845d0dd8fbd88800fd3a22c40f66954c93adf602ef7878409a586a10 not found: ID does not exist" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.941397 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 08:48:23 crc kubenswrapper[4978]: E0225 08:48:23.941942 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf80876-ea3d-4885-90fa-cd977f45aee4" containerName="nova-cell1-conductor-db-sync" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.941960 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf80876-ea3d-4885-90fa-cd977f45aee4" containerName="nova-cell1-conductor-db-sync" Feb 25 08:48:23 crc kubenswrapper[4978]: E0225 08:48:23.941988 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" containerName="nova-metadata-metadata" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.941996 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" containerName="nova-metadata-metadata" Feb 25 08:48:23 crc kubenswrapper[4978]: E0225 08:48:23.942045 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" containerName="nova-metadata-log" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.942052 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" containerName="nova-metadata-log" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.942283 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cf80876-ea3d-4885-90fa-cd977f45aee4" containerName="nova-cell1-conductor-db-sync" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.942301 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" containerName="nova-metadata-metadata" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.942313 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" containerName="nova-metadata-log" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.943144 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.947675 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.966182 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.968170 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.971953 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.972154 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.976648 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 08:48:23 crc kubenswrapper[4978]: I0225 08:48:23.989586 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.031530 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63ec4e73-ad23-451a-9894-e79c25fdf76f-logs\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.031568 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.031787 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.031868 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-config-data\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.032026 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.032111 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d4tv\" (UniqueName: \"kubernetes.io/projected/63ec4e73-ad23-451a-9894-e79c25fdf76f-kube-api-access-7d4tv\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.032223 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh85w\" (UniqueName: \"kubernetes.io/projected/57345093-7635-4852-bf51-5f0610196ab3-kube-api-access-qh85w\") pod \"nova-cell1-conductor-0\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.032289 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.135020 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63ec4e73-ad23-451a-9894-e79c25fdf76f-logs\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.135054 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.135087 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.135107 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-config-data\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.135143 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.135193 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d4tv\" (UniqueName: \"kubernetes.io/projected/63ec4e73-ad23-451a-9894-e79c25fdf76f-kube-api-access-7d4tv\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.135226 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh85w\" (UniqueName: \"kubernetes.io/projected/57345093-7635-4852-bf51-5f0610196ab3-kube-api-access-qh85w\") pod \"nova-cell1-conductor-0\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.135248 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.137510 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63ec4e73-ad23-451a-9894-e79c25fdf76f-logs\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.141074 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.149051 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.149172 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.149544 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-config-data\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.149717 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.151132 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d4tv\" (UniqueName: \"kubernetes.io/projected/63ec4e73-ad23-451a-9894-e79c25fdf76f-kube-api-access-7d4tv\") pod \"nova-metadata-0\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.157303 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh85w\" (UniqueName: \"kubernetes.io/projected/57345093-7635-4852-bf51-5f0610196ab3-kube-api-access-qh85w\") pod \"nova-cell1-conductor-0\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.245551 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.287698 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.295939 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.338099 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5dfh\" (UniqueName: \"kubernetes.io/projected/55021e7b-713c-4697-a80b-49e0a23cff45-kube-api-access-z5dfh\") pod \"55021e7b-713c-4697-a80b-49e0a23cff45\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.338149 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-combined-ca-bundle\") pod \"55021e7b-713c-4697-a80b-49e0a23cff45\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.338183 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-scripts\") pod \"55021e7b-713c-4697-a80b-49e0a23cff45\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.338388 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-config-data\") pod \"55021e7b-713c-4697-a80b-49e0a23cff45\" (UID: \"55021e7b-713c-4697-a80b-49e0a23cff45\") " Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.345767 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55021e7b-713c-4697-a80b-49e0a23cff45-kube-api-access-z5dfh" (OuterVolumeSpecName: "kube-api-access-z5dfh") pod "55021e7b-713c-4697-a80b-49e0a23cff45" (UID: "55021e7b-713c-4697-a80b-49e0a23cff45"). InnerVolumeSpecName "kube-api-access-z5dfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.346938 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-scripts" (OuterVolumeSpecName: "scripts") pod "55021e7b-713c-4697-a80b-49e0a23cff45" (UID: "55021e7b-713c-4697-a80b-49e0a23cff45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.373868 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-config-data" (OuterVolumeSpecName: "config-data") pod "55021e7b-713c-4697-a80b-49e0a23cff45" (UID: "55021e7b-713c-4697-a80b-49e0a23cff45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.374700 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55021e7b-713c-4697-a80b-49e0a23cff45" (UID: "55021e7b-713c-4697-a80b-49e0a23cff45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.441763 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.441809 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5dfh\" (UniqueName: \"kubernetes.io/projected/55021e7b-713c-4697-a80b-49e0a23cff45-kube-api-access-z5dfh\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.441832 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.441851 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55021e7b-713c-4697-a80b-49e0a23cff45-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.762894 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.845338 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 08:48:24 crc kubenswrapper[4978]: W0225 08:48:24.850153 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57345093_7635_4852_bf51_5f0610196ab3.slice/crio-4b0a930d88c1f1a5e447360c5190cb1c46abb0fc034bb84801375adf1bdd5d42 WatchSource:0}: Error finding container 4b0a930d88c1f1a5e447360c5190cb1c46abb0fc034bb84801375adf1bdd5d42: Status 404 returned error can't find the container with id 4b0a930d88c1f1a5e447360c5190cb1c46abb0fc034bb84801375adf1bdd5d42 Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.861127 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6lvjg" event={"ID":"55021e7b-713c-4697-a80b-49e0a23cff45","Type":"ContainerDied","Data":"da45351cc69bea9cab3e6cca9ef2e876f2639df065a5ceefaa289b7db1040442"} Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.861169 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da45351cc69bea9cab3e6cca9ef2e876f2639df065a5ceefaa289b7db1040442" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.861242 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6lvjg" Feb 25 08:48:24 crc kubenswrapper[4978]: I0225 08:48:24.863753 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63ec4e73-ad23-451a-9894-e79c25fdf76f","Type":"ContainerStarted","Data":"ee6ec20bb726b8f6486576220eae137b7a21ff1472efe113afbde5f62b291c31"} Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.116274 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.116659 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" containerName="nova-api-log" containerID="cri-o://0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1" gracePeriod=30 Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.117281 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" containerName="nova-api-api" containerID="cri-o://b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee" gracePeriod=30 Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.131595 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.131960 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="386f01a1-7e55-498a-adda-977e1ada8872" containerName="nova-scheduler-scheduler" containerID="cri-o://c8465ec81edabb8d55345d8e1a838840f4c03b5d88087767bd4016e883901e12" gracePeriod=30 Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.148791 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.337182 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec1ca54c-26b2-4635-9dd1-20b2eeab454b" path="/var/lib/kubelet/pods/ec1ca54c-26b2-4635-9dd1-20b2eeab454b/volumes" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.678753 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.766994 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fnr7\" (UniqueName: \"kubernetes.io/projected/598d57b6-5348-4ff3-b5ba-b4df010ef25f-kube-api-access-4fnr7\") pod \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.767364 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-config-data\") pod \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.767413 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/598d57b6-5348-4ff3-b5ba-b4df010ef25f-logs\") pod \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.767518 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-combined-ca-bundle\") pod \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\" (UID: \"598d57b6-5348-4ff3-b5ba-b4df010ef25f\") " Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.768319 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/598d57b6-5348-4ff3-b5ba-b4df010ef25f-logs" (OuterVolumeSpecName: "logs") pod "598d57b6-5348-4ff3-b5ba-b4df010ef25f" (UID: "598d57b6-5348-4ff3-b5ba-b4df010ef25f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.773728 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/598d57b6-5348-4ff3-b5ba-b4df010ef25f-kube-api-access-4fnr7" (OuterVolumeSpecName: "kube-api-access-4fnr7") pod "598d57b6-5348-4ff3-b5ba-b4df010ef25f" (UID: "598d57b6-5348-4ff3-b5ba-b4df010ef25f"). InnerVolumeSpecName "kube-api-access-4fnr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.797945 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-config-data" (OuterVolumeSpecName: "config-data") pod "598d57b6-5348-4ff3-b5ba-b4df010ef25f" (UID: "598d57b6-5348-4ff3-b5ba-b4df010ef25f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.820240 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "598d57b6-5348-4ff3-b5ba-b4df010ef25f" (UID: "598d57b6-5348-4ff3-b5ba-b4df010ef25f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.869443 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fnr7\" (UniqueName: \"kubernetes.io/projected/598d57b6-5348-4ff3-b5ba-b4df010ef25f-kube-api-access-4fnr7\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.869476 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.869491 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/598d57b6-5348-4ff3-b5ba-b4df010ef25f-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.869502 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/598d57b6-5348-4ff3-b5ba-b4df010ef25f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.879037 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63ec4e73-ad23-451a-9894-e79c25fdf76f","Type":"ContainerStarted","Data":"99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830"} Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.879091 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63ec4e73-ad23-451a-9894-e79c25fdf76f","Type":"ContainerStarted","Data":"a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17"} Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.879888 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="63ec4e73-ad23-451a-9894-e79c25fdf76f" containerName="nova-metadata-log" containerID="cri-o://a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17" gracePeriod=30 Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.880669 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="63ec4e73-ad23-451a-9894-e79c25fdf76f" containerName="nova-metadata-metadata" containerID="cri-o://99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830" gracePeriod=30 Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.895667 4978 generic.go:334] "Generic (PLEG): container finished" podID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" containerID="b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee" exitCode=0 Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.895706 4978 generic.go:334] "Generic (PLEG): container finished" podID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" containerID="0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1" exitCode=143 Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.895750 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"598d57b6-5348-4ff3-b5ba-b4df010ef25f","Type":"ContainerDied","Data":"b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee"} Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.895780 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"598d57b6-5348-4ff3-b5ba-b4df010ef25f","Type":"ContainerDied","Data":"0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1"} Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.895795 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"598d57b6-5348-4ff3-b5ba-b4df010ef25f","Type":"ContainerDied","Data":"3d7d28139723b25d70b7082fee577c7377865d3177acef8639cf9ce9d385f10b"} Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.895812 4978 scope.go:117] "RemoveContainer" containerID="b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.895983 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.904338 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.904323172 podStartE2EDuration="2.904323172s" podCreationTimestamp="2026-02-25 08:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:48:25.901858946 +0000 UTC m=+7399.341115425" watchObservedRunningTime="2026-02-25 08:48:25.904323172 +0000 UTC m=+7399.343579631" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.905408 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"57345093-7635-4852-bf51-5f0610196ab3","Type":"ContainerStarted","Data":"660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2"} Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.905445 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"57345093-7635-4852-bf51-5f0610196ab3","Type":"ContainerStarted","Data":"4b0a930d88c1f1a5e447360c5190cb1c46abb0fc034bb84801375adf1bdd5d42"} Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.905727 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.931204 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.931181044 podStartE2EDuration="2.931181044s" podCreationTimestamp="2026-02-25 08:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:48:25.922930118 +0000 UTC m=+7399.362186577" watchObservedRunningTime="2026-02-25 08:48:25.931181044 +0000 UTC m=+7399.370437503" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.938963 4978 scope.go:117] "RemoveContainer" containerID="0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.950902 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.954616 4978 scope.go:117] "RemoveContainer" containerID="b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee" Feb 25 08:48:25 crc kubenswrapper[4978]: E0225 08:48:25.956168 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee\": container with ID starting with b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee not found: ID does not exist" containerID="b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.956207 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee"} err="failed to get container status \"b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee\": rpc error: code = NotFound desc = could not find container \"b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee\": container with ID starting with b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee not found: ID does not exist" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.956229 4978 scope.go:117] "RemoveContainer" containerID="0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1" Feb 25 08:48:25 crc kubenswrapper[4978]: E0225 08:48:25.956506 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1\": container with ID starting with 0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1 not found: ID does not exist" containerID="0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.956535 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1"} err="failed to get container status \"0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1\": rpc error: code = NotFound desc = could not find container \"0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1\": container with ID starting with 0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1 not found: ID does not exist" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.956550 4978 scope.go:117] "RemoveContainer" containerID="b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.956760 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee"} err="failed to get container status \"b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee\": rpc error: code = NotFound desc = could not find container \"b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee\": container with ID starting with b4ede9ed6e15469f55c4f4d329b3a68ac1acd760971b429dfddb7cacbc4cffee not found: ID does not exist" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.956784 4978 scope.go:117] "RemoveContainer" containerID="0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.956995 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1"} err="failed to get container status \"0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1\": rpc error: code = NotFound desc = could not find container \"0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1\": container with ID starting with 0813cbe5771d05ad213fe0010fe847838b0d366c54630e05c218e7c7a27718a1 not found: ID does not exist" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.965849 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.985517 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 08:48:25 crc kubenswrapper[4978]: E0225 08:48:25.985924 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" containerName="nova-api-log" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.985940 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" containerName="nova-api-log" Feb 25 08:48:25 crc kubenswrapper[4978]: E0225 08:48:25.985960 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" containerName="nova-api-api" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.985965 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" containerName="nova-api-api" Feb 25 08:48:25 crc kubenswrapper[4978]: E0225 08:48:25.985977 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55021e7b-713c-4697-a80b-49e0a23cff45" containerName="nova-manage" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.985985 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="55021e7b-713c-4697-a80b-49e0a23cff45" containerName="nova-manage" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.986149 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" containerName="nova-api-api" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.986166 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="55021e7b-713c-4697-a80b-49e0a23cff45" containerName="nova-manage" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.986181 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" containerName="nova-api-log" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.987114 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:48:25 crc kubenswrapper[4978]: I0225 08:48:25.989664 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.020211 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.072975 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.073037 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjhlv\" (UniqueName: \"kubernetes.io/projected/e5136c0c-233c-4454-a076-d26cd2538149-kube-api-access-qjhlv\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.073108 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5136c0c-233c-4454-a076-d26cd2538149-logs\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.073136 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-config-data\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.175024 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5136c0c-233c-4454-a076-d26cd2538149-logs\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.175240 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-config-data\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.175454 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.175540 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjhlv\" (UniqueName: \"kubernetes.io/projected/e5136c0c-233c-4454-a076-d26cd2538149-kube-api-access-qjhlv\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.175459 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5136c0c-233c-4454-a076-d26cd2538149-logs\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.181262 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-config-data\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.182167 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.199752 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjhlv\" (UniqueName: \"kubernetes.io/projected/e5136c0c-233c-4454-a076-d26cd2538149-kube-api-access-qjhlv\") pod \"nova-api-0\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.317665 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.412269 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.480848 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d4tv\" (UniqueName: \"kubernetes.io/projected/63ec4e73-ad23-451a-9894-e79c25fdf76f-kube-api-access-7d4tv\") pod \"63ec4e73-ad23-451a-9894-e79c25fdf76f\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.480906 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-combined-ca-bundle\") pod \"63ec4e73-ad23-451a-9894-e79c25fdf76f\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.480963 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63ec4e73-ad23-451a-9894-e79c25fdf76f-logs\") pod \"63ec4e73-ad23-451a-9894-e79c25fdf76f\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.481025 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-config-data\") pod \"63ec4e73-ad23-451a-9894-e79c25fdf76f\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.481106 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-nova-metadata-tls-certs\") pod \"63ec4e73-ad23-451a-9894-e79c25fdf76f\" (UID: \"63ec4e73-ad23-451a-9894-e79c25fdf76f\") " Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.482052 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ec4e73-ad23-451a-9894-e79c25fdf76f-logs" (OuterVolumeSpecName: "logs") pod "63ec4e73-ad23-451a-9894-e79c25fdf76f" (UID: "63ec4e73-ad23-451a-9894-e79c25fdf76f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.485637 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63ec4e73-ad23-451a-9894-e79c25fdf76f-kube-api-access-7d4tv" (OuterVolumeSpecName: "kube-api-access-7d4tv") pod "63ec4e73-ad23-451a-9894-e79c25fdf76f" (UID: "63ec4e73-ad23-451a-9894-e79c25fdf76f"). InnerVolumeSpecName "kube-api-access-7d4tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.510559 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63ec4e73-ad23-451a-9894-e79c25fdf76f" (UID: "63ec4e73-ad23-451a-9894-e79c25fdf76f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.511990 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-config-data" (OuterVolumeSpecName: "config-data") pod "63ec4e73-ad23-451a-9894-e79c25fdf76f" (UID: "63ec4e73-ad23-451a-9894-e79c25fdf76f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.553529 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "63ec4e73-ad23-451a-9894-e79c25fdf76f" (UID: "63ec4e73-ad23-451a-9894-e79c25fdf76f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.583524 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d4tv\" (UniqueName: \"kubernetes.io/projected/63ec4e73-ad23-451a-9894-e79c25fdf76f-kube-api-access-7d4tv\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.583759 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.583843 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63ec4e73-ad23-451a-9894-e79c25fdf76f-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.583909 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.583976 4978 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63ec4e73-ad23-451a-9894-e79c25fdf76f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.758744 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:48:26 crc kubenswrapper[4978]: W0225 08:48:26.764914 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5136c0c_233c_4454_a076_d26cd2538149.slice/crio-80a7ee599321ba51dd990251a2d52f506e01252d39a500dd0ee2338b7666a8b1 WatchSource:0}: Error finding container 80a7ee599321ba51dd990251a2d52f506e01252d39a500dd0ee2338b7666a8b1: Status 404 returned error can't find the container with id 80a7ee599321ba51dd990251a2d52f506e01252d39a500dd0ee2338b7666a8b1 Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.921786 4978 generic.go:334] "Generic (PLEG): container finished" podID="63ec4e73-ad23-451a-9894-e79c25fdf76f" containerID="99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830" exitCode=0 Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.922044 4978 generic.go:334] "Generic (PLEG): container finished" podID="63ec4e73-ad23-451a-9894-e79c25fdf76f" containerID="a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17" exitCode=143 Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.922079 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63ec4e73-ad23-451a-9894-e79c25fdf76f","Type":"ContainerDied","Data":"99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830"} Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.922104 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63ec4e73-ad23-451a-9894-e79c25fdf76f","Type":"ContainerDied","Data":"a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17"} Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.922114 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63ec4e73-ad23-451a-9894-e79c25fdf76f","Type":"ContainerDied","Data":"ee6ec20bb726b8f6486576220eae137b7a21ff1472efe113afbde5f62b291c31"} Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.922130 4978 scope.go:117] "RemoveContainer" containerID="99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.922218 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.932847 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5136c0c-233c-4454-a076-d26cd2538149","Type":"ContainerStarted","Data":"80a7ee599321ba51dd990251a2d52f506e01252d39a500dd0ee2338b7666a8b1"} Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.947457 4978 scope.go:117] "RemoveContainer" containerID="a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17" Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.959235 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.967568 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:26 crc kubenswrapper[4978]: I0225 08:48:26.999845 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:27 crc kubenswrapper[4978]: E0225 08:48:27.000231 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ec4e73-ad23-451a-9894-e79c25fdf76f" containerName="nova-metadata-metadata" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.000250 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ec4e73-ad23-451a-9894-e79c25fdf76f" containerName="nova-metadata-metadata" Feb 25 08:48:27 crc kubenswrapper[4978]: E0225 08:48:27.000270 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ec4e73-ad23-451a-9894-e79c25fdf76f" containerName="nova-metadata-log" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.000277 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ec4e73-ad23-451a-9894-e79c25fdf76f" containerName="nova-metadata-log" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.000529 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ec4e73-ad23-451a-9894-e79c25fdf76f" containerName="nova-metadata-log" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.000559 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ec4e73-ad23-451a-9894-e79c25fdf76f" containerName="nova-metadata-metadata" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.001531 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.003077 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.010141 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.030570 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.046695 4978 scope.go:117] "RemoveContainer" containerID="99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830" Feb 25 08:48:27 crc kubenswrapper[4978]: E0225 08:48:27.047239 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830\": container with ID starting with 99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830 not found: ID does not exist" containerID="99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.047277 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830"} err="failed to get container status \"99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830\": rpc error: code = NotFound desc = could not find container \"99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830\": container with ID starting with 99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830 not found: ID does not exist" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.047298 4978 scope.go:117] "RemoveContainer" containerID="a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17" Feb 25 08:48:27 crc kubenswrapper[4978]: E0225 08:48:27.047786 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17\": container with ID starting with a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17 not found: ID does not exist" containerID="a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.047824 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17"} err="failed to get container status \"a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17\": rpc error: code = NotFound desc = could not find container \"a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17\": container with ID starting with a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17 not found: ID does not exist" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.047846 4978 scope.go:117] "RemoveContainer" containerID="99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.048176 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830"} err="failed to get container status \"99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830\": rpc error: code = NotFound desc = could not find container \"99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830\": container with ID starting with 99862ca40a69f951bacac7e22aa77428b725672fac7f683b06508959ff828830 not found: ID does not exist" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.048232 4978 scope.go:117] "RemoveContainer" containerID="a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.048772 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17"} err="failed to get container status \"a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17\": rpc error: code = NotFound desc = could not find container \"a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17\": container with ID starting with a2eafaa37a7773ff35641ec14285572b81bad1e6a1598b3a5ca239ac13ea7a17 not found: ID does not exist" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.098746 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.098873 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.099122 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-config-data\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.099278 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkrxj\" (UniqueName: \"kubernetes.io/projected/c3c2da42-6edb-4b26-9708-9d08c61190fe-kube-api-access-fkrxj\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.099338 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3c2da42-6edb-4b26-9708-9d08c61190fe-logs\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.202093 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.202315 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.202600 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-config-data\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.202842 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkrxj\" (UniqueName: \"kubernetes.io/projected/c3c2da42-6edb-4b26-9708-9d08c61190fe-kube-api-access-fkrxj\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.203015 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3c2da42-6edb-4b26-9708-9d08c61190fe-logs\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.203758 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3c2da42-6edb-4b26-9708-9d08c61190fe-logs\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.207340 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-config-data\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.207433 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.207905 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.223162 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkrxj\" (UniqueName: \"kubernetes.io/projected/c3c2da42-6edb-4b26-9708-9d08c61190fe-kube-api-access-fkrxj\") pod \"nova-metadata-0\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.285544 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.333188 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.337174 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="598d57b6-5348-4ff3-b5ba-b4df010ef25f" path="/var/lib/kubelet/pods/598d57b6-5348-4ff3-b5ba-b4df010ef25f/volumes" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.337900 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63ec4e73-ad23-451a-9894-e79c25fdf76f" path="/var/lib/kubelet/pods/63ec4e73-ad23-451a-9894-e79c25fdf76f/volumes" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.338486 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.385064 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.476188 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778bcd845c-fw9q9"] Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.476482 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" podUID="b985867e-5709-4c7d-8c1d-facadb04db8a" containerName="dnsmasq-dns" containerID="cri-o://5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734" gracePeriod=10 Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.892600 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:48:27 crc kubenswrapper[4978]: W0225 08:48:27.900481 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3c2da42_6edb_4b26_9708_9d08c61190fe.slice/crio-ffad03a748f2eefdf2033d9d6c4a27915e7b55da2e9e29f354258ee7f1f42e15 WatchSource:0}: Error finding container ffad03a748f2eefdf2033d9d6c4a27915e7b55da2e9e29f354258ee7f1f42e15: Status 404 returned error can't find the container with id ffad03a748f2eefdf2033d9d6c4a27915e7b55da2e9e29f354258ee7f1f42e15 Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.964238 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:48:27 crc kubenswrapper[4978]: I0225 08:48:27.966851 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3c2da42-6edb-4b26-9708-9d08c61190fe","Type":"ContainerStarted","Data":"ffad03a748f2eefdf2033d9d6c4a27915e7b55da2e9e29f354258ee7f1f42e15"} Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.000782 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5136c0c-233c-4454-a076-d26cd2538149","Type":"ContainerStarted","Data":"d79740afa03266125c91e4c3d98e3e7f2d069644235f21b99695051739bc9195"} Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.000826 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5136c0c-233c-4454-a076-d26cd2538149","Type":"ContainerStarted","Data":"d222424a73352f3f33aeca1f9d171cf20976192c64c4991a9a728d355c919149"} Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.023243 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztjf6\" (UniqueName: \"kubernetes.io/projected/b985867e-5709-4c7d-8c1d-facadb04db8a-kube-api-access-ztjf6\") pod \"b985867e-5709-4c7d-8c1d-facadb04db8a\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.023498 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-config\") pod \"b985867e-5709-4c7d-8c1d-facadb04db8a\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.023556 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-dns-svc\") pod \"b985867e-5709-4c7d-8c1d-facadb04db8a\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.023616 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-nb\") pod \"b985867e-5709-4c7d-8c1d-facadb04db8a\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.023653 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-sb\") pod \"b985867e-5709-4c7d-8c1d-facadb04db8a\" (UID: \"b985867e-5709-4c7d-8c1d-facadb04db8a\") " Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.023767 4978 generic.go:334] "Generic (PLEG): container finished" podID="b985867e-5709-4c7d-8c1d-facadb04db8a" containerID="5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734" exitCode=0 Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.023813 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" event={"ID":"b985867e-5709-4c7d-8c1d-facadb04db8a","Type":"ContainerDied","Data":"5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734"} Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.023841 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" event={"ID":"b985867e-5709-4c7d-8c1d-facadb04db8a","Type":"ContainerDied","Data":"7972d043d2ab1afc98d40f405b1d7f910b8cb5a143785807a462a7fc96383b02"} Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.023857 4978 scope.go:117] "RemoveContainer" containerID="5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.024010 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-778bcd845c-fw9q9" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.036600 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b985867e-5709-4c7d-8c1d-facadb04db8a-kube-api-access-ztjf6" (OuterVolumeSpecName: "kube-api-access-ztjf6") pod "b985867e-5709-4c7d-8c1d-facadb04db8a" (UID: "b985867e-5709-4c7d-8c1d-facadb04db8a"). InnerVolumeSpecName "kube-api-access-ztjf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.087621 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.087605402 podStartE2EDuration="3.087605402s" podCreationTimestamp="2026-02-25 08:48:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:48:28.085554579 +0000 UTC m=+7401.524811038" watchObservedRunningTime="2026-02-25 08:48:28.087605402 +0000 UTC m=+7401.526861861" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.126683 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztjf6\" (UniqueName: \"kubernetes.io/projected/b985867e-5709-4c7d-8c1d-facadb04db8a-kube-api-access-ztjf6\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.200222 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b985867e-5709-4c7d-8c1d-facadb04db8a" (UID: "b985867e-5709-4c7d-8c1d-facadb04db8a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.216244 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b985867e-5709-4c7d-8c1d-facadb04db8a" (UID: "b985867e-5709-4c7d-8c1d-facadb04db8a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.226173 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b985867e-5709-4c7d-8c1d-facadb04db8a" (UID: "b985867e-5709-4c7d-8c1d-facadb04db8a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.229768 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.229790 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.229802 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.238920 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-config" (OuterVolumeSpecName: "config") pod "b985867e-5709-4c7d-8c1d-facadb04db8a" (UID: "b985867e-5709-4c7d-8c1d-facadb04db8a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.334935 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b985867e-5709-4c7d-8c1d-facadb04db8a-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.343816 4978 scope.go:117] "RemoveContainer" containerID="13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.367576 4978 scope.go:117] "RemoveContainer" containerID="5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734" Feb 25 08:48:28 crc kubenswrapper[4978]: E0225 08:48:28.368398 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734\": container with ID starting with 5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734 not found: ID does not exist" containerID="5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.368465 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734"} err="failed to get container status \"5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734\": rpc error: code = NotFound desc = could not find container \"5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734\": container with ID starting with 5e2cfc930799dd9f9c1b3ccdcca2123329aedb8677b927dd3ac62ffaee039734 not found: ID does not exist" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.368509 4978 scope.go:117] "RemoveContainer" containerID="13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4" Feb 25 08:48:28 crc kubenswrapper[4978]: E0225 08:48:28.372090 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4\": container with ID starting with 13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4 not found: ID does not exist" containerID="13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.372142 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4"} err="failed to get container status \"13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4\": rpc error: code = NotFound desc = could not find container \"13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4\": container with ID starting with 13dc61a37ee50b5a18fc2095f7fdf9ae0a0907d87657e54c5ed1f537833060c4 not found: ID does not exist" Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.380732 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-778bcd845c-fw9q9"] Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.395050 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-778bcd845c-fw9q9"] Feb 25 08:48:28 crc kubenswrapper[4978]: I0225 08:48:28.691018 4978 scope.go:117] "RemoveContainer" containerID="1d5d7c730c018c8087296a28ee707405223e3b0b7b5323b3f21b49ee7c1e7234" Feb 25 08:48:29 crc kubenswrapper[4978]: I0225 08:48:29.065644 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3c2da42-6edb-4b26-9708-9d08c61190fe","Type":"ContainerStarted","Data":"98927fb5e3c62da34936eccdd2ce919109bc411a5349a3650951c61f82284c69"} Feb 25 08:48:29 crc kubenswrapper[4978]: I0225 08:48:29.065710 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3c2da42-6edb-4b26-9708-9d08c61190fe","Type":"ContainerStarted","Data":"597623758b07b4c8c035ef722b65ea7046000a79638c143d94ef7e1e89de4336"} Feb 25 08:48:29 crc kubenswrapper[4978]: I0225 08:48:29.115926 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.11589819 podStartE2EDuration="3.11589819s" podCreationTimestamp="2026-02-25 08:48:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:48:29.098002105 +0000 UTC m=+7402.537258594" watchObservedRunningTime="2026-02-25 08:48:29.11589819 +0000 UTC m=+7402.555154689" Feb 25 08:48:29 crc kubenswrapper[4978]: I0225 08:48:29.325044 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 25 08:48:29 crc kubenswrapper[4978]: I0225 08:48:29.343421 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b985867e-5709-4c7d-8c1d-facadb04db8a" path="/var/lib/kubelet/pods/b985867e-5709-4c7d-8c1d-facadb04db8a/volumes" Feb 25 08:48:32 crc kubenswrapper[4978]: I0225 08:48:32.333606 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 08:48:32 crc kubenswrapper[4978]: I0225 08:48:32.333951 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 08:48:36 crc kubenswrapper[4978]: I0225 08:48:36.319553 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 08:48:36 crc kubenswrapper[4978]: I0225 08:48:36.319909 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 08:48:37 crc kubenswrapper[4978]: I0225 08:48:37.337137 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 08:48:37 crc kubenswrapper[4978]: I0225 08:48:37.337440 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 08:48:37 crc kubenswrapper[4978]: I0225 08:48:37.402574 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.154:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:37 crc kubenswrapper[4978]: I0225 08:48:37.402941 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.154:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:38 crc kubenswrapper[4978]: I0225 08:48:38.343518 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.155:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:38 crc kubenswrapper[4978]: I0225 08:48:38.343608 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.155:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:47 crc kubenswrapper[4978]: I0225 08:48:47.401680 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.154:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:47 crc kubenswrapper[4978]: I0225 08:48:47.402463 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.154:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:48 crc kubenswrapper[4978]: I0225 08:48:48.354594 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.155:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:48 crc kubenswrapper[4978]: I0225 08:48:48.354664 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.155:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:52 crc kubenswrapper[4978]: W0225 08:48:52.917994 4978 container.go:586] Failed to update stats for container "/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaf834ac_cc3b_4728_b4c7_bd19abd9a82a.slice/crio-fcd95ab490fba2c35b13104e3a02d27b1acbbbc83ba32387f007fa29e627c0e4": error while statting cgroup v2: [unable to parse /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaf834ac_cc3b_4728_b4c7_bd19abd9a82a.slice/crio-fcd95ab490fba2c35b13104e3a02d27b1acbbbc83ba32387f007fa29e627c0e4/memory.stat: read /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaf834ac_cc3b_4728_b4c7_bd19abd9a82a.slice/crio-fcd95ab490fba2c35b13104e3a02d27b1acbbbc83ba32387f007fa29e627c0e4/memory.stat: no such device], continuing to push stats Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.312878 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.348879 4978 generic.go:334] "Generic (PLEG): container finished" podID="caf834ac-cc3b-4728-b4c7-bd19abd9a82a" containerID="b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64" exitCode=137 Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.348937 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"caf834ac-cc3b-4728-b4c7-bd19abd9a82a","Type":"ContainerDied","Data":"b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64"} Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.348979 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.349016 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"caf834ac-cc3b-4728-b4c7-bd19abd9a82a","Type":"ContainerDied","Data":"fcd95ab490fba2c35b13104e3a02d27b1acbbbc83ba32387f007fa29e627c0e4"} Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.349050 4978 scope.go:117] "RemoveContainer" containerID="b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.376443 4978 scope.go:117] "RemoveContainer" containerID="b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64" Feb 25 08:48:53 crc kubenswrapper[4978]: E0225 08:48:53.377208 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64\": container with ID starting with b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64 not found: ID does not exist" containerID="b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.377246 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64"} err="failed to get container status \"b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64\": rpc error: code = NotFound desc = could not find container \"b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64\": container with ID starting with b3b96e6bc10a23d7b29116ffc0d19c73f0c05a3a3158c5a814eb1c7705510e64 not found: ID does not exist" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.431546 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-config-data\") pod \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.431636 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d9nk\" (UniqueName: \"kubernetes.io/projected/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-kube-api-access-6d9nk\") pod \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.431846 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-combined-ca-bundle\") pod \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\" (UID: \"caf834ac-cc3b-4728-b4c7-bd19abd9a82a\") " Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.438787 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-kube-api-access-6d9nk" (OuterVolumeSpecName: "kube-api-access-6d9nk") pod "caf834ac-cc3b-4728-b4c7-bd19abd9a82a" (UID: "caf834ac-cc3b-4728-b4c7-bd19abd9a82a"). InnerVolumeSpecName "kube-api-access-6d9nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.465836 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "caf834ac-cc3b-4728-b4c7-bd19abd9a82a" (UID: "caf834ac-cc3b-4728-b4c7-bd19abd9a82a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.474574 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-config-data" (OuterVolumeSpecName: "config-data") pod "caf834ac-cc3b-4728-b4c7-bd19abd9a82a" (UID: "caf834ac-cc3b-4728-b4c7-bd19abd9a82a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.534227 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.534257 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.534268 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d9nk\" (UniqueName: \"kubernetes.io/projected/caf834ac-cc3b-4728-b4c7-bd19abd9a82a-kube-api-access-6d9nk\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.707568 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.735945 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.759011 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 08:48:53 crc kubenswrapper[4978]: E0225 08:48:53.759676 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b985867e-5709-4c7d-8c1d-facadb04db8a" containerName="dnsmasq-dns" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.759708 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b985867e-5709-4c7d-8c1d-facadb04db8a" containerName="dnsmasq-dns" Feb 25 08:48:53 crc kubenswrapper[4978]: E0225 08:48:53.759733 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b985867e-5709-4c7d-8c1d-facadb04db8a" containerName="init" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.759746 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b985867e-5709-4c7d-8c1d-facadb04db8a" containerName="init" Feb 25 08:48:53 crc kubenswrapper[4978]: E0225 08:48:53.759808 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf834ac-cc3b-4728-b4c7-bd19abd9a82a" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.759821 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf834ac-cc3b-4728-b4c7-bd19abd9a82a" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.760128 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf834ac-cc3b-4728-b4c7-bd19abd9a82a" containerName="nova-cell1-novncproxy-novncproxy" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.760154 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b985867e-5709-4c7d-8c1d-facadb04db8a" containerName="dnsmasq-dns" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.761221 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.764485 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.764606 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.765147 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.785499 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.840837 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.841257 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.841682 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.841876 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.842156 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cggd\" (UniqueName: \"kubernetes.io/projected/ddac0d7e-f9ca-4094-854a-e008472a187a-kube-api-access-7cggd\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.944644 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.944702 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.944785 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cggd\" (UniqueName: \"kubernetes.io/projected/ddac0d7e-f9ca-4094-854a-e008472a187a-kube-api-access-7cggd\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.944840 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.944865 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.951260 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.951442 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.952447 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.952644 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddac0d7e-f9ca-4094-854a-e008472a187a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:53 crc kubenswrapper[4978]: I0225 08:48:53.971608 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cggd\" (UniqueName: \"kubernetes.io/projected/ddac0d7e-f9ca-4094-854a-e008472a187a-kube-api-access-7cggd\") pod \"nova-cell1-novncproxy-0\" (UID: \"ddac0d7e-f9ca-4094-854a-e008472a187a\") " pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:54 crc kubenswrapper[4978]: I0225 08:48:54.072041 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-608f-account-create-update-cwl9c"] Feb 25 08:48:54 crc kubenswrapper[4978]: I0225 08:48:54.088837 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-znvxz"] Feb 25 08:48:54 crc kubenswrapper[4978]: I0225 08:48:54.093937 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:48:54 crc kubenswrapper[4978]: I0225 08:48:54.099909 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-608f-account-create-update-cwl9c"] Feb 25 08:48:54 crc kubenswrapper[4978]: I0225 08:48:54.108275 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-znvxz"] Feb 25 08:48:54 crc kubenswrapper[4978]: I0225 08:48:54.574656 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.351805 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47319f12-b693-4085-84d3-8d0d1942b36c" path="/var/lib/kubelet/pods/47319f12-b693-4085-84d3-8d0d1942b36c/volumes" Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.353106 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caf834ac-cc3b-4728-b4c7-bd19abd9a82a" path="/var/lib/kubelet/pods/caf834ac-cc3b-4728-b4c7-bd19abd9a82a/volumes" Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.353838 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdab92b0-d524-4536-bf2d-a474a8cd2787" path="/var/lib/kubelet/pods/cdab92b0-d524-4536-bf2d-a474a8cd2787/volumes" Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.367736 4978 generic.go:334] "Generic (PLEG): container finished" podID="386f01a1-7e55-498a-adda-977e1ada8872" containerID="c8465ec81edabb8d55345d8e1a838840f4c03b5d88087767bd4016e883901e12" exitCode=137 Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.367801 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"386f01a1-7e55-498a-adda-977e1ada8872","Type":"ContainerDied","Data":"c8465ec81edabb8d55345d8e1a838840f4c03b5d88087767bd4016e883901e12"} Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.369762 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ddac0d7e-f9ca-4094-854a-e008472a187a","Type":"ContainerStarted","Data":"ae9be74f789f011a7413d83513e50f413c74d50ada9fbcf6db331971d0142598"} Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.369794 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ddac0d7e-f9ca-4094-854a-e008472a187a","Type":"ContainerStarted","Data":"b95d7fa34cb68de670a2fb7dd3814a6dfcb75b98ac274d9b1db25f3549f11583"} Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.397494 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.397447096 podStartE2EDuration="2.397447096s" podCreationTimestamp="2026-02-25 08:48:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:48:55.394611128 +0000 UTC m=+7428.833867597" watchObservedRunningTime="2026-02-25 08:48:55.397447096 +0000 UTC m=+7428.836703565" Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.602247 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.706259 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-config-data\") pod \"386f01a1-7e55-498a-adda-977e1ada8872\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.706684 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-combined-ca-bundle\") pod \"386f01a1-7e55-498a-adda-977e1ada8872\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.706774 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjs6g\" (UniqueName: \"kubernetes.io/projected/386f01a1-7e55-498a-adda-977e1ada8872-kube-api-access-mjs6g\") pod \"386f01a1-7e55-498a-adda-977e1ada8872\" (UID: \"386f01a1-7e55-498a-adda-977e1ada8872\") " Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.715535 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/386f01a1-7e55-498a-adda-977e1ada8872-kube-api-access-mjs6g" (OuterVolumeSpecName: "kube-api-access-mjs6g") pod "386f01a1-7e55-498a-adda-977e1ada8872" (UID: "386f01a1-7e55-498a-adda-977e1ada8872"). InnerVolumeSpecName "kube-api-access-mjs6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.734298 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-config-data" (OuterVolumeSpecName: "config-data") pod "386f01a1-7e55-498a-adda-977e1ada8872" (UID: "386f01a1-7e55-498a-adda-977e1ada8872"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.735881 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "386f01a1-7e55-498a-adda-977e1ada8872" (UID: "386f01a1-7e55-498a-adda-977e1ada8872"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.810320 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.810384 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjs6g\" (UniqueName: \"kubernetes.io/projected/386f01a1-7e55-498a-adda-977e1ada8872-kube-api-access-mjs6g\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:55 crc kubenswrapper[4978]: I0225 08:48:55.810406 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/386f01a1-7e55-498a-adda-977e1ada8872-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.320441 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.320807 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.387678 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"386f01a1-7e55-498a-adda-977e1ada8872","Type":"ContainerDied","Data":"1f8e69bf38c24d1e93b4a09573b8d3b9f74b2df859e1387cc95dff44f1394fcc"} Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.387790 4978 scope.go:117] "RemoveContainer" containerID="c8465ec81edabb8d55345d8e1a838840f4c03b5d88087767bd4016e883901e12" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.387788 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.453937 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.470459 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.497732 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:48:56 crc kubenswrapper[4978]: E0225 08:48:56.498164 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386f01a1-7e55-498a-adda-977e1ada8872" containerName="nova-scheduler-scheduler" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.498179 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="386f01a1-7e55-498a-adda-977e1ada8872" containerName="nova-scheduler-scheduler" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.498427 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="386f01a1-7e55-498a-adda-977e1ada8872" containerName="nova-scheduler-scheduler" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.504462 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.507793 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.516963 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.630054 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrzkq\" (UniqueName: \"kubernetes.io/projected/c6ce4389-5362-49ba-8f3d-c8249031107b-kube-api-access-jrzkq\") pod \"nova-scheduler-0\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.630157 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.630261 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-config-data\") pod \"nova-scheduler-0\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.732919 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.733037 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-config-data\") pod \"nova-scheduler-0\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.733328 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrzkq\" (UniqueName: \"kubernetes.io/projected/c6ce4389-5362-49ba-8f3d-c8249031107b-kube-api-access-jrzkq\") pod \"nova-scheduler-0\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.741162 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.743606 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-config-data\") pod \"nova-scheduler-0\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.758177 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrzkq\" (UniqueName: \"kubernetes.io/projected/c6ce4389-5362-49ba-8f3d-c8249031107b-kube-api-access-jrzkq\") pod \"nova-scheduler-0\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " pod="openstack/nova-scheduler-0" Feb 25 08:48:56 crc kubenswrapper[4978]: I0225 08:48:56.828730 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 08:48:57 crc kubenswrapper[4978]: I0225 08:48:57.351293 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="386f01a1-7e55-498a-adda-977e1ada8872" path="/var/lib/kubelet/pods/386f01a1-7e55-498a-adda-977e1ada8872/volumes" Feb 25 08:48:57 crc kubenswrapper[4978]: I0225 08:48:57.371139 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:48:57 crc kubenswrapper[4978]: W0225 08:48:57.382267 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6ce4389_5362_49ba_8f3d_c8249031107b.slice/crio-43ff1af6c301d05b95cfcbe9d0cb927d73f4e1ca3e71a0faadc52457e1e09f29 WatchSource:0}: Error finding container 43ff1af6c301d05b95cfcbe9d0cb927d73f4e1ca3e71a0faadc52457e1e09f29: Status 404 returned error can't find the container with id 43ff1af6c301d05b95cfcbe9d0cb927d73f4e1ca3e71a0faadc52457e1e09f29 Feb 25 08:48:57 crc kubenswrapper[4978]: I0225 08:48:57.398283 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6ce4389-5362-49ba-8f3d-c8249031107b","Type":"ContainerStarted","Data":"43ff1af6c301d05b95cfcbe9d0cb927d73f4e1ca3e71a0faadc52457e1e09f29"} Feb 25 08:48:57 crc kubenswrapper[4978]: I0225 08:48:57.406075 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.154:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:57 crc kubenswrapper[4978]: I0225 08:48:57.407310 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.154:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:58 crc kubenswrapper[4978]: I0225 08:48:58.342577 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.155:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:58 crc kubenswrapper[4978]: I0225 08:48:58.342609 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.155:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:48:58 crc kubenswrapper[4978]: I0225 08:48:58.417325 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6ce4389-5362-49ba-8f3d-c8249031107b","Type":"ContainerStarted","Data":"74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d"} Feb 25 08:48:58 crc kubenswrapper[4978]: I0225 08:48:58.476733 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.476709703 podStartE2EDuration="2.476709703s" podCreationTimestamp="2026-02-25 08:48:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:48:58.44045469 +0000 UTC m=+7431.879711209" watchObservedRunningTime="2026-02-25 08:48:58.476709703 +0000 UTC m=+7431.915966202" Feb 25 08:48:59 crc kubenswrapper[4978]: I0225 08:48:59.094233 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:49:01 crc kubenswrapper[4978]: I0225 08:49:01.828995 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.094554 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.129154 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.519331 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.750341 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-d794z"] Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.751685 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.759245 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.759626 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.777174 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-d794z"] Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.811902 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-config-data\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.811985 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.812017 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-scripts\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.812060 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdg5q\" (UniqueName: \"kubernetes.io/projected/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-kube-api-access-mdg5q\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.913260 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-config-data\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.913702 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.913882 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-scripts\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.914075 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdg5q\" (UniqueName: \"kubernetes.io/projected/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-kube-api-access-mdg5q\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.919394 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.919485 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-config-data\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.920237 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-scripts\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:04 crc kubenswrapper[4978]: I0225 08:49:04.934502 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdg5q\" (UniqueName: \"kubernetes.io/projected/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-kube-api-access-mdg5q\") pod \"nova-cell1-cell-mapping-d794z\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:05 crc kubenswrapper[4978]: I0225 08:49:05.054943 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-l9hnq"] Feb 25 08:49:05 crc kubenswrapper[4978]: I0225 08:49:05.063511 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-l9hnq"] Feb 25 08:49:05 crc kubenswrapper[4978]: I0225 08:49:05.071464 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:05 crc kubenswrapper[4978]: I0225 08:49:05.340142 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2e7af06-5b02-45cf-bdad-685aee4a49f4" path="/var/lib/kubelet/pods/d2e7af06-5b02-45cf-bdad-685aee4a49f4/volumes" Feb 25 08:49:05 crc kubenswrapper[4978]: I0225 08:49:05.594528 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-d794z"] Feb 25 08:49:06 crc kubenswrapper[4978]: I0225 08:49:06.514435 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d794z" event={"ID":"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2","Type":"ContainerStarted","Data":"1185ba4cd40e1397a90efca975c8396c8c484410a8e717821da832987a228fb6"} Feb 25 08:49:06 crc kubenswrapper[4978]: I0225 08:49:06.515279 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d794z" event={"ID":"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2","Type":"ContainerStarted","Data":"f6cc87c3afb07cfd838aa44f31c1fa90086bab92263651f0693b2dd7c5ee2b86"} Feb 25 08:49:06 crc kubenswrapper[4978]: I0225 08:49:06.557337 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-d794z" podStartSLOduration=2.5573105270000003 podStartE2EDuration="2.557310527s" podCreationTimestamp="2026-02-25 08:49:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:49:06.540273779 +0000 UTC m=+7439.979530298" watchObservedRunningTime="2026-02-25 08:49:06.557310527 +0000 UTC m=+7439.996567026" Feb 25 08:49:06 crc kubenswrapper[4978]: I0225 08:49:06.829314 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 25 08:49:06 crc kubenswrapper[4978]: I0225 08:49:06.867606 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 25 08:49:07 crc kubenswrapper[4978]: I0225 08:49:07.402975 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.154:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:49:07 crc kubenswrapper[4978]: I0225 08:49:07.404460 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.154:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:49:07 crc kubenswrapper[4978]: I0225 08:49:07.565826 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 25 08:49:08 crc kubenswrapper[4978]: I0225 08:49:08.346564 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.155:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:49:08 crc kubenswrapper[4978]: I0225 08:49:08.346553 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.155:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:49:10 crc kubenswrapper[4978]: I0225 08:49:10.561714 4978 generic.go:334] "Generic (PLEG): container finished" podID="339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2" containerID="1185ba4cd40e1397a90efca975c8396c8c484410a8e717821da832987a228fb6" exitCode=0 Feb 25 08:49:10 crc kubenswrapper[4978]: I0225 08:49:10.562030 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d794z" event={"ID":"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2","Type":"ContainerDied","Data":"1185ba4cd40e1397a90efca975c8396c8c484410a8e717821da832987a228fb6"} Feb 25 08:49:11 crc kubenswrapper[4978]: I0225 08:49:11.982075 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.073456 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-config-data\") pod \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.073515 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-combined-ca-bundle\") pod \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.073690 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdg5q\" (UniqueName: \"kubernetes.io/projected/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-kube-api-access-mdg5q\") pod \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.073869 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-scripts\") pod \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\" (UID: \"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2\") " Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.081051 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-kube-api-access-mdg5q" (OuterVolumeSpecName: "kube-api-access-mdg5q") pod "339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2" (UID: "339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2"). InnerVolumeSpecName "kube-api-access-mdg5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.082189 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-scripts" (OuterVolumeSpecName: "scripts") pod "339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2" (UID: "339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.106749 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-config-data" (OuterVolumeSpecName: "config-data") pod "339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2" (UID: "339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.126751 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2" (UID: "339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.176265 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdg5q\" (UniqueName: \"kubernetes.io/projected/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-kube-api-access-mdg5q\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.176301 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.176315 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.176327 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.589304 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d794z" event={"ID":"339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2","Type":"ContainerDied","Data":"f6cc87c3afb07cfd838aa44f31c1fa90086bab92263651f0693b2dd7c5ee2b86"} Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.589362 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6cc87c3afb07cfd838aa44f31c1fa90086bab92263651f0693b2dd7c5ee2b86" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.589410 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d794z" Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.800856 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.801124 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-log" containerID="cri-o://d222424a73352f3f33aeca1f9d171cf20976192c64c4991a9a728d355c919149" gracePeriod=30 Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.801603 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-api" containerID="cri-o://d79740afa03266125c91e4c3d98e3e7f2d069644235f21b99695051739bc9195" gracePeriod=30 Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.841099 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.841413 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerName="nova-scheduler-scheduler" containerID="cri-o://74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" gracePeriod=30 Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.934683 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.934988 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-log" containerID="cri-o://597623758b07b4c8c035ef722b65ea7046000a79638c143d94ef7e1e89de4336" gracePeriod=30 Feb 25 08:49:12 crc kubenswrapper[4978]: I0225 08:49:12.935143 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-metadata" containerID="cri-o://98927fb5e3c62da34936eccdd2ce919109bc411a5349a3650951c61f82284c69" gracePeriod=30 Feb 25 08:49:13 crc kubenswrapper[4978]: I0225 08:49:13.607778 4978 generic.go:334] "Generic (PLEG): container finished" podID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerID="597623758b07b4c8c035ef722b65ea7046000a79638c143d94ef7e1e89de4336" exitCode=143 Feb 25 08:49:13 crc kubenswrapper[4978]: I0225 08:49:13.607894 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3c2da42-6edb-4b26-9708-9d08c61190fe","Type":"ContainerDied","Data":"597623758b07b4c8c035ef722b65ea7046000a79638c143d94ef7e1e89de4336"} Feb 25 08:49:13 crc kubenswrapper[4978]: I0225 08:49:13.616757 4978 generic.go:334] "Generic (PLEG): container finished" podID="e5136c0c-233c-4454-a076-d26cd2538149" containerID="d222424a73352f3f33aeca1f9d171cf20976192c64c4991a9a728d355c919149" exitCode=143 Feb 25 08:49:13 crc kubenswrapper[4978]: I0225 08:49:13.616804 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5136c0c-233c-4454-a076-d26cd2538149","Type":"ContainerDied","Data":"d222424a73352f3f33aeca1f9d171cf20976192c64c4991a9a728d355c919149"} Feb 25 08:49:16 crc kubenswrapper[4978]: E0225 08:49:16.832181 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:16 crc kubenswrapper[4978]: E0225 08:49:16.836540 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:16 crc kubenswrapper[4978]: E0225 08:49:16.838897 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:16 crc kubenswrapper[4978]: E0225 08:49:16.839001 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerName="nova-scheduler-scheduler" Feb 25 08:49:19 crc kubenswrapper[4978]: I0225 08:49:19.053963 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5nqzd"] Feb 25 08:49:19 crc kubenswrapper[4978]: I0225 08:49:19.096161 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5nqzd"] Feb 25 08:49:19 crc kubenswrapper[4978]: I0225 08:49:19.339298 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e73d580e-bb87-4270-aec3-ad2c35da47ad" path="/var/lib/kubelet/pods/e73d580e-bb87-4270-aec3-ad2c35da47ad/volumes" Feb 25 08:49:21 crc kubenswrapper[4978]: E0225 08:49:21.831982 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:21 crc kubenswrapper[4978]: E0225 08:49:21.834962 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:21 crc kubenswrapper[4978]: E0225 08:49:21.840543 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:21 crc kubenswrapper[4978]: E0225 08:49:21.840607 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerName="nova-scheduler-scheduler" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.768439 4978 generic.go:334] "Generic (PLEG): container finished" podID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerID="98927fb5e3c62da34936eccdd2ce919109bc411a5349a3650951c61f82284c69" exitCode=0 Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.768487 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3c2da42-6edb-4b26-9708-9d08c61190fe","Type":"ContainerDied","Data":"98927fb5e3c62da34936eccdd2ce919109bc411a5349a3650951c61f82284c69"} Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.768943 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c3c2da42-6edb-4b26-9708-9d08c61190fe","Type":"ContainerDied","Data":"ffad03a748f2eefdf2033d9d6c4a27915e7b55da2e9e29f354258ee7f1f42e15"} Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.768959 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffad03a748f2eefdf2033d9d6c4a27915e7b55da2e9e29f354258ee7f1f42e15" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.771354 4978 generic.go:334] "Generic (PLEG): container finished" podID="e5136c0c-233c-4454-a076-d26cd2538149" containerID="d79740afa03266125c91e4c3d98e3e7f2d069644235f21b99695051739bc9195" exitCode=0 Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.771388 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5136c0c-233c-4454-a076-d26cd2538149","Type":"ContainerDied","Data":"d79740afa03266125c91e4c3d98e3e7f2d069644235f21b99695051739bc9195"} Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.771417 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5136c0c-233c-4454-a076-d26cd2538149","Type":"ContainerDied","Data":"80a7ee599321ba51dd990251a2d52f506e01252d39a500dd0ee2338b7666a8b1"} Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.771429 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80a7ee599321ba51dd990251a2d52f506e01252d39a500dd0ee2338b7666a8b1" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.797191 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.801130 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:49:26 crc kubenswrapper[4978]: E0225 08:49:26.831211 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:26 crc kubenswrapper[4978]: E0225 08:49:26.861096 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:26 crc kubenswrapper[4978]: E0225 08:49:26.870580 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:26 crc kubenswrapper[4978]: E0225 08:49:26.870651 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerName="nova-scheduler-scheduler" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.910142 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-config-data\") pod \"c3c2da42-6edb-4b26-9708-9d08c61190fe\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.910286 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5136c0c-233c-4454-a076-d26cd2538149-logs\") pod \"e5136c0c-233c-4454-a076-d26cd2538149\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.910326 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-config-data\") pod \"e5136c0c-233c-4454-a076-d26cd2538149\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.910353 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-combined-ca-bundle\") pod \"c3c2da42-6edb-4b26-9708-9d08c61190fe\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.910395 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-nova-metadata-tls-certs\") pod \"c3c2da42-6edb-4b26-9708-9d08c61190fe\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.910413 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjhlv\" (UniqueName: \"kubernetes.io/projected/e5136c0c-233c-4454-a076-d26cd2538149-kube-api-access-qjhlv\") pod \"e5136c0c-233c-4454-a076-d26cd2538149\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.910445 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-combined-ca-bundle\") pod \"e5136c0c-233c-4454-a076-d26cd2538149\" (UID: \"e5136c0c-233c-4454-a076-d26cd2538149\") " Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.910477 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkrxj\" (UniqueName: \"kubernetes.io/projected/c3c2da42-6edb-4b26-9708-9d08c61190fe-kube-api-access-fkrxj\") pod \"c3c2da42-6edb-4b26-9708-9d08c61190fe\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.910569 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3c2da42-6edb-4b26-9708-9d08c61190fe-logs\") pod \"c3c2da42-6edb-4b26-9708-9d08c61190fe\" (UID: \"c3c2da42-6edb-4b26-9708-9d08c61190fe\") " Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.911412 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3c2da42-6edb-4b26-9708-9d08c61190fe-logs" (OuterVolumeSpecName: "logs") pod "c3c2da42-6edb-4b26-9708-9d08c61190fe" (UID: "c3c2da42-6edb-4b26-9708-9d08c61190fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.911569 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5136c0c-233c-4454-a076-d26cd2538149-logs" (OuterVolumeSpecName: "logs") pod "e5136c0c-233c-4454-a076-d26cd2538149" (UID: "e5136c0c-233c-4454-a076-d26cd2538149"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.916855 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5136c0c-233c-4454-a076-d26cd2538149-kube-api-access-qjhlv" (OuterVolumeSpecName: "kube-api-access-qjhlv") pod "e5136c0c-233c-4454-a076-d26cd2538149" (UID: "e5136c0c-233c-4454-a076-d26cd2538149"). InnerVolumeSpecName "kube-api-access-qjhlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.917750 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3c2da42-6edb-4b26-9708-9d08c61190fe-kube-api-access-fkrxj" (OuterVolumeSpecName: "kube-api-access-fkrxj") pod "c3c2da42-6edb-4b26-9708-9d08c61190fe" (UID: "c3c2da42-6edb-4b26-9708-9d08c61190fe"). InnerVolumeSpecName "kube-api-access-fkrxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.936298 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3c2da42-6edb-4b26-9708-9d08c61190fe" (UID: "c3c2da42-6edb-4b26-9708-9d08c61190fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.942542 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5136c0c-233c-4454-a076-d26cd2538149" (UID: "e5136c0c-233c-4454-a076-d26cd2538149"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.942932 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-config-data" (OuterVolumeSpecName: "config-data") pod "e5136c0c-233c-4454-a076-d26cd2538149" (UID: "e5136c0c-233c-4454-a076-d26cd2538149"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.944062 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-config-data" (OuterVolumeSpecName: "config-data") pod "c3c2da42-6edb-4b26-9708-9d08c61190fe" (UID: "c3c2da42-6edb-4b26-9708-9d08c61190fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:26 crc kubenswrapper[4978]: I0225 08:49:26.969115 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "c3c2da42-6edb-4b26-9708-9d08c61190fe" (UID: "c3c2da42-6edb-4b26-9708-9d08c61190fe"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.013167 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5136c0c-233c-4454-a076-d26cd2538149-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.013212 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.013234 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.013255 4978 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.013289 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjhlv\" (UniqueName: \"kubernetes.io/projected/e5136c0c-233c-4454-a076-d26cd2538149-kube-api-access-qjhlv\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.013308 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5136c0c-233c-4454-a076-d26cd2538149-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.013326 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkrxj\" (UniqueName: \"kubernetes.io/projected/c3c2da42-6edb-4b26-9708-9d08c61190fe-kube-api-access-fkrxj\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.013343 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3c2da42-6edb-4b26-9708-9d08c61190fe-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.013358 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3c2da42-6edb-4b26-9708-9d08c61190fe-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.782731 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.782815 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.816839 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.828086 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.843639 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.859561 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872000 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:49:27 crc kubenswrapper[4978]: E0225 08:49:27.872440 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-log" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872459 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-log" Feb 25 08:49:27 crc kubenswrapper[4978]: E0225 08:49:27.872477 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-api" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872483 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-api" Feb 25 08:49:27 crc kubenswrapper[4978]: E0225 08:49:27.872509 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2" containerName="nova-manage" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872516 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2" containerName="nova-manage" Feb 25 08:49:27 crc kubenswrapper[4978]: E0225 08:49:27.872526 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-log" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872532 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-log" Feb 25 08:49:27 crc kubenswrapper[4978]: E0225 08:49:27.872552 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-metadata" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872558 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-metadata" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872733 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-api" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872748 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5136c0c-233c-4454-a076-d26cd2538149" containerName="nova-api-log" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872765 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2" containerName="nova-manage" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872777 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-metadata" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.872789 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" containerName="nova-metadata-log" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.873747 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.877632 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.877863 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.911679 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.913208 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.916766 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.922432 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:49:27 crc kubenswrapper[4978]: I0225 08:49:27.961118 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.043529 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5bwt\" (UniqueName: \"kubernetes.io/projected/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-kube-api-access-f5bwt\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.043609 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-logs\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.043688 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.043754 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.043825 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-config-data\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.043850 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.043888 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bb32ab4-b355-4406-a6b7-d738373e66d9-logs\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.043935 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-config-data\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.043995 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shtpq\" (UniqueName: \"kubernetes.io/projected/1bb32ab4-b355-4406-a6b7-d738373e66d9-kube-api-access-shtpq\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.145054 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.145124 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-config-data\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.145143 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.145189 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bb32ab4-b355-4406-a6b7-d738373e66d9-logs\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.145215 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-config-data\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.145287 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shtpq\" (UniqueName: \"kubernetes.io/projected/1bb32ab4-b355-4406-a6b7-d738373e66d9-kube-api-access-shtpq\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.145332 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5bwt\" (UniqueName: \"kubernetes.io/projected/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-kube-api-access-f5bwt\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.145396 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-logs\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.145420 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.146778 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bb32ab4-b355-4406-a6b7-d738373e66d9-logs\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.146831 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-logs\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.150442 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-config-data\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.152116 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.153424 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.153653 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.154501 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-config-data\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.163497 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5bwt\" (UniqueName: \"kubernetes.io/projected/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-kube-api-access-f5bwt\") pod \"nova-metadata-0\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.175787 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shtpq\" (UniqueName: \"kubernetes.io/projected/1bb32ab4-b355-4406-a6b7-d738373e66d9-kube-api-access-shtpq\") pod \"nova-api-0\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.199865 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.226147 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.704272 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.781993 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:28 crc kubenswrapper[4978]: W0225 08:49:28.782436 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bb32ab4_b355_4406_a6b7_d738373e66d9.slice/crio-9b9d4098b074e86b34a7b60ab57a4b9d428e94b9742254b69b13e71e5a313317 WatchSource:0}: Error finding container 9b9d4098b074e86b34a7b60ab57a4b9d428e94b9742254b69b13e71e5a313317: Status 404 returned error can't find the container with id 9b9d4098b074e86b34a7b60ab57a4b9d428e94b9742254b69b13e71e5a313317 Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.794950 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1bb32ab4-b355-4406-a6b7-d738373e66d9","Type":"ContainerStarted","Data":"9b9d4098b074e86b34a7b60ab57a4b9d428e94b9742254b69b13e71e5a313317"} Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.796439 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bb14a83b-160a-473c-bbe7-d4720ed8bcb4","Type":"ContainerStarted","Data":"e312c4b0b988910ebedf21b46037fce4cf3817540f99e05ea2c2c9fac23a21a9"} Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.845602 4978 scope.go:117] "RemoveContainer" containerID="e802925c232e3151d8073c8b7c34dc5a6827acd379157082050a59043f6fda37" Feb 25 08:49:28 crc kubenswrapper[4978]: I0225 08:49:28.959545 4978 scope.go:117] "RemoveContainer" containerID="2e3aadcf94ca6bd2f03f7793def413ca633a70114ffae9b6d4e415755c398d25" Feb 25 08:49:29 crc kubenswrapper[4978]: I0225 08:49:29.003988 4978 scope.go:117] "RemoveContainer" containerID="e484ad914e91a8f456a4df58a588f86fd699f9f1247438b959006b04c1a1ddd7" Feb 25 08:49:29 crc kubenswrapper[4978]: I0225 08:49:29.030321 4978 scope.go:117] "RemoveContainer" containerID="b502eadcbed72640c53debfbc5666578bfc0a91bdf42f7286b86218df0a76f9c" Feb 25 08:49:29 crc kubenswrapper[4978]: I0225 08:49:29.342014 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3c2da42-6edb-4b26-9708-9d08c61190fe" path="/var/lib/kubelet/pods/c3c2da42-6edb-4b26-9708-9d08c61190fe/volumes" Feb 25 08:49:29 crc kubenswrapper[4978]: I0225 08:49:29.343268 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5136c0c-233c-4454-a076-d26cd2538149" path="/var/lib/kubelet/pods/e5136c0c-233c-4454-a076-d26cd2538149/volumes" Feb 25 08:49:29 crc kubenswrapper[4978]: I0225 08:49:29.809196 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bb14a83b-160a-473c-bbe7-d4720ed8bcb4","Type":"ContainerStarted","Data":"2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19"} Feb 25 08:49:29 crc kubenswrapper[4978]: I0225 08:49:29.809592 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bb14a83b-160a-473c-bbe7-d4720ed8bcb4","Type":"ContainerStarted","Data":"dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2"} Feb 25 08:49:29 crc kubenswrapper[4978]: I0225 08:49:29.810955 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1bb32ab4-b355-4406-a6b7-d738373e66d9","Type":"ContainerStarted","Data":"729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947"} Feb 25 08:49:29 crc kubenswrapper[4978]: I0225 08:49:29.811007 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1bb32ab4-b355-4406-a6b7-d738373e66d9","Type":"ContainerStarted","Data":"e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404"} Feb 25 08:49:29 crc kubenswrapper[4978]: I0225 08:49:29.834437 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.834408972 podStartE2EDuration="2.834408972s" podCreationTimestamp="2026-02-25 08:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:49:29.825156006 +0000 UTC m=+7463.264412495" watchObservedRunningTime="2026-02-25 08:49:29.834408972 +0000 UTC m=+7463.273665471" Feb 25 08:49:29 crc kubenswrapper[4978]: I0225 08:49:29.859096 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.859066167 podStartE2EDuration="2.859066167s" podCreationTimestamp="2026-02-25 08:49:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:49:29.854318629 +0000 UTC m=+7463.293575128" watchObservedRunningTime="2026-02-25 08:49:29.859066167 +0000 UTC m=+7463.298322666" Feb 25 08:49:31 crc kubenswrapper[4978]: E0225 08:49:31.831290 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:31 crc kubenswrapper[4978]: E0225 08:49:31.833065 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:31 crc kubenswrapper[4978]: E0225 08:49:31.834629 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:31 crc kubenswrapper[4978]: E0225 08:49:31.834662 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerName="nova-scheduler-scheduler" Feb 25 08:49:33 crc kubenswrapper[4978]: I0225 08:49:33.200693 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 08:49:33 crc kubenswrapper[4978]: I0225 08:49:33.200741 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 08:49:36 crc kubenswrapper[4978]: E0225 08:49:36.831823 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:36 crc kubenswrapper[4978]: E0225 08:49:36.837271 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:36 crc kubenswrapper[4978]: E0225 08:49:36.839718 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:36 crc kubenswrapper[4978]: E0225 08:49:36.839800 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerName="nova-scheduler-scheduler" Feb 25 08:49:38 crc kubenswrapper[4978]: I0225 08:49:38.200532 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 08:49:38 crc kubenswrapper[4978]: I0225 08:49:38.200604 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 08:49:38 crc kubenswrapper[4978]: I0225 08:49:38.226447 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 08:49:38 crc kubenswrapper[4978]: I0225 08:49:38.226544 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 08:49:39 crc kubenswrapper[4978]: I0225 08:49:39.216671 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.159:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:49:39 crc kubenswrapper[4978]: I0225 08:49:39.216659 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.159:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:49:39 crc kubenswrapper[4978]: I0225 08:49:39.308619 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.160:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:49:39 crc kubenswrapper[4978]: I0225 08:49:39.308652 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.160:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 08:49:41 crc kubenswrapper[4978]: E0225 08:49:41.832024 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:41 crc kubenswrapper[4978]: E0225 08:49:41.834160 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:41 crc kubenswrapper[4978]: E0225 08:49:41.835945 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 08:49:41 crc kubenswrapper[4978]: E0225 08:49:41.836048 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerName="nova-scheduler-scheduler" Feb 25 08:49:42 crc kubenswrapper[4978]: I0225 08:49:42.977317 4978 generic.go:334] "Generic (PLEG): container finished" podID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" exitCode=137 Feb 25 08:49:42 crc kubenswrapper[4978]: I0225 08:49:42.977405 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6ce4389-5362-49ba-8f3d-c8249031107b","Type":"ContainerDied","Data":"74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d"} Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.325380 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.494701 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-config-data\") pod \"c6ce4389-5362-49ba-8f3d-c8249031107b\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.494847 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrzkq\" (UniqueName: \"kubernetes.io/projected/c6ce4389-5362-49ba-8f3d-c8249031107b-kube-api-access-jrzkq\") pod \"c6ce4389-5362-49ba-8f3d-c8249031107b\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.495080 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-combined-ca-bundle\") pod \"c6ce4389-5362-49ba-8f3d-c8249031107b\" (UID: \"c6ce4389-5362-49ba-8f3d-c8249031107b\") " Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.523652 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6ce4389-5362-49ba-8f3d-c8249031107b-kube-api-access-jrzkq" (OuterVolumeSpecName: "kube-api-access-jrzkq") pod "c6ce4389-5362-49ba-8f3d-c8249031107b" (UID: "c6ce4389-5362-49ba-8f3d-c8249031107b"). InnerVolumeSpecName "kube-api-access-jrzkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.597731 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrzkq\" (UniqueName: \"kubernetes.io/projected/c6ce4389-5362-49ba-8f3d-c8249031107b-kube-api-access-jrzkq\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.633908 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6ce4389-5362-49ba-8f3d-c8249031107b" (UID: "c6ce4389-5362-49ba-8f3d-c8249031107b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.636607 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-config-data" (OuterVolumeSpecName: "config-data") pod "c6ce4389-5362-49ba-8f3d-c8249031107b" (UID: "c6ce4389-5362-49ba-8f3d-c8249031107b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.699106 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.699141 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6ce4389-5362-49ba-8f3d-c8249031107b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.990430 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c6ce4389-5362-49ba-8f3d-c8249031107b","Type":"ContainerDied","Data":"43ff1af6c301d05b95cfcbe9d0cb927d73f4e1ca3e71a0faadc52457e1e09f29"} Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.991413 4978 scope.go:117] "RemoveContainer" containerID="74b3e02c09f5ec531a3d97610eeb3095f7fa0bbab851f7aeea5ba21264f8518d" Feb 25 08:49:43 crc kubenswrapper[4978]: I0225 08:49:43.990571 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.057883 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.071566 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.081498 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:49:44 crc kubenswrapper[4978]: E0225 08:49:44.082121 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerName="nova-scheduler-scheduler" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.082151 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerName="nova-scheduler-scheduler" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.082537 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" containerName="nova-scheduler-scheduler" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.083649 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.091210 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.093443 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.208036 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-config-data\") pod \"nova-scheduler-0\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.208094 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.208261 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvgxc\" (UniqueName: \"kubernetes.io/projected/c14894d9-d8c4-49df-a6e4-d9823efe8f50-kube-api-access-tvgxc\") pod \"nova-scheduler-0\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.310748 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-config-data\") pod \"nova-scheduler-0\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.311153 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.311213 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvgxc\" (UniqueName: \"kubernetes.io/projected/c14894d9-d8c4-49df-a6e4-d9823efe8f50-kube-api-access-tvgxc\") pod \"nova-scheduler-0\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.316280 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.327930 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-config-data\") pod \"nova-scheduler-0\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.328962 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvgxc\" (UniqueName: \"kubernetes.io/projected/c14894d9-d8c4-49df-a6e4-d9823efe8f50-kube-api-access-tvgxc\") pod \"nova-scheduler-0\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.418221 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 08:49:44 crc kubenswrapper[4978]: I0225 08:49:44.844427 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 08:49:44 crc kubenswrapper[4978]: W0225 08:49:44.847615 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc14894d9_d8c4_49df_a6e4_d9823efe8f50.slice/crio-6155a77c902297b891038fe236796ea8ede8b14eafd0460abaed99c9b196ee58 WatchSource:0}: Error finding container 6155a77c902297b891038fe236796ea8ede8b14eafd0460abaed99c9b196ee58: Status 404 returned error can't find the container with id 6155a77c902297b891038fe236796ea8ede8b14eafd0460abaed99c9b196ee58 Feb 25 08:49:45 crc kubenswrapper[4978]: I0225 08:49:45.022930 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c14894d9-d8c4-49df-a6e4-d9823efe8f50","Type":"ContainerStarted","Data":"6155a77c902297b891038fe236796ea8ede8b14eafd0460abaed99c9b196ee58"} Feb 25 08:49:45 crc kubenswrapper[4978]: I0225 08:49:45.337627 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6ce4389-5362-49ba-8f3d-c8249031107b" path="/var/lib/kubelet/pods/c6ce4389-5362-49ba-8f3d-c8249031107b/volumes" Feb 25 08:49:46 crc kubenswrapper[4978]: I0225 08:49:46.035526 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c14894d9-d8c4-49df-a6e4-d9823efe8f50","Type":"ContainerStarted","Data":"5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9"} Feb 25 08:49:46 crc kubenswrapper[4978]: I0225 08:49:46.064174 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.064146234 podStartE2EDuration="2.064146234s" podCreationTimestamp="2026-02-25 08:49:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:49:46.05852165 +0000 UTC m=+7479.497778169" watchObservedRunningTime="2026-02-25 08:49:46.064146234 +0000 UTC m=+7479.503402763" Feb 25 08:49:48 crc kubenswrapper[4978]: I0225 08:49:48.208459 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 08:49:48 crc kubenswrapper[4978]: I0225 08:49:48.210408 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 08:49:48 crc kubenswrapper[4978]: I0225 08:49:48.217282 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 08:49:48 crc kubenswrapper[4978]: I0225 08:49:48.237308 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 08:49:48 crc kubenswrapper[4978]: I0225 08:49:48.237820 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 08:49:48 crc kubenswrapper[4978]: I0225 08:49:48.239258 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 08:49:48 crc kubenswrapper[4978]: I0225 08:49:48.243762 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.069248 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.073595 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.076116 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.272432 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-797c8cd5-csvwh"] Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.273830 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.287215 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797c8cd5-csvwh"] Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.420609 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.422471 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-config\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.422508 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xmqk\" (UniqueName: \"kubernetes.io/projected/5714c95c-8a9f-4885-b450-3867dc4ce904-kube-api-access-9xmqk\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.422542 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-sb\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.422810 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-nb\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.423339 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-dns-svc\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.525580 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-config\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.525686 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xmqk\" (UniqueName: \"kubernetes.io/projected/5714c95c-8a9f-4885-b450-3867dc4ce904-kube-api-access-9xmqk\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.525725 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-sb\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.525792 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-nb\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.525871 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-dns-svc\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.526491 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-config\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.526682 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-dns-svc\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.526682 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-nb\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.527201 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-sb\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.557850 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xmqk\" (UniqueName: \"kubernetes.io/projected/5714c95c-8a9f-4885-b450-3867dc4ce904-kube-api-access-9xmqk\") pod \"dnsmasq-dns-797c8cd5-csvwh\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:49 crc kubenswrapper[4978]: I0225 08:49:49.599619 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:50 crc kubenswrapper[4978]: I0225 08:49:50.093181 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797c8cd5-csvwh"] Feb 25 08:49:51 crc kubenswrapper[4978]: I0225 08:49:51.087666 4978 generic.go:334] "Generic (PLEG): container finished" podID="5714c95c-8a9f-4885-b450-3867dc4ce904" containerID="798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b" exitCode=0 Feb 25 08:49:51 crc kubenswrapper[4978]: I0225 08:49:51.087957 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" event={"ID":"5714c95c-8a9f-4885-b450-3867dc4ce904","Type":"ContainerDied","Data":"798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b"} Feb 25 08:49:51 crc kubenswrapper[4978]: I0225 08:49:51.088132 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" event={"ID":"5714c95c-8a9f-4885-b450-3867dc4ce904","Type":"ContainerStarted","Data":"54bebd2292834da68bbe0ddbb37867737e5747c9d3ef13214f983162572b288b"} Feb 25 08:49:52 crc kubenswrapper[4978]: I0225 08:49:52.104569 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" event={"ID":"5714c95c-8a9f-4885-b450-3867dc4ce904","Type":"ContainerStarted","Data":"2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4"} Feb 25 08:49:52 crc kubenswrapper[4978]: I0225 08:49:52.105017 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:52 crc kubenswrapper[4978]: I0225 08:49:52.106831 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:52 crc kubenswrapper[4978]: I0225 08:49:52.107193 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerName="nova-api-api" containerID="cri-o://729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947" gracePeriod=30 Feb 25 08:49:52 crc kubenswrapper[4978]: I0225 08:49:52.107172 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerName="nova-api-log" containerID="cri-o://e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404" gracePeriod=30 Feb 25 08:49:52 crc kubenswrapper[4978]: I0225 08:49:52.145454 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" podStartSLOduration=3.145437708 podStartE2EDuration="3.145437708s" podCreationTimestamp="2026-02-25 08:49:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:49:52.135424167 +0000 UTC m=+7485.574680636" watchObservedRunningTime="2026-02-25 08:49:52.145437708 +0000 UTC m=+7485.584694167" Feb 25 08:49:53 crc kubenswrapper[4978]: I0225 08:49:53.119489 4978 generic.go:334] "Generic (PLEG): container finished" podID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerID="e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404" exitCode=143 Feb 25 08:49:53 crc kubenswrapper[4978]: I0225 08:49:53.119731 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1bb32ab4-b355-4406-a6b7-d738373e66d9","Type":"ContainerDied","Data":"e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404"} Feb 25 08:49:54 crc kubenswrapper[4978]: I0225 08:49:54.418777 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 25 08:49:54 crc kubenswrapper[4978]: I0225 08:49:54.466690 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 25 08:49:55 crc kubenswrapper[4978]: I0225 08:49:55.191572 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 25 08:49:55 crc kubenswrapper[4978]: I0225 08:49:55.779098 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:49:55 crc kubenswrapper[4978]: I0225 08:49:55.968420 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-config-data\") pod \"1bb32ab4-b355-4406-a6b7-d738373e66d9\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " Feb 25 08:49:55 crc kubenswrapper[4978]: I0225 08:49:55.968503 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shtpq\" (UniqueName: \"kubernetes.io/projected/1bb32ab4-b355-4406-a6b7-d738373e66d9-kube-api-access-shtpq\") pod \"1bb32ab4-b355-4406-a6b7-d738373e66d9\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " Feb 25 08:49:55 crc kubenswrapper[4978]: I0225 08:49:55.968621 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-combined-ca-bundle\") pod \"1bb32ab4-b355-4406-a6b7-d738373e66d9\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " Feb 25 08:49:55 crc kubenswrapper[4978]: I0225 08:49:55.968774 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bb32ab4-b355-4406-a6b7-d738373e66d9-logs\") pod \"1bb32ab4-b355-4406-a6b7-d738373e66d9\" (UID: \"1bb32ab4-b355-4406-a6b7-d738373e66d9\") " Feb 25 08:49:55 crc kubenswrapper[4978]: I0225 08:49:55.969462 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bb32ab4-b355-4406-a6b7-d738373e66d9-logs" (OuterVolumeSpecName: "logs") pod "1bb32ab4-b355-4406-a6b7-d738373e66d9" (UID: "1bb32ab4-b355-4406-a6b7-d738373e66d9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:49:55 crc kubenswrapper[4978]: I0225 08:49:55.978310 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bb32ab4-b355-4406-a6b7-d738373e66d9-kube-api-access-shtpq" (OuterVolumeSpecName: "kube-api-access-shtpq") pod "1bb32ab4-b355-4406-a6b7-d738373e66d9" (UID: "1bb32ab4-b355-4406-a6b7-d738373e66d9"). InnerVolumeSpecName "kube-api-access-shtpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:49:55 crc kubenswrapper[4978]: I0225 08:49:55.991703 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-config-data" (OuterVolumeSpecName: "config-data") pod "1bb32ab4-b355-4406-a6b7-d738373e66d9" (UID: "1bb32ab4-b355-4406-a6b7-d738373e66d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:55 crc kubenswrapper[4978]: I0225 08:49:55.995655 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bb32ab4-b355-4406-a6b7-d738373e66d9" (UID: "1bb32ab4-b355-4406-a6b7-d738373e66d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.070316 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.070345 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bb32ab4-b355-4406-a6b7-d738373e66d9-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.070358 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bb32ab4-b355-4406-a6b7-d738373e66d9-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.070373 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shtpq\" (UniqueName: \"kubernetes.io/projected/1bb32ab4-b355-4406-a6b7-d738373e66d9-kube-api-access-shtpq\") on node \"crc\" DevicePath \"\"" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.157777 4978 generic.go:334] "Generic (PLEG): container finished" podID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerID="729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947" exitCode=0 Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.158527 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1bb32ab4-b355-4406-a6b7-d738373e66d9","Type":"ContainerDied","Data":"729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947"} Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.158577 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1bb32ab4-b355-4406-a6b7-d738373e66d9","Type":"ContainerDied","Data":"9b9d4098b074e86b34a7b60ab57a4b9d428e94b9742254b69b13e71e5a313317"} Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.158574 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.158597 4978 scope.go:117] "RemoveContainer" containerID="729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.180224 4978 scope.go:117] "RemoveContainer" containerID="e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.207249 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.214188 4978 scope.go:117] "RemoveContainer" containerID="729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947" Feb 25 08:49:56 crc kubenswrapper[4978]: E0225 08:49:56.214709 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947\": container with ID starting with 729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947 not found: ID does not exist" containerID="729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.214739 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947"} err="failed to get container status \"729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947\": rpc error: code = NotFound desc = could not find container \"729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947\": container with ID starting with 729214b8385255dc4d4e5d6aadf52f2bebef397c37c9c535b187b13475338947 not found: ID does not exist" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.214763 4978 scope.go:117] "RemoveContainer" containerID="e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404" Feb 25 08:49:56 crc kubenswrapper[4978]: E0225 08:49:56.215145 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404\": container with ID starting with e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404 not found: ID does not exist" containerID="e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.215167 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404"} err="failed to get container status \"e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404\": rpc error: code = NotFound desc = could not find container \"e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404\": container with ID starting with e5b1883cead43e29957ba55c46c740474b67ef82238443c5c5a091f1fbf78404 not found: ID does not exist" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.217167 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.232217 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:56 crc kubenswrapper[4978]: E0225 08:49:56.232627 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerName="nova-api-log" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.232642 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerName="nova-api-log" Feb 25 08:49:56 crc kubenswrapper[4978]: E0225 08:49:56.232673 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerName="nova-api-api" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.232680 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerName="nova-api-api" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.232903 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerName="nova-api-log" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.232932 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" containerName="nova-api-api" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.233845 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.241897 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.241950 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.242280 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.242282 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.375363 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.375468 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-public-tls-certs\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.375602 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.375652 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57wmn\" (UniqueName: \"kubernetes.io/projected/22a39e3e-3d98-4240-b845-75e2d5456ac6-kube-api-access-57wmn\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.375676 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-config-data\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.375721 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a39e3e-3d98-4240-b845-75e2d5456ac6-logs\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.477709 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.477827 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-public-tls-certs\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.477971 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.478031 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57wmn\" (UniqueName: \"kubernetes.io/projected/22a39e3e-3d98-4240-b845-75e2d5456ac6-kube-api-access-57wmn\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.478074 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-config-data\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.478144 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a39e3e-3d98-4240-b845-75e2d5456ac6-logs\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.478891 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a39e3e-3d98-4240-b845-75e2d5456ac6-logs\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.482527 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-public-tls-certs\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.483070 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-config-data\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.483835 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.488568 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.509228 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57wmn\" (UniqueName: \"kubernetes.io/projected/22a39e3e-3d98-4240-b845-75e2d5456ac6-kube-api-access-57wmn\") pod \"nova-api-0\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " pod="openstack/nova-api-0" Feb 25 08:49:56 crc kubenswrapper[4978]: I0225 08:49:56.557100 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 08:49:57 crc kubenswrapper[4978]: I0225 08:49:57.043291 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 08:49:57 crc kubenswrapper[4978]: I0225 08:49:57.171791 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22a39e3e-3d98-4240-b845-75e2d5456ac6","Type":"ContainerStarted","Data":"4f87c102a46a718d446be361acd2387a2e51c003b5f39634a40c6746dd3650b8"} Feb 25 08:49:57 crc kubenswrapper[4978]: I0225 08:49:57.346824 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bb32ab4-b355-4406-a6b7-d738373e66d9" path="/var/lib/kubelet/pods/1bb32ab4-b355-4406-a6b7-d738373e66d9/volumes" Feb 25 08:49:58 crc kubenswrapper[4978]: I0225 08:49:58.189032 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22a39e3e-3d98-4240-b845-75e2d5456ac6","Type":"ContainerStarted","Data":"9a241ef1e2aa2ef3236b4dbcd65d86341b67ad51d21eae2baaada557ac2a0e09"} Feb 25 08:49:58 crc kubenswrapper[4978]: I0225 08:49:58.189099 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22a39e3e-3d98-4240-b845-75e2d5456ac6","Type":"ContainerStarted","Data":"6f3953c76379bf4d3f3b4386500608827daf936585aaadf1474621514f22f5f4"} Feb 25 08:49:58 crc kubenswrapper[4978]: I0225 08:49:58.238777 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.238746594 podStartE2EDuration="2.238746594s" podCreationTimestamp="2026-02-25 08:49:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:49:58.220292893 +0000 UTC m=+7491.659549432" watchObservedRunningTime="2026-02-25 08:49:58.238746594 +0000 UTC m=+7491.678003083" Feb 25 08:49:59 crc kubenswrapper[4978]: I0225 08:49:59.602657 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:49:59 crc kubenswrapper[4978]: I0225 08:49:59.703784 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-677484c4fc-lxtjs"] Feb 25 08:49:59 crc kubenswrapper[4978]: I0225 08:49:59.704199 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" podUID="ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" containerName="dnsmasq-dns" containerID="cri-o://2856f9525243e46948863bde61e1a085af07ea2bf627e25e0b1b1a6ee119a68d" gracePeriod=10 Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.148697 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533490-mv8mf"] Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.161296 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533490-mv8mf" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.163110 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.164735 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.164889 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.206534 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533490-mv8mf"] Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.213267 4978 generic.go:334] "Generic (PLEG): container finished" podID="ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" containerID="2856f9525243e46948863bde61e1a085af07ea2bf627e25e0b1b1a6ee119a68d" exitCode=0 Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.213307 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" event={"ID":"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8","Type":"ContainerDied","Data":"2856f9525243e46948863bde61e1a085af07ea2bf627e25e0b1b1a6ee119a68d"} Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.213332 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" event={"ID":"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8","Type":"ContainerDied","Data":"5947db7210ae49dca8dbcc6d90a3386aa8ea309e8c24c14842095de5d3876b0a"} Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.213344 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5947db7210ae49dca8dbcc6d90a3386aa8ea309e8c24c14842095de5d3876b0a" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.242812 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.291136 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2wt9\" (UniqueName: \"kubernetes.io/projected/e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0-kube-api-access-h2wt9\") pod \"auto-csr-approver-29533490-mv8mf\" (UID: \"e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0\") " pod="openshift-infra/auto-csr-approver-29533490-mv8mf" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.392657 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-sb\") pod \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.392727 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-dns-svc\") pod \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.392837 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-nb\") pod \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.392882 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-config\") pod \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.393009 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhwvz\" (UniqueName: \"kubernetes.io/projected/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-kube-api-access-nhwvz\") pod \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\" (UID: \"ea086a6b-f29e-49b1-a4ee-07c7db3df9f8\") " Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.393779 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2wt9\" (UniqueName: \"kubernetes.io/projected/e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0-kube-api-access-h2wt9\") pod \"auto-csr-approver-29533490-mv8mf\" (UID: \"e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0\") " pod="openshift-infra/auto-csr-approver-29533490-mv8mf" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.408765 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-kube-api-access-nhwvz" (OuterVolumeSpecName: "kube-api-access-nhwvz") pod "ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" (UID: "ea086a6b-f29e-49b1-a4ee-07c7db3df9f8"). InnerVolumeSpecName "kube-api-access-nhwvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.412671 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2wt9\" (UniqueName: \"kubernetes.io/projected/e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0-kube-api-access-h2wt9\") pod \"auto-csr-approver-29533490-mv8mf\" (UID: \"e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0\") " pod="openshift-infra/auto-csr-approver-29533490-mv8mf" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.448255 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" (UID: "ea086a6b-f29e-49b1-a4ee-07c7db3df9f8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.456114 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" (UID: "ea086a6b-f29e-49b1-a4ee-07c7db3df9f8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.473585 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-config" (OuterVolumeSpecName: "config") pod "ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" (UID: "ea086a6b-f29e-49b1-a4ee-07c7db3df9f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.475122 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" (UID: "ea086a6b-f29e-49b1-a4ee-07c7db3df9f8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.496725 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.497696 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.497714 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.497726 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.497737 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhwvz\" (UniqueName: \"kubernetes.io/projected/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8-kube-api-access-nhwvz\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.554421 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533490-mv8mf" Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.931137 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533490-mv8mf"] Feb 25 08:50:00 crc kubenswrapper[4978]: I0225 08:50:00.974700 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 08:50:01 crc kubenswrapper[4978]: I0225 08:50:01.229926 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-677484c4fc-lxtjs" Feb 25 08:50:01 crc kubenswrapper[4978]: I0225 08:50:01.230105 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533490-mv8mf" event={"ID":"e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0","Type":"ContainerStarted","Data":"310d07da72dd564c3579927f37698ccfd87da668eeb23ccfd4df841489be327a"} Feb 25 08:50:01 crc kubenswrapper[4978]: I0225 08:50:01.275679 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-677484c4fc-lxtjs"] Feb 25 08:50:01 crc kubenswrapper[4978]: I0225 08:50:01.283483 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-677484c4fc-lxtjs"] Feb 25 08:50:01 crc kubenswrapper[4978]: I0225 08:50:01.344197 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" path="/var/lib/kubelet/pods/ea086a6b-f29e-49b1-a4ee-07c7db3df9f8/volumes" Feb 25 08:50:03 crc kubenswrapper[4978]: I0225 08:50:03.255484 4978 generic.go:334] "Generic (PLEG): container finished" podID="e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0" containerID="225b0c9dd1125c7894b6a0800aae02d676551673a00a78b8b17ba28d18dca189" exitCode=0 Feb 25 08:50:03 crc kubenswrapper[4978]: I0225 08:50:03.255736 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533490-mv8mf" event={"ID":"e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0","Type":"ContainerDied","Data":"225b0c9dd1125c7894b6a0800aae02d676551673a00a78b8b17ba28d18dca189"} Feb 25 08:50:04 crc kubenswrapper[4978]: I0225 08:50:04.671540 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533490-mv8mf" Feb 25 08:50:04 crc kubenswrapper[4978]: I0225 08:50:04.801161 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2wt9\" (UniqueName: \"kubernetes.io/projected/e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0-kube-api-access-h2wt9\") pod \"e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0\" (UID: \"e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0\") " Feb 25 08:50:04 crc kubenswrapper[4978]: I0225 08:50:04.811937 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0-kube-api-access-h2wt9" (OuterVolumeSpecName: "kube-api-access-h2wt9") pod "e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0" (UID: "e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0"). InnerVolumeSpecName "kube-api-access-h2wt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:50:04 crc kubenswrapper[4978]: I0225 08:50:04.904960 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2wt9\" (UniqueName: \"kubernetes.io/projected/e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0-kube-api-access-h2wt9\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:05 crc kubenswrapper[4978]: I0225 08:50:05.280781 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533490-mv8mf" event={"ID":"e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0","Type":"ContainerDied","Data":"310d07da72dd564c3579927f37698ccfd87da668eeb23ccfd4df841489be327a"} Feb 25 08:50:05 crc kubenswrapper[4978]: I0225 08:50:05.280842 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533490-mv8mf" Feb 25 08:50:05 crc kubenswrapper[4978]: I0225 08:50:05.280846 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="310d07da72dd564c3579927f37698ccfd87da668eeb23ccfd4df841489be327a" Feb 25 08:50:05 crc kubenswrapper[4978]: I0225 08:50:05.759475 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533484-55fdb"] Feb 25 08:50:05 crc kubenswrapper[4978]: I0225 08:50:05.775888 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533484-55fdb"] Feb 25 08:50:06 crc kubenswrapper[4978]: I0225 08:50:06.557505 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 08:50:06 crc kubenswrapper[4978]: I0225 08:50:06.558060 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 08:50:07 crc kubenswrapper[4978]: I0225 08:50:07.347612 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="763ee025-3136-4ba3-8d91-53d980116649" path="/var/lib/kubelet/pods/763ee025-3136-4ba3-8d91-53d980116649/volumes" Feb 25 08:50:07 crc kubenswrapper[4978]: I0225 08:50:07.567771 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.163:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:50:07 crc kubenswrapper[4978]: I0225 08:50:07.576728 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.163:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 08:50:16 crc kubenswrapper[4978]: I0225 08:50:16.569993 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 08:50:16 crc kubenswrapper[4978]: I0225 08:50:16.571731 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 08:50:16 crc kubenswrapper[4978]: I0225 08:50:16.573522 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 08:50:16 crc kubenswrapper[4978]: I0225 08:50:16.584652 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 08:50:17 crc kubenswrapper[4978]: I0225 08:50:17.416331 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 08:50:17 crc kubenswrapper[4978]: I0225 08:50:17.426032 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.260097 4978 scope.go:117] "RemoveContainer" containerID="63d433593b576c533c54000588288ce057fe6ee2c09db10b46ac262f5e311ef2" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.292494 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f588cc4d7-jpzfn"] Feb 25 08:50:29 crc kubenswrapper[4978]: E0225 08:50:29.292999 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" containerName="init" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.293025 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" containerName="init" Feb 25 08:50:29 crc kubenswrapper[4978]: E0225 08:50:29.293065 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0" containerName="oc" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.293085 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0" containerName="oc" Feb 25 08:50:29 crc kubenswrapper[4978]: E0225 08:50:29.293096 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" containerName="dnsmasq-dns" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.293106 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" containerName="dnsmasq-dns" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.293909 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea086a6b-f29e-49b1-a4ee-07c7db3df9f8" containerName="dnsmasq-dns" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.293932 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0" containerName="oc" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.297732 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.301261 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.301508 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.301915 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-vcq4g" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.302323 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.339875 4978 scope.go:117] "RemoveContainer" containerID="76f0115e6d9038f7e9d58a18039109ea26258971f5e059062456304cdfa92e31" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.341256 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f588cc4d7-jpzfn"] Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.350278 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.350528 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6a38b26a-a122-4440-8887-528e8614c0e3" containerName="glance-log" containerID="cri-o://44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133" gracePeriod=30 Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.350943 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6a38b26a-a122-4440-8887-528e8614c0e3" containerName="glance-httpd" containerID="cri-o://6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c" gracePeriod=30 Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.404443 4978 scope.go:117] "RemoveContainer" containerID="75bcb874ea68ae0d4be3642cfc3b4c24733abbfd3402a2fb29189fde4f2eca81" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.408126 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-69dcc5b84c-j4bkp"] Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.409839 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.445664 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69dcc5b84c-j4bkp"] Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.462195 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.462459 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" containerName="glance-log" containerID="cri-o://73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5" gracePeriod=30 Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.462587 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" containerName="glance-httpd" containerID="cri-o://d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42" gracePeriod=30 Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.484199 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-scripts\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.484302 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-config-data\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.484413 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17065e03-40e6-419a-82b6-f30993bfd59a-horizon-secret-key\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.484453 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqgdg\" (UniqueName: \"kubernetes.io/projected/17065e03-40e6-419a-82b6-f30993bfd59a-kube-api-access-mqgdg\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.484497 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17065e03-40e6-419a-82b6-f30993bfd59a-logs\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.547580 4978 generic.go:334] "Generic (PLEG): container finished" podID="6a38b26a-a122-4440-8887-528e8614c0e3" containerID="44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133" exitCode=143 Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.547652 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a38b26a-a122-4440-8887-528e8614c0e3","Type":"ContainerDied","Data":"44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133"} Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.586634 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17065e03-40e6-419a-82b6-f30993bfd59a-horizon-secret-key\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.587782 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqgdg\" (UniqueName: \"kubernetes.io/projected/17065e03-40e6-419a-82b6-f30993bfd59a-kube-api-access-mqgdg\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.587862 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0018d938-52d1-443d-a492-c141a4a6654f-logs\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.587914 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17065e03-40e6-419a-82b6-f30993bfd59a-logs\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.587969 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-scripts\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.588005 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvqwv\" (UniqueName: \"kubernetes.io/projected/0018d938-52d1-443d-a492-c141a4a6654f-kube-api-access-kvqwv\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.588052 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0018d938-52d1-443d-a492-c141a4a6654f-horizon-secret-key\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.588106 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-config-data\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.588157 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-scripts\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.588206 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-config-data\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.588726 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17065e03-40e6-419a-82b6-f30993bfd59a-logs\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.589432 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-scripts\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.590284 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-config-data\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.594926 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17065e03-40e6-419a-82b6-f30993bfd59a-horizon-secret-key\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.606824 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqgdg\" (UniqueName: \"kubernetes.io/projected/17065e03-40e6-419a-82b6-f30993bfd59a-kube-api-access-mqgdg\") pod \"horizon-f588cc4d7-jpzfn\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.661847 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.689294 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-scripts\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.689475 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-config-data\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.689578 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0018d938-52d1-443d-a492-c141a4a6654f-logs\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.689642 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvqwv\" (UniqueName: \"kubernetes.io/projected/0018d938-52d1-443d-a492-c141a4a6654f-kube-api-access-kvqwv\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.689680 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0018d938-52d1-443d-a492-c141a4a6654f-horizon-secret-key\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.690063 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0018d938-52d1-443d-a492-c141a4a6654f-logs\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.690330 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-scripts\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.690910 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-config-data\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.693007 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0018d938-52d1-443d-a492-c141a4a6654f-horizon-secret-key\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.705100 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvqwv\" (UniqueName: \"kubernetes.io/projected/0018d938-52d1-443d-a492-c141a4a6654f-kube-api-access-kvqwv\") pod \"horizon-69dcc5b84c-j4bkp\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:29 crc kubenswrapper[4978]: I0225 08:50:29.746030 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:30 crc kubenswrapper[4978]: I0225 08:50:30.047922 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69dcc5b84c-j4bkp"] Feb 25 08:50:30 crc kubenswrapper[4978]: W0225 08:50:30.050552 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0018d938_52d1_443d_a492_c141a4a6654f.slice/crio-13d5b5238a9182f8e99f29692e3bc07e4b96b1cfc975cef823575e7672b6a8e7 WatchSource:0}: Error finding container 13d5b5238a9182f8e99f29692e3bc07e4b96b1cfc975cef823575e7672b6a8e7: Status 404 returned error can't find the container with id 13d5b5238a9182f8e99f29692e3bc07e4b96b1cfc975cef823575e7672b6a8e7 Feb 25 08:50:30 crc kubenswrapper[4978]: I0225 08:50:30.105715 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f588cc4d7-jpzfn"] Feb 25 08:50:30 crc kubenswrapper[4978]: I0225 08:50:30.559926 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69dcc5b84c-j4bkp" event={"ID":"0018d938-52d1-443d-a492-c141a4a6654f","Type":"ContainerStarted","Data":"13d5b5238a9182f8e99f29692e3bc07e4b96b1cfc975cef823575e7672b6a8e7"} Feb 25 08:50:30 crc kubenswrapper[4978]: I0225 08:50:30.561623 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f588cc4d7-jpzfn" event={"ID":"17065e03-40e6-419a-82b6-f30993bfd59a","Type":"ContainerStarted","Data":"460e327b9bc01b11de80c7235e72e718bc481648796ae60ade3445bfc82f801d"} Feb 25 08:50:30 crc kubenswrapper[4978]: I0225 08:50:30.565297 4978 generic.go:334] "Generic (PLEG): container finished" podID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" containerID="73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5" exitCode=143 Feb 25 08:50:30 crc kubenswrapper[4978]: I0225 08:50:30.565331 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e0da9b3-9cc8-475d-8f9c-f5d228575271","Type":"ContainerDied","Data":"73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5"} Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.505435 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69dcc5b84c-j4bkp"] Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.539729 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6994658d78-bcdzg"] Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.541221 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.543563 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.567138 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6994658d78-bcdzg"] Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.600204 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f588cc4d7-jpzfn"] Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.634415 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7cf5bdb968-klhbw"] Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.636095 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.657039 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cf5bdb968-klhbw"] Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.736342 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-tls-certs\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.737241 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-secret-key\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.737380 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01581086-b21b-4f8a-8390-df51be3a5999-logs\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.737601 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-combined-ca-bundle\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.737722 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-secret-key\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.737799 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-scripts\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.737855 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-config-data\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.737882 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-config-data\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.737934 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpnxb\" (UniqueName: \"kubernetes.io/projected/01581086-b21b-4f8a-8390-df51be3a5999-kube-api-access-tpnxb\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.738020 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-tls-certs\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.738084 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-combined-ca-bundle\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.738139 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lwqp\" (UniqueName: \"kubernetes.io/projected/e8cba710-0510-48af-a7e5-40285f95acb2-kube-api-access-6lwqp\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.738170 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8cba710-0510-48af-a7e5-40285f95acb2-logs\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.738218 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-scripts\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.840746 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-combined-ca-bundle\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.840827 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-secret-key\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.840859 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-scripts\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.840886 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-config-data\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.840903 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-config-data\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.840921 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpnxb\" (UniqueName: \"kubernetes.io/projected/01581086-b21b-4f8a-8390-df51be3a5999-kube-api-access-tpnxb\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.840959 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-tls-certs\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.840980 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-combined-ca-bundle\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.840997 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lwqp\" (UniqueName: \"kubernetes.io/projected/e8cba710-0510-48af-a7e5-40285f95acb2-kube-api-access-6lwqp\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.841014 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8cba710-0510-48af-a7e5-40285f95acb2-logs\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.841029 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-scripts\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.841067 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-tls-certs\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.841095 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-secret-key\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.841110 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01581086-b21b-4f8a-8390-df51be3a5999-logs\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.842045 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01581086-b21b-4f8a-8390-df51be3a5999-logs\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.842824 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-scripts\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.843082 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-config-data\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.843166 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-scripts\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.843444 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8cba710-0510-48af-a7e5-40285f95acb2-logs\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.844166 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-config-data\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.856149 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-tls-certs\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.856323 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-combined-ca-bundle\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.856631 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-secret-key\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.856731 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-secret-key\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.859914 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-tls-certs\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.860976 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpnxb\" (UniqueName: \"kubernetes.io/projected/01581086-b21b-4f8a-8390-df51be3a5999-kube-api-access-tpnxb\") pod \"horizon-7cf5bdb968-klhbw\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.861175 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-combined-ca-bundle\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.863503 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lwqp\" (UniqueName: \"kubernetes.io/projected/e8cba710-0510-48af-a7e5-40285f95acb2-kube-api-access-6lwqp\") pod \"horizon-6994658d78-bcdzg\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:31 crc kubenswrapper[4978]: I0225 08:50:31.964051 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:32 crc kubenswrapper[4978]: I0225 08:50:32.162700 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:32 crc kubenswrapper[4978]: I0225 08:50:32.439440 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cf5bdb968-klhbw"] Feb 25 08:50:32 crc kubenswrapper[4978]: W0225 08:50:32.449596 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01581086_b21b_4f8a_8390_df51be3a5999.slice/crio-1db00d1aa44876704b44d201dbf9acd246a0d4795ecb68201c60d654a3883f53 WatchSource:0}: Error finding container 1db00d1aa44876704b44d201dbf9acd246a0d4795ecb68201c60d654a3883f53: Status 404 returned error can't find the container with id 1db00d1aa44876704b44d201dbf9acd246a0d4795ecb68201c60d654a3883f53 Feb 25 08:50:32 crc kubenswrapper[4978]: I0225 08:50:32.588518 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cf5bdb968-klhbw" event={"ID":"01581086-b21b-4f8a-8390-df51be3a5999","Type":"ContainerStarted","Data":"1db00d1aa44876704b44d201dbf9acd246a0d4795ecb68201c60d654a3883f53"} Feb 25 08:50:32 crc kubenswrapper[4978]: I0225 08:50:32.600262 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6994658d78-bcdzg"] Feb 25 08:50:32 crc kubenswrapper[4978]: W0225 08:50:32.610699 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8cba710_0510_48af_a7e5_40285f95acb2.slice/crio-b41156409c16489cd08ccbc1de0a2fc3ef330f4f13b2099f3bcbf5ef63541bbf WatchSource:0}: Error finding container b41156409c16489cd08ccbc1de0a2fc3ef330f4f13b2099f3bcbf5ef63541bbf: Status 404 returned error can't find the container with id b41156409c16489cd08ccbc1de0a2fc3ef330f4f13b2099f3bcbf5ef63541bbf Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.144077 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.274230 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-logs\") pod \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.274789 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-logs" (OuterVolumeSpecName: "logs") pod "6e0da9b3-9cc8-475d-8f9c-f5d228575271" (UID: "6e0da9b3-9cc8-475d-8f9c-f5d228575271"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.275411 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fjd4\" (UniqueName: \"kubernetes.io/projected/6e0da9b3-9cc8-475d-8f9c-f5d228575271-kube-api-access-2fjd4\") pod \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.276135 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-scripts\") pod \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.277033 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-httpd-run\") pod \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.277890 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-config-data\") pod \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.277976 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-internal-tls-certs\") pod \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.278064 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-combined-ca-bundle\") pod \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\" (UID: \"6e0da9b3-9cc8-475d-8f9c-f5d228575271\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.278771 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.279012 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6e0da9b3-9cc8-475d-8f9c-f5d228575271" (UID: "6e0da9b3-9cc8-475d-8f9c-f5d228575271"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.282129 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-scripts" (OuterVolumeSpecName: "scripts") pod "6e0da9b3-9cc8-475d-8f9c-f5d228575271" (UID: "6e0da9b3-9cc8-475d-8f9c-f5d228575271"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.282501 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e0da9b3-9cc8-475d-8f9c-f5d228575271-kube-api-access-2fjd4" (OuterVolumeSpecName: "kube-api-access-2fjd4") pod "6e0da9b3-9cc8-475d-8f9c-f5d228575271" (UID: "6e0da9b3-9cc8-475d-8f9c-f5d228575271"). InnerVolumeSpecName "kube-api-access-2fjd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.319167 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.328135 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e0da9b3-9cc8-475d-8f9c-f5d228575271" (UID: "6e0da9b3-9cc8-475d-8f9c-f5d228575271"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.350301 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6e0da9b3-9cc8-475d-8f9c-f5d228575271" (UID: "6e0da9b3-9cc8-475d-8f9c-f5d228575271"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.352649 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-config-data" (OuterVolumeSpecName: "config-data") pod "6e0da9b3-9cc8-475d-8f9c-f5d228575271" (UID: "6e0da9b3-9cc8-475d-8f9c-f5d228575271"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.380190 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.380219 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e0da9b3-9cc8-475d-8f9c-f5d228575271-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.380233 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.380261 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.380272 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0da9b3-9cc8-475d-8f9c-f5d228575271-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.380282 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fjd4\" (UniqueName: \"kubernetes.io/projected/6e0da9b3-9cc8-475d-8f9c-f5d228575271-kube-api-access-2fjd4\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.481085 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-scripts\") pod \"6a38b26a-a122-4440-8887-528e8614c0e3\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.481247 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r97gt\" (UniqueName: \"kubernetes.io/projected/6a38b26a-a122-4440-8887-528e8614c0e3-kube-api-access-r97gt\") pod \"6a38b26a-a122-4440-8887-528e8614c0e3\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.481325 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-combined-ca-bundle\") pod \"6a38b26a-a122-4440-8887-528e8614c0e3\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.481349 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-config-data\") pod \"6a38b26a-a122-4440-8887-528e8614c0e3\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.481486 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-public-tls-certs\") pod \"6a38b26a-a122-4440-8887-528e8614c0e3\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.481551 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-httpd-run\") pod \"6a38b26a-a122-4440-8887-528e8614c0e3\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.481579 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-logs\") pod \"6a38b26a-a122-4440-8887-528e8614c0e3\" (UID: \"6a38b26a-a122-4440-8887-528e8614c0e3\") " Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.482544 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6a38b26a-a122-4440-8887-528e8614c0e3" (UID: "6a38b26a-a122-4440-8887-528e8614c0e3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.483122 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-logs" (OuterVolumeSpecName: "logs") pod "6a38b26a-a122-4440-8887-528e8614c0e3" (UID: "6a38b26a-a122-4440-8887-528e8614c0e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.485763 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-scripts" (OuterVolumeSpecName: "scripts") pod "6a38b26a-a122-4440-8887-528e8614c0e3" (UID: "6a38b26a-a122-4440-8887-528e8614c0e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.487175 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a38b26a-a122-4440-8887-528e8614c0e3-kube-api-access-r97gt" (OuterVolumeSpecName: "kube-api-access-r97gt") pod "6a38b26a-a122-4440-8887-528e8614c0e3" (UID: "6a38b26a-a122-4440-8887-528e8614c0e3"). InnerVolumeSpecName "kube-api-access-r97gt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.513317 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a38b26a-a122-4440-8887-528e8614c0e3" (UID: "6a38b26a-a122-4440-8887-528e8614c0e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.529354 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6a38b26a-a122-4440-8887-528e8614c0e3" (UID: "6a38b26a-a122-4440-8887-528e8614c0e3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.529591 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-config-data" (OuterVolumeSpecName: "config-data") pod "6a38b26a-a122-4440-8887-528e8614c0e3" (UID: "6a38b26a-a122-4440-8887-528e8614c0e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.584273 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.584313 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r97gt\" (UniqueName: \"kubernetes.io/projected/6a38b26a-a122-4440-8887-528e8614c0e3-kube-api-access-r97gt\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.584329 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.584337 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.584348 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a38b26a-a122-4440-8887-528e8614c0e3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.584356 4978 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.584383 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a38b26a-a122-4440-8887-528e8614c0e3-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.599331 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6994658d78-bcdzg" event={"ID":"e8cba710-0510-48af-a7e5-40285f95acb2","Type":"ContainerStarted","Data":"b41156409c16489cd08ccbc1de0a2fc3ef330f4f13b2099f3bcbf5ef63541bbf"} Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.601562 4978 generic.go:334] "Generic (PLEG): container finished" podID="6a38b26a-a122-4440-8887-528e8614c0e3" containerID="6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c" exitCode=0 Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.601645 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a38b26a-a122-4440-8887-528e8614c0e3","Type":"ContainerDied","Data":"6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c"} Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.601667 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a38b26a-a122-4440-8887-528e8614c0e3","Type":"ContainerDied","Data":"cb0cc2cfd295cdb62c2a2b8e5ef9cdd1ac980280a0d12c184af89737f7306139"} Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.601673 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.601684 4978 scope.go:117] "RemoveContainer" containerID="6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.606019 4978 generic.go:334] "Generic (PLEG): container finished" podID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" containerID="d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42" exitCode=0 Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.606065 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e0da9b3-9cc8-475d-8f9c-f5d228575271","Type":"ContainerDied","Data":"d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42"} Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.606096 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e0da9b3-9cc8-475d-8f9c-f5d228575271","Type":"ContainerDied","Data":"35ea630b460af7b9bc53081a57206072b5392b5d571b573b07241fa8ad7b5aa9"} Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.606179 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.645552 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.684609 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.694417 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.703325 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.711891 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:50:33 crc kubenswrapper[4978]: E0225 08:50:33.712313 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" containerName="glance-log" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.712328 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" containerName="glance-log" Feb 25 08:50:33 crc kubenswrapper[4978]: E0225 08:50:33.712351 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" containerName="glance-httpd" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.712357 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" containerName="glance-httpd" Feb 25 08:50:33 crc kubenswrapper[4978]: E0225 08:50:33.712386 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a38b26a-a122-4440-8887-528e8614c0e3" containerName="glance-log" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.712392 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a38b26a-a122-4440-8887-528e8614c0e3" containerName="glance-log" Feb 25 08:50:33 crc kubenswrapper[4978]: E0225 08:50:33.712411 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a38b26a-a122-4440-8887-528e8614c0e3" containerName="glance-httpd" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.712418 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a38b26a-a122-4440-8887-528e8614c0e3" containerName="glance-httpd" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.712580 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a38b26a-a122-4440-8887-528e8614c0e3" containerName="glance-httpd" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.712591 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" containerName="glance-httpd" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.712602 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" containerName="glance-log" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.712622 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a38b26a-a122-4440-8887-528e8614c0e3" containerName="glance-log" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.716236 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.717846 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.718789 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.719204 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.719336 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-p5b46" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.725631 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.739107 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.740652 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.742140 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.742166 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.764803 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889642 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889700 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-config-data\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889723 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889744 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889776 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqgw7\" (UniqueName: \"kubernetes.io/projected/6bd498f7-e4ee-4062-8db7-cade3fe823af-kube-api-access-bqgw7\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889817 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889856 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd498f7-e4ee-4062-8db7-cade3fe823af-logs\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889878 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6bd498f7-e4ee-4062-8db7-cade3fe823af-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889895 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d4zf\" (UniqueName: \"kubernetes.io/projected/84c35447-f585-4c1d-b04e-a0e6a86b8541-kube-api-access-2d4zf\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889914 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-scripts\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889946 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889961 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84c35447-f585-4c1d-b04e-a0e6a86b8541-logs\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889974 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84c35447-f585-4c1d-b04e-a0e6a86b8541-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.889996 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.991897 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqgw7\" (UniqueName: \"kubernetes.io/projected/6bd498f7-e4ee-4062-8db7-cade3fe823af-kube-api-access-bqgw7\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.991968 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992008 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd498f7-e4ee-4062-8db7-cade3fe823af-logs\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992031 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d4zf\" (UniqueName: \"kubernetes.io/projected/84c35447-f585-4c1d-b04e-a0e6a86b8541-kube-api-access-2d4zf\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992049 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6bd498f7-e4ee-4062-8db7-cade3fe823af-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992067 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-scripts\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992096 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992112 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84c35447-f585-4c1d-b04e-a0e6a86b8541-logs\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992125 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84c35447-f585-4c1d-b04e-a0e6a86b8541-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992146 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992190 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992212 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-config-data\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992229 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992248 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.992970 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bd498f7-e4ee-4062-8db7-cade3fe823af-logs\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.993143 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84c35447-f585-4c1d-b04e-a0e6a86b8541-logs\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.993240 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6bd498f7-e4ee-4062-8db7-cade3fe823af-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.994016 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/84c35447-f585-4c1d-b04e-a0e6a86b8541-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.997058 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.997185 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-scripts\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.998036 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-config-data\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.998630 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:33 crc kubenswrapper[4978]: I0225 08:50:33.999265 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bd498f7-e4ee-4062-8db7-cade3fe823af-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:34 crc kubenswrapper[4978]: I0225 08:50:34.000131 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:34 crc kubenswrapper[4978]: I0225 08:50:34.000505 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-config-data\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:34 crc kubenswrapper[4978]: I0225 08:50:34.001077 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/84c35447-f585-4c1d-b04e-a0e6a86b8541-scripts\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:34 crc kubenswrapper[4978]: I0225 08:50:34.009496 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqgw7\" (UniqueName: \"kubernetes.io/projected/6bd498f7-e4ee-4062-8db7-cade3fe823af-kube-api-access-bqgw7\") pod \"glance-default-external-api-0\" (UID: \"6bd498f7-e4ee-4062-8db7-cade3fe823af\") " pod="openstack/glance-default-external-api-0" Feb 25 08:50:34 crc kubenswrapper[4978]: I0225 08:50:34.011102 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d4zf\" (UniqueName: \"kubernetes.io/projected/84c35447-f585-4c1d-b04e-a0e6a86b8541-kube-api-access-2d4zf\") pod \"glance-default-internal-api-0\" (UID: \"84c35447-f585-4c1d-b04e-a0e6a86b8541\") " pod="openstack/glance-default-internal-api-0" Feb 25 08:50:34 crc kubenswrapper[4978]: I0225 08:50:34.037020 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:34 crc kubenswrapper[4978]: I0225 08:50:34.057460 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 25 08:50:35 crc kubenswrapper[4978]: I0225 08:50:35.345086 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a38b26a-a122-4440-8887-528e8614c0e3" path="/var/lib/kubelet/pods/6a38b26a-a122-4440-8887-528e8614c0e3/volumes" Feb 25 08:50:35 crc kubenswrapper[4978]: I0225 08:50:35.347356 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e0da9b3-9cc8-475d-8f9c-f5d228575271" path="/var/lib/kubelet/pods/6e0da9b3-9cc8-475d-8f9c-f5d228575271/volumes" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.176966 4978 scope.go:117] "RemoveContainer" containerID="44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.272774 4978 scope.go:117] "RemoveContainer" containerID="6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c" Feb 25 08:50:38 crc kubenswrapper[4978]: E0225 08:50:38.273689 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c\": container with ID starting with 6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c not found: ID does not exist" containerID="6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.273730 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c"} err="failed to get container status \"6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c\": rpc error: code = NotFound desc = could not find container \"6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c\": container with ID starting with 6b77c06b8c258efc1628e72e38d9071fbd54271a82dee955405494761793f48c not found: ID does not exist" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.273758 4978 scope.go:117] "RemoveContainer" containerID="44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133" Feb 25 08:50:38 crc kubenswrapper[4978]: E0225 08:50:38.274813 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133\": container with ID starting with 44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133 not found: ID does not exist" containerID="44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.274851 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133"} err="failed to get container status \"44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133\": rpc error: code = NotFound desc = could not find container \"44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133\": container with ID starting with 44afb35ae41bbce0b44a490bdcb3014ca3bf004d7d97d94e6209b0dd8dc1f133 not found: ID does not exist" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.274873 4978 scope.go:117] "RemoveContainer" containerID="d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.432495 4978 scope.go:117] "RemoveContainer" containerID="73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.483081 4978 scope.go:117] "RemoveContainer" containerID="d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42" Feb 25 08:50:38 crc kubenswrapper[4978]: E0225 08:50:38.483881 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42\": container with ID starting with d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42 not found: ID does not exist" containerID="d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.483920 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42"} err="failed to get container status \"d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42\": rpc error: code = NotFound desc = could not find container \"d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42\": container with ID starting with d2a29d2070b176eb92a7162a36964086c9cfca7852e32961337a632140235b42 not found: ID does not exist" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.483945 4978 scope.go:117] "RemoveContainer" containerID="73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5" Feb 25 08:50:38 crc kubenswrapper[4978]: E0225 08:50:38.484270 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5\": container with ID starting with 73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5 not found: ID does not exist" containerID="73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.484300 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5"} err="failed to get container status \"73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5\": rpc error: code = NotFound desc = could not find container \"73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5\": container with ID starting with 73427f9459882159ffc5e5dc499686f465b356908019d90a2e6b60170d2176e5 not found: ID does not exist" Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.655158 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69dcc5b84c-j4bkp" event={"ID":"0018d938-52d1-443d-a492-c141a4a6654f","Type":"ContainerStarted","Data":"f098cafbe86677d59b945b64a54282fcb046381f8da289257f228d608745ca00"} Feb 25 08:50:38 crc kubenswrapper[4978]: I0225 08:50:38.917229 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 25 08:50:38 crc kubenswrapper[4978]: W0225 08:50:38.924022 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6bd498f7_e4ee_4062_8db7_cade3fe823af.slice/crio-4b1947177363a0f74b764993524d3ce8170c035937b7f141d95a287e30b3610b WatchSource:0}: Error finding container 4b1947177363a0f74b764993524d3ce8170c035937b7f141d95a287e30b3610b: Status 404 returned error can't find the container with id 4b1947177363a0f74b764993524d3ce8170c035937b7f141d95a287e30b3610b Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.676967 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cf5bdb968-klhbw" event={"ID":"01581086-b21b-4f8a-8390-df51be3a5999","Type":"ContainerStarted","Data":"1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa"} Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.677405 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cf5bdb968-klhbw" event={"ID":"01581086-b21b-4f8a-8390-df51be3a5999","Type":"ContainerStarted","Data":"191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b"} Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.689206 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69dcc5b84c-j4bkp" event={"ID":"0018d938-52d1-443d-a492-c141a4a6654f","Type":"ContainerStarted","Data":"436ce56e900d6ebeaf73e1d03ed222eaf0b322089a9c8957f0d6071a5f642d87"} Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.689338 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69dcc5b84c-j4bkp" podUID="0018d938-52d1-443d-a492-c141a4a6654f" containerName="horizon-log" containerID="cri-o://f098cafbe86677d59b945b64a54282fcb046381f8da289257f228d608745ca00" gracePeriod=30 Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.689568 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69dcc5b84c-j4bkp" podUID="0018d938-52d1-443d-a492-c141a4a6654f" containerName="horizon" containerID="cri-o://436ce56e900d6ebeaf73e1d03ed222eaf0b322089a9c8957f0d6071a5f642d87" gracePeriod=30 Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.702995 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7cf5bdb968-klhbw" podStartSLOduration=2.722176703 podStartE2EDuration="8.702972593s" podCreationTimestamp="2026-02-25 08:50:31 +0000 UTC" firstStartedPulling="2026-02-25 08:50:32.453019574 +0000 UTC m=+7525.892276033" lastFinishedPulling="2026-02-25 08:50:38.433815464 +0000 UTC m=+7531.873071923" observedRunningTime="2026-02-25 08:50:39.694137821 +0000 UTC m=+7533.133394280" watchObservedRunningTime="2026-02-25 08:50:39.702972593 +0000 UTC m=+7533.142229052" Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.715323 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f588cc4d7-jpzfn" event={"ID":"17065e03-40e6-419a-82b6-f30993bfd59a","Type":"ContainerStarted","Data":"0d36dc01a6cd595254ece55d27282f4ff5e7694aa6b385f6850e80dc1af2e441"} Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.715390 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f588cc4d7-jpzfn" event={"ID":"17065e03-40e6-419a-82b6-f30993bfd59a","Type":"ContainerStarted","Data":"d84fc21fc07dca7e451337b353842cbf9f1221ba197c47c2a50c805805c451ff"} Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.715719 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f588cc4d7-jpzfn" podUID="17065e03-40e6-419a-82b6-f30993bfd59a" containerName="horizon-log" containerID="cri-o://d84fc21fc07dca7e451337b353842cbf9f1221ba197c47c2a50c805805c451ff" gracePeriod=30 Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.715799 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f588cc4d7-jpzfn" podUID="17065e03-40e6-419a-82b6-f30993bfd59a" containerName="horizon" containerID="cri-o://0d36dc01a6cd595254ece55d27282f4ff5e7694aa6b385f6850e80dc1af2e441" gracePeriod=30 Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.722415 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6994658d78-bcdzg" event={"ID":"e8cba710-0510-48af-a7e5-40285f95acb2","Type":"ContainerStarted","Data":"be36bd631a4e297b69dcefa831029241ef977d0822fe47c6d1a8aa57777c9a5b"} Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.722457 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6994658d78-bcdzg" event={"ID":"e8cba710-0510-48af-a7e5-40285f95acb2","Type":"ContainerStarted","Data":"43cfe060d0263ac53cbc88b9ecfbc0593efe41c3d9f37a65d6e884ed621f69b7"} Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.722499 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-69dcc5b84c-j4bkp" podStartSLOduration=2.460392541 podStartE2EDuration="10.722479948s" podCreationTimestamp="2026-02-25 08:50:29 +0000 UTC" firstStartedPulling="2026-02-25 08:50:30.053450413 +0000 UTC m=+7523.492706882" lastFinishedPulling="2026-02-25 08:50:38.31553783 +0000 UTC m=+7531.754794289" observedRunningTime="2026-02-25 08:50:39.714957245 +0000 UTC m=+7533.154213704" watchObservedRunningTime="2026-02-25 08:50:39.722479948 +0000 UTC m=+7533.161736407" Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.739006 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6bd498f7-e4ee-4062-8db7-cade3fe823af","Type":"ContainerStarted","Data":"6cf6037bd190cdc15d893c1f9f0186ef0062f92729ff82e185bb21679cdf129e"} Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.739061 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6bd498f7-e4ee-4062-8db7-cade3fe823af","Type":"ContainerStarted","Data":"4b1947177363a0f74b764993524d3ce8170c035937b7f141d95a287e30b3610b"} Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.747426 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.762178 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-f588cc4d7-jpzfn" podStartSLOduration=2.55655647 podStartE2EDuration="10.762155927s" podCreationTimestamp="2026-02-25 08:50:29 +0000 UTC" firstStartedPulling="2026-02-25 08:50:30.123265945 +0000 UTC m=+7523.562522404" lastFinishedPulling="2026-02-25 08:50:38.328865402 +0000 UTC m=+7531.768121861" observedRunningTime="2026-02-25 08:50:39.7386554 +0000 UTC m=+7533.177911869" watchObservedRunningTime="2026-02-25 08:50:39.762155927 +0000 UTC m=+7533.201412386" Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.778164 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6994658d78-bcdzg" podStartSLOduration=3.04319925 podStartE2EDuration="8.778147393s" podCreationTimestamp="2026-02-25 08:50:31 +0000 UTC" firstStartedPulling="2026-02-25 08:50:32.613702632 +0000 UTC m=+7526.052959091" lastFinishedPulling="2026-02-25 08:50:38.348650775 +0000 UTC m=+7531.787907234" observedRunningTime="2026-02-25 08:50:39.756527153 +0000 UTC m=+7533.195783612" watchObservedRunningTime="2026-02-25 08:50:39.778147393 +0000 UTC m=+7533.217403852" Feb 25 08:50:39 crc kubenswrapper[4978]: I0225 08:50:39.812849 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 25 08:50:40 crc kubenswrapper[4978]: I0225 08:50:40.753240 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84c35447-f585-4c1d-b04e-a0e6a86b8541","Type":"ContainerStarted","Data":"47e947790f5e6ac1b7ebf0977ac7f13d09e2259f3fee313c72c85a3c8d5a4562"} Feb 25 08:50:40 crc kubenswrapper[4978]: I0225 08:50:40.753859 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84c35447-f585-4c1d-b04e-a0e6a86b8541","Type":"ContainerStarted","Data":"8246a9a500c37d7dcfd0718a1f28a1506af8f4a54ae880aae89b0443022ddf6d"} Feb 25 08:50:40 crc kubenswrapper[4978]: I0225 08:50:40.756322 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6bd498f7-e4ee-4062-8db7-cade3fe823af","Type":"ContainerStarted","Data":"c7e0c3e8c0b631738c6e17160e2aac02917788ab94e7ccbb125def849f0b66aa"} Feb 25 08:50:40 crc kubenswrapper[4978]: I0225 08:50:40.791758 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.791733715 podStartE2EDuration="7.791733715s" podCreationTimestamp="2026-02-25 08:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:50:40.780028042 +0000 UTC m=+7534.219284521" watchObservedRunningTime="2026-02-25 08:50:40.791733715 +0000 UTC m=+7534.230990174" Feb 25 08:50:41 crc kubenswrapper[4978]: I0225 08:50:41.773133 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"84c35447-f585-4c1d-b04e-a0e6a86b8541","Type":"ContainerStarted","Data":"31788f5f9f98f3c2a838569442af76af2352cb3692140b2d4637fa439b6616cf"} Feb 25 08:50:41 crc kubenswrapper[4978]: I0225 08:50:41.808995 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.808969369 podStartE2EDuration="8.808969369s" podCreationTimestamp="2026-02-25 08:50:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:50:41.807702151 +0000 UTC m=+7535.246958640" watchObservedRunningTime="2026-02-25 08:50:41.808969369 +0000 UTC m=+7535.248225838" Feb 25 08:50:41 crc kubenswrapper[4978]: I0225 08:50:41.964501 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:41 crc kubenswrapper[4978]: I0225 08:50:41.964805 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:50:42 crc kubenswrapper[4978]: I0225 08:50:42.163548 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:42 crc kubenswrapper[4978]: I0225 08:50:42.163596 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.038745 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.038828 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.058660 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.058926 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.080504 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.097012 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.098700 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.145856 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.818964 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.819107 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.819139 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 08:50:44 crc kubenswrapper[4978]: I0225 08:50:44.819151 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 25 08:50:46 crc kubenswrapper[4978]: I0225 08:50:46.540609 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:50:46 crc kubenswrapper[4978]: I0225 08:50:46.540940 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:50:47 crc kubenswrapper[4978]: I0225 08:50:47.079703 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 08:50:47 crc kubenswrapper[4978]: I0225 08:50:47.083520 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 25 08:50:47 crc kubenswrapper[4978]: I0225 08:50:47.924992 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:48 crc kubenswrapper[4978]: I0225 08:50:48.007098 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 25 08:50:49 crc kubenswrapper[4978]: I0225 08:50:49.663334 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:50:51 crc kubenswrapper[4978]: I0225 08:50:51.965744 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7cf5bdb968-klhbw" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.168:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.168:8443: connect: connection refused" Feb 25 08:50:52 crc kubenswrapper[4978]: I0225 08:50:52.165227 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6994658d78-bcdzg" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.167:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.167:8443: connect: connection refused" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.134064 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2g6w4"] Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.136557 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.159210 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2g6w4"] Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.160171 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kqdg\" (UniqueName: \"kubernetes.io/projected/116f38fe-742e-41e2-9ae0-bff8f9105e49-kube-api-access-6kqdg\") pod \"community-operators-2g6w4\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.160251 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-catalog-content\") pod \"community-operators-2g6w4\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.160300 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-utilities\") pod \"community-operators-2g6w4\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.264215 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kqdg\" (UniqueName: \"kubernetes.io/projected/116f38fe-742e-41e2-9ae0-bff8f9105e49-kube-api-access-6kqdg\") pod \"community-operators-2g6w4\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.264350 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-catalog-content\") pod \"community-operators-2g6w4\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.264445 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-utilities\") pod \"community-operators-2g6w4\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.265009 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-utilities\") pod \"community-operators-2g6w4\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.265258 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-catalog-content\") pod \"community-operators-2g6w4\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.287277 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kqdg\" (UniqueName: \"kubernetes.io/projected/116f38fe-742e-41e2-9ae0-bff8f9105e49-kube-api-access-6kqdg\") pod \"community-operators-2g6w4\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:55 crc kubenswrapper[4978]: I0225 08:50:55.459211 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:50:56 crc kubenswrapper[4978]: I0225 08:50:56.053670 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2g6w4"] Feb 25 08:50:56 crc kubenswrapper[4978]: I0225 08:50:56.957341 4978 generic.go:334] "Generic (PLEG): container finished" podID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerID="3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a" exitCode=0 Feb 25 08:50:56 crc kubenswrapper[4978]: I0225 08:50:56.957397 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2g6w4" event={"ID":"116f38fe-742e-41e2-9ae0-bff8f9105e49","Type":"ContainerDied","Data":"3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a"} Feb 25 08:50:56 crc kubenswrapper[4978]: I0225 08:50:56.957628 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2g6w4" event={"ID":"116f38fe-742e-41e2-9ae0-bff8f9105e49","Type":"ContainerStarted","Data":"980fd520a5d4274b55ea71c788bcfc42dc264ab8f4f0b500edb501972a6a55b4"} Feb 25 08:50:57 crc kubenswrapper[4978]: I0225 08:50:57.967119 4978 generic.go:334] "Generic (PLEG): container finished" podID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerID="62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a" exitCode=0 Feb 25 08:50:57 crc kubenswrapper[4978]: I0225 08:50:57.967157 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2g6w4" event={"ID":"116f38fe-742e-41e2-9ae0-bff8f9105e49","Type":"ContainerDied","Data":"62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a"} Feb 25 08:50:58 crc kubenswrapper[4978]: I0225 08:50:58.978325 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2g6w4" event={"ID":"116f38fe-742e-41e2-9ae0-bff8f9105e49","Type":"ContainerStarted","Data":"c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442"} Feb 25 08:50:58 crc kubenswrapper[4978]: I0225 08:50:58.995651 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2g6w4" podStartSLOduration=2.314807674 podStartE2EDuration="3.995633948s" podCreationTimestamp="2026-02-25 08:50:55 +0000 UTC" firstStartedPulling="2026-02-25 08:50:56.95994168 +0000 UTC m=+7550.399198139" lastFinishedPulling="2026-02-25 08:50:58.640767954 +0000 UTC m=+7552.080024413" observedRunningTime="2026-02-25 08:50:58.994212954 +0000 UTC m=+7552.433469423" watchObservedRunningTime="2026-02-25 08:50:58.995633948 +0000 UTC m=+7552.434890397" Feb 25 08:51:04 crc kubenswrapper[4978]: I0225 08:51:04.540586 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:51:04 crc kubenswrapper[4978]: I0225 08:51:04.599052 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:51:05 crc kubenswrapper[4978]: I0225 08:51:05.459869 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:51:05 crc kubenswrapper[4978]: I0225 08:51:05.460173 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:51:05 crc kubenswrapper[4978]: I0225 08:51:05.532344 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:51:06 crc kubenswrapper[4978]: I0225 08:51:06.095726 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:51:06 crc kubenswrapper[4978]: I0225 08:51:06.153139 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2g6w4"] Feb 25 08:51:06 crc kubenswrapper[4978]: I0225 08:51:06.552127 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:51:06 crc kubenswrapper[4978]: I0225 08:51:06.567992 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:51:06 crc kubenswrapper[4978]: I0225 08:51:06.627823 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6994658d78-bcdzg"] Feb 25 08:51:07 crc kubenswrapper[4978]: I0225 08:51:07.059880 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6994658d78-bcdzg" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon-log" containerID="cri-o://43cfe060d0263ac53cbc88b9ecfbc0593efe41c3d9f37a65d6e884ed621f69b7" gracePeriod=30 Feb 25 08:51:07 crc kubenswrapper[4978]: I0225 08:51:07.059985 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6994658d78-bcdzg" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon" containerID="cri-o://be36bd631a4e297b69dcefa831029241ef977d0822fe47c6d1a8aa57777c9a5b" gracePeriod=30 Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.071357 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2g6w4" podUID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerName="registry-server" containerID="cri-o://c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442" gracePeriod=2 Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.583496 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.758134 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-utilities\") pod \"116f38fe-742e-41e2-9ae0-bff8f9105e49\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.758257 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kqdg\" (UniqueName: \"kubernetes.io/projected/116f38fe-742e-41e2-9ae0-bff8f9105e49-kube-api-access-6kqdg\") pod \"116f38fe-742e-41e2-9ae0-bff8f9105e49\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.758295 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-catalog-content\") pod \"116f38fe-742e-41e2-9ae0-bff8f9105e49\" (UID: \"116f38fe-742e-41e2-9ae0-bff8f9105e49\") " Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.759719 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-utilities" (OuterVolumeSpecName: "utilities") pod "116f38fe-742e-41e2-9ae0-bff8f9105e49" (UID: "116f38fe-742e-41e2-9ae0-bff8f9105e49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.765320 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/116f38fe-742e-41e2-9ae0-bff8f9105e49-kube-api-access-6kqdg" (OuterVolumeSpecName: "kube-api-access-6kqdg") pod "116f38fe-742e-41e2-9ae0-bff8f9105e49" (UID: "116f38fe-742e-41e2-9ae0-bff8f9105e49"). InnerVolumeSpecName "kube-api-access-6kqdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.813753 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "116f38fe-742e-41e2-9ae0-bff8f9105e49" (UID: "116f38fe-742e-41e2-9ae0-bff8f9105e49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.861222 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.861253 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kqdg\" (UniqueName: \"kubernetes.io/projected/116f38fe-742e-41e2-9ae0-bff8f9105e49-kube-api-access-6kqdg\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:08 crc kubenswrapper[4978]: I0225 08:51:08.861266 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/116f38fe-742e-41e2-9ae0-bff8f9105e49-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.085819 4978 generic.go:334] "Generic (PLEG): container finished" podID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerID="c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442" exitCode=0 Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.085850 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2g6w4" event={"ID":"116f38fe-742e-41e2-9ae0-bff8f9105e49","Type":"ContainerDied","Data":"c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442"} Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.085892 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2g6w4" event={"ID":"116f38fe-742e-41e2-9ae0-bff8f9105e49","Type":"ContainerDied","Data":"980fd520a5d4274b55ea71c788bcfc42dc264ab8f4f0b500edb501972a6a55b4"} Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.085910 4978 scope.go:117] "RemoveContainer" containerID="c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.085907 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2g6w4" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.123993 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2g6w4"] Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.131682 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2g6w4"] Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.155005 4978 scope.go:117] "RemoveContainer" containerID="62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.180244 4978 scope.go:117] "RemoveContainer" containerID="3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.218839 4978 scope.go:117] "RemoveContainer" containerID="c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442" Feb 25 08:51:09 crc kubenswrapper[4978]: E0225 08:51:09.219193 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442\": container with ID starting with c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442 not found: ID does not exist" containerID="c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.219231 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442"} err="failed to get container status \"c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442\": rpc error: code = NotFound desc = could not find container \"c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442\": container with ID starting with c1bfb9bda17b9cc5470731abd6f4c4416266bc0b62f0cd2a46220bb86a9e5442 not found: ID does not exist" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.219259 4978 scope.go:117] "RemoveContainer" containerID="62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a" Feb 25 08:51:09 crc kubenswrapper[4978]: E0225 08:51:09.219493 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a\": container with ID starting with 62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a not found: ID does not exist" containerID="62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.219517 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a"} err="failed to get container status \"62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a\": rpc error: code = NotFound desc = could not find container \"62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a\": container with ID starting with 62a6a17b7b6b04d424a3b9e99b3535a69c6cb38cac3f87b683d3adab6585aa2a not found: ID does not exist" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.219550 4978 scope.go:117] "RemoveContainer" containerID="3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a" Feb 25 08:51:09 crc kubenswrapper[4978]: E0225 08:51:09.219731 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a\": container with ID starting with 3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a not found: ID does not exist" containerID="3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.219757 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a"} err="failed to get container status \"3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a\": rpc error: code = NotFound desc = could not find container \"3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a\": container with ID starting with 3f0a2b0d41918dee5db04318115c538e2122f76223ae72545c98f14011f4dc5a not found: ID does not exist" Feb 25 08:51:09 crc kubenswrapper[4978]: I0225 08:51:09.338007 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="116f38fe-742e-41e2-9ae0-bff8f9105e49" path="/var/lib/kubelet/pods/116f38fe-742e-41e2-9ae0-bff8f9105e49/volumes" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.098777 4978 generic.go:334] "Generic (PLEG): container finished" podID="0018d938-52d1-443d-a492-c141a4a6654f" containerID="436ce56e900d6ebeaf73e1d03ed222eaf0b322089a9c8957f0d6071a5f642d87" exitCode=137 Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.099044 4978 generic.go:334] "Generic (PLEG): container finished" podID="0018d938-52d1-443d-a492-c141a4a6654f" containerID="f098cafbe86677d59b945b64a54282fcb046381f8da289257f228d608745ca00" exitCode=137 Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.098835 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69dcc5b84c-j4bkp" event={"ID":"0018d938-52d1-443d-a492-c141a4a6654f","Type":"ContainerDied","Data":"436ce56e900d6ebeaf73e1d03ed222eaf0b322089a9c8957f0d6071a5f642d87"} Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.099088 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69dcc5b84c-j4bkp" event={"ID":"0018d938-52d1-443d-a492-c141a4a6654f","Type":"ContainerDied","Data":"f098cafbe86677d59b945b64a54282fcb046381f8da289257f228d608745ca00"} Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.115964 4978 generic.go:334] "Generic (PLEG): container finished" podID="17065e03-40e6-419a-82b6-f30993bfd59a" containerID="0d36dc01a6cd595254ece55d27282f4ff5e7694aa6b385f6850e80dc1af2e441" exitCode=137 Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.115998 4978 generic.go:334] "Generic (PLEG): container finished" podID="17065e03-40e6-419a-82b6-f30993bfd59a" containerID="d84fc21fc07dca7e451337b353842cbf9f1221ba197c47c2a50c805805c451ff" exitCode=137 Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.116067 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f588cc4d7-jpzfn" event={"ID":"17065e03-40e6-419a-82b6-f30993bfd59a","Type":"ContainerDied","Data":"0d36dc01a6cd595254ece55d27282f4ff5e7694aa6b385f6850e80dc1af2e441"} Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.116100 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f588cc4d7-jpzfn" event={"ID":"17065e03-40e6-419a-82b6-f30993bfd59a","Type":"ContainerDied","Data":"d84fc21fc07dca7e451337b353842cbf9f1221ba197c47c2a50c805805c451ff"} Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.215320 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.354092 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.388986 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvqwv\" (UniqueName: \"kubernetes.io/projected/0018d938-52d1-443d-a492-c141a4a6654f-kube-api-access-kvqwv\") pod \"0018d938-52d1-443d-a492-c141a4a6654f\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.389231 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0018d938-52d1-443d-a492-c141a4a6654f-logs\") pod \"0018d938-52d1-443d-a492-c141a4a6654f\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.389274 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0018d938-52d1-443d-a492-c141a4a6654f-horizon-secret-key\") pod \"0018d938-52d1-443d-a492-c141a4a6654f\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.389401 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-scripts\") pod \"0018d938-52d1-443d-a492-c141a4a6654f\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.389532 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-config-data\") pod \"0018d938-52d1-443d-a492-c141a4a6654f\" (UID: \"0018d938-52d1-443d-a492-c141a4a6654f\") " Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.390098 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0018d938-52d1-443d-a492-c141a4a6654f-logs" (OuterVolumeSpecName: "logs") pod "0018d938-52d1-443d-a492-c141a4a6654f" (UID: "0018d938-52d1-443d-a492-c141a4a6654f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.411102 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0018d938-52d1-443d-a492-c141a4a6654f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0018d938-52d1-443d-a492-c141a4a6654f" (UID: "0018d938-52d1-443d-a492-c141a4a6654f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.416324 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0018d938-52d1-443d-a492-c141a4a6654f-kube-api-access-kvqwv" (OuterVolumeSpecName: "kube-api-access-kvqwv") pod "0018d938-52d1-443d-a492-c141a4a6654f" (UID: "0018d938-52d1-443d-a492-c141a4a6654f"). InnerVolumeSpecName "kube-api-access-kvqwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.418041 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-scripts" (OuterVolumeSpecName: "scripts") pod "0018d938-52d1-443d-a492-c141a4a6654f" (UID: "0018d938-52d1-443d-a492-c141a4a6654f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.434502 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-config-data" (OuterVolumeSpecName: "config-data") pod "0018d938-52d1-443d-a492-c141a4a6654f" (UID: "0018d938-52d1-443d-a492-c141a4a6654f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.491097 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17065e03-40e6-419a-82b6-f30993bfd59a-horizon-secret-key\") pod \"17065e03-40e6-419a-82b6-f30993bfd59a\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.491189 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17065e03-40e6-419a-82b6-f30993bfd59a-logs\") pod \"17065e03-40e6-419a-82b6-f30993bfd59a\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.491237 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-scripts\") pod \"17065e03-40e6-419a-82b6-f30993bfd59a\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.491304 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-config-data\") pod \"17065e03-40e6-419a-82b6-f30993bfd59a\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.491389 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqgdg\" (UniqueName: \"kubernetes.io/projected/17065e03-40e6-419a-82b6-f30993bfd59a-kube-api-access-mqgdg\") pod \"17065e03-40e6-419a-82b6-f30993bfd59a\" (UID: \"17065e03-40e6-419a-82b6-f30993bfd59a\") " Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.491691 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17065e03-40e6-419a-82b6-f30993bfd59a-logs" (OuterVolumeSpecName: "logs") pod "17065e03-40e6-419a-82b6-f30993bfd59a" (UID: "17065e03-40e6-419a-82b6-f30993bfd59a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.491749 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0018d938-52d1-443d-a492-c141a4a6654f-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.491835 4978 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0018d938-52d1-443d-a492-c141a4a6654f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.491898 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.491958 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0018d938-52d1-443d-a492-c141a4a6654f-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.492011 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvqwv\" (UniqueName: \"kubernetes.io/projected/0018d938-52d1-443d-a492-c141a4a6654f-kube-api-access-kvqwv\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.495498 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17065e03-40e6-419a-82b6-f30993bfd59a-kube-api-access-mqgdg" (OuterVolumeSpecName: "kube-api-access-mqgdg") pod "17065e03-40e6-419a-82b6-f30993bfd59a" (UID: "17065e03-40e6-419a-82b6-f30993bfd59a"). InnerVolumeSpecName "kube-api-access-mqgdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.496698 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17065e03-40e6-419a-82b6-f30993bfd59a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "17065e03-40e6-419a-82b6-f30993bfd59a" (UID: "17065e03-40e6-419a-82b6-f30993bfd59a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.513113 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-config-data" (OuterVolumeSpecName: "config-data") pod "17065e03-40e6-419a-82b6-f30993bfd59a" (UID: "17065e03-40e6-419a-82b6-f30993bfd59a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.518969 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-scripts" (OuterVolumeSpecName: "scripts") pod "17065e03-40e6-419a-82b6-f30993bfd59a" (UID: "17065e03-40e6-419a-82b6-f30993bfd59a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.593674 4978 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/17065e03-40e6-419a-82b6-f30993bfd59a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.593730 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17065e03-40e6-419a-82b6-f30993bfd59a-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.593751 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.593768 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/17065e03-40e6-419a-82b6-f30993bfd59a-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:10 crc kubenswrapper[4978]: I0225 08:51:10.593785 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqgdg\" (UniqueName: \"kubernetes.io/projected/17065e03-40e6-419a-82b6-f30993bfd59a-kube-api-access-mqgdg\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.129600 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69dcc5b84c-j4bkp" event={"ID":"0018d938-52d1-443d-a492-c141a4a6654f","Type":"ContainerDied","Data":"13d5b5238a9182f8e99f29692e3bc07e4b96b1cfc975cef823575e7672b6a8e7"} Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.129924 4978 scope.go:117] "RemoveContainer" containerID="436ce56e900d6ebeaf73e1d03ed222eaf0b322089a9c8957f0d6071a5f642d87" Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.129945 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69dcc5b84c-j4bkp" Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.132285 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f588cc4d7-jpzfn" event={"ID":"17065e03-40e6-419a-82b6-f30993bfd59a","Type":"ContainerDied","Data":"460e327b9bc01b11de80c7235e72e718bc481648796ae60ade3445bfc82f801d"} Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.132312 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f588cc4d7-jpzfn" Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.137782 4978 generic.go:334] "Generic (PLEG): container finished" podID="e8cba710-0510-48af-a7e5-40285f95acb2" containerID="be36bd631a4e297b69dcefa831029241ef977d0822fe47c6d1a8aa57777c9a5b" exitCode=0 Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.137821 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6994658d78-bcdzg" event={"ID":"e8cba710-0510-48af-a7e5-40285f95acb2","Type":"ContainerDied","Data":"be36bd631a4e297b69dcefa831029241ef977d0822fe47c6d1a8aa57777c9a5b"} Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.176408 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f588cc4d7-jpzfn"] Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.201451 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-f588cc4d7-jpzfn"] Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.212429 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69dcc5b84c-j4bkp"] Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.238727 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-69dcc5b84c-j4bkp"] Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.338957 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0018d938-52d1-443d-a492-c141a4a6654f" path="/var/lib/kubelet/pods/0018d938-52d1-443d-a492-c141a4a6654f/volumes" Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.340171 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17065e03-40e6-419a-82b6-f30993bfd59a" path="/var/lib/kubelet/pods/17065e03-40e6-419a-82b6-f30993bfd59a/volumes" Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.364819 4978 scope.go:117] "RemoveContainer" containerID="f098cafbe86677d59b945b64a54282fcb046381f8da289257f228d608745ca00" Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.390506 4978 scope.go:117] "RemoveContainer" containerID="0d36dc01a6cd595254ece55d27282f4ff5e7694aa6b385f6850e80dc1af2e441" Feb 25 08:51:11 crc kubenswrapper[4978]: I0225 08:51:11.615178 4978 scope.go:117] "RemoveContainer" containerID="d84fc21fc07dca7e451337b353842cbf9f1221ba197c47c2a50c805805c451ff" Feb 25 08:51:12 crc kubenswrapper[4978]: I0225 08:51:12.163457 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6994658d78-bcdzg" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.167:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.167:8443: connect: connection refused" Feb 25 08:51:16 crc kubenswrapper[4978]: I0225 08:51:16.541307 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:51:16 crc kubenswrapper[4978]: I0225 08:51:16.542075 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:51:22 crc kubenswrapper[4978]: I0225 08:51:22.163516 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6994658d78-bcdzg" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.167:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.167:8443: connect: connection refused" Feb 25 08:51:30 crc kubenswrapper[4978]: E0225 08:51:30.130714 4978 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/3a27625777f6476aaca94eb8fe0e41b21e1045dd2d3da81c8fc7faa67b3b5a57/diff" to get inode usage: stat /var/lib/containers/storage/overlay/3a27625777f6476aaca94eb8fe0e41b21e1045dd2d3da81c8fc7faa67b3b5a57/diff: no such file or directory, extraDiskErr: Feb 25 08:51:32 crc kubenswrapper[4978]: I0225 08:51:32.164230 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6994658d78-bcdzg" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.167:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.167:8443: connect: connection refused" Feb 25 08:51:32 crc kubenswrapper[4978]: I0225 08:51:32.164554 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:51:35 crc kubenswrapper[4978]: I0225 08:51:35.072964 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-62mng"] Feb 25 08:51:35 crc kubenswrapper[4978]: I0225 08:51:35.087191 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-fe52-account-create-update-fbhjz"] Feb 25 08:51:35 crc kubenswrapper[4978]: I0225 08:51:35.101328 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-62mng"] Feb 25 08:51:35 crc kubenswrapper[4978]: I0225 08:51:35.112431 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-fe52-account-create-update-fbhjz"] Feb 25 08:51:35 crc kubenswrapper[4978]: I0225 08:51:35.342268 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f1472a5-fa2d-42d9-8a56-7ea61b2190f4" path="/var/lib/kubelet/pods/2f1472a5-fa2d-42d9-8a56-7ea61b2190f4/volumes" Feb 25 08:51:35 crc kubenswrapper[4978]: I0225 08:51:35.343804 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62fca817-3596-45b2-9649-aec49cee89be" path="/var/lib/kubelet/pods/62fca817-3596-45b2-9649-aec49cee89be/volumes" Feb 25 08:51:37 crc kubenswrapper[4978]: E0225 08:51:37.096808 4978 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0018d938_52d1_443d_a492_c141a4a6654f.slice/crio-13d5b5238a9182f8e99f29692e3bc07e4b96b1cfc975cef823575e7672b6a8e7: Error finding container 13d5b5238a9182f8e99f29692e3bc07e4b96b1cfc975cef823575e7672b6a8e7: Status 404 returned error can't find the container with id 13d5b5238a9182f8e99f29692e3bc07e4b96b1cfc975cef823575e7672b6a8e7 Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.429843 4978 generic.go:334] "Generic (PLEG): container finished" podID="e8cba710-0510-48af-a7e5-40285f95acb2" containerID="43cfe060d0263ac53cbc88b9ecfbc0593efe41c3d9f37a65d6e884ed621f69b7" exitCode=137 Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.430014 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6994658d78-bcdzg" event={"ID":"e8cba710-0510-48af-a7e5-40285f95acb2","Type":"ContainerDied","Data":"43cfe060d0263ac53cbc88b9ecfbc0593efe41c3d9f37a65d6e884ed621f69b7"} Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.568048 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.691454 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-scripts\") pod \"e8cba710-0510-48af-a7e5-40285f95acb2\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.691532 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lwqp\" (UniqueName: \"kubernetes.io/projected/e8cba710-0510-48af-a7e5-40285f95acb2-kube-api-access-6lwqp\") pod \"e8cba710-0510-48af-a7e5-40285f95acb2\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.691647 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-combined-ca-bundle\") pod \"e8cba710-0510-48af-a7e5-40285f95acb2\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.691678 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-secret-key\") pod \"e8cba710-0510-48af-a7e5-40285f95acb2\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.691714 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8cba710-0510-48af-a7e5-40285f95acb2-logs\") pod \"e8cba710-0510-48af-a7e5-40285f95acb2\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.691812 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-config-data\") pod \"e8cba710-0510-48af-a7e5-40285f95acb2\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.691836 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-tls-certs\") pod \"e8cba710-0510-48af-a7e5-40285f95acb2\" (UID: \"e8cba710-0510-48af-a7e5-40285f95acb2\") " Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.692307 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8cba710-0510-48af-a7e5-40285f95acb2-logs" (OuterVolumeSpecName: "logs") pod "e8cba710-0510-48af-a7e5-40285f95acb2" (UID: "e8cba710-0510-48af-a7e5-40285f95acb2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.698575 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8cba710-0510-48af-a7e5-40285f95acb2-kube-api-access-6lwqp" (OuterVolumeSpecName: "kube-api-access-6lwqp") pod "e8cba710-0510-48af-a7e5-40285f95acb2" (UID: "e8cba710-0510-48af-a7e5-40285f95acb2"). InnerVolumeSpecName "kube-api-access-6lwqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.702687 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e8cba710-0510-48af-a7e5-40285f95acb2" (UID: "e8cba710-0510-48af-a7e5-40285f95acb2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.723827 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8cba710-0510-48af-a7e5-40285f95acb2" (UID: "e8cba710-0510-48af-a7e5-40285f95acb2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.733891 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-config-data" (OuterVolumeSpecName: "config-data") pod "e8cba710-0510-48af-a7e5-40285f95acb2" (UID: "e8cba710-0510-48af-a7e5-40285f95acb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.734501 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-scripts" (OuterVolumeSpecName: "scripts") pod "e8cba710-0510-48af-a7e5-40285f95acb2" (UID: "e8cba710-0510-48af-a7e5-40285f95acb2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.777129 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "e8cba710-0510-48af-a7e5-40285f95acb2" (UID: "e8cba710-0510-48af-a7e5-40285f95acb2"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.794616 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.794672 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lwqp\" (UniqueName: \"kubernetes.io/projected/e8cba710-0510-48af-a7e5-40285f95acb2-kube-api-access-6lwqp\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.794712 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.794728 4978 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.794793 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8cba710-0510-48af-a7e5-40285f95acb2-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.794805 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e8cba710-0510-48af-a7e5-40285f95acb2-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:37 crc kubenswrapper[4978]: I0225 08:51:37.794822 4978 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8cba710-0510-48af-a7e5-40285f95acb2-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:38 crc kubenswrapper[4978]: I0225 08:51:38.446194 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6994658d78-bcdzg" event={"ID":"e8cba710-0510-48af-a7e5-40285f95acb2","Type":"ContainerDied","Data":"b41156409c16489cd08ccbc1de0a2fc3ef330f4f13b2099f3bcbf5ef63541bbf"} Feb 25 08:51:38 crc kubenswrapper[4978]: I0225 08:51:38.446543 4978 scope.go:117] "RemoveContainer" containerID="be36bd631a4e297b69dcefa831029241ef977d0822fe47c6d1a8aa57777c9a5b" Feb 25 08:51:38 crc kubenswrapper[4978]: I0225 08:51:38.446287 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6994658d78-bcdzg" Feb 25 08:51:38 crc kubenswrapper[4978]: I0225 08:51:38.505344 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6994658d78-bcdzg"] Feb 25 08:51:38 crc kubenswrapper[4978]: I0225 08:51:38.518075 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6994658d78-bcdzg"] Feb 25 08:51:38 crc kubenswrapper[4978]: I0225 08:51:38.700636 4978 scope.go:117] "RemoveContainer" containerID="43cfe060d0263ac53cbc88b9ecfbc0593efe41c3d9f37a65d6e884ed621f69b7" Feb 25 08:51:39 crc kubenswrapper[4978]: I0225 08:51:39.350115 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" path="/var/lib/kubelet/pods/e8cba710-0510-48af-a7e5-40285f95acb2/volumes" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.045522 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-8dpnv"] Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.058219 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-8dpnv"] Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454015 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d9787759d-6l97q"] Feb 25 08:51:46 crc kubenswrapper[4978]: E0225 08:51:46.454383 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0018d938-52d1-443d-a492-c141a4a6654f" containerName="horizon-log" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454397 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0018d938-52d1-443d-a492-c141a4a6654f" containerName="horizon-log" Feb 25 08:51:46 crc kubenswrapper[4978]: E0225 08:51:46.454479 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerName="extract-content" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454488 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerName="extract-content" Feb 25 08:51:46 crc kubenswrapper[4978]: E0225 08:51:46.454497 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerName="registry-server" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454503 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerName="registry-server" Feb 25 08:51:46 crc kubenswrapper[4978]: E0225 08:51:46.454517 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerName="extract-utilities" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454523 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerName="extract-utilities" Feb 25 08:51:46 crc kubenswrapper[4978]: E0225 08:51:46.454542 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17065e03-40e6-419a-82b6-f30993bfd59a" containerName="horizon-log" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454548 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="17065e03-40e6-419a-82b6-f30993bfd59a" containerName="horizon-log" Feb 25 08:51:46 crc kubenswrapper[4978]: E0225 08:51:46.454558 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454565 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon" Feb 25 08:51:46 crc kubenswrapper[4978]: E0225 08:51:46.454579 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17065e03-40e6-419a-82b6-f30993bfd59a" containerName="horizon" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454585 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="17065e03-40e6-419a-82b6-f30993bfd59a" containerName="horizon" Feb 25 08:51:46 crc kubenswrapper[4978]: E0225 08:51:46.454598 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon-log" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454603 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon-log" Feb 25 08:51:46 crc kubenswrapper[4978]: E0225 08:51:46.454614 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0018d938-52d1-443d-a492-c141a4a6654f" containerName="horizon" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454619 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0018d938-52d1-443d-a492-c141a4a6654f" containerName="horizon" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454782 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon-log" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454797 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0018d938-52d1-443d-a492-c141a4a6654f" containerName="horizon-log" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454813 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8cba710-0510-48af-a7e5-40285f95acb2" containerName="horizon" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454823 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0018d938-52d1-443d-a492-c141a4a6654f" containerName="horizon" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454834 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="17065e03-40e6-419a-82b6-f30993bfd59a" containerName="horizon" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454840 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="116f38fe-742e-41e2-9ae0-bff8f9105e49" containerName="registry-server" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.454848 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="17065e03-40e6-419a-82b6-f30993bfd59a" containerName="horizon-log" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.455768 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.472893 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d9787759d-6l97q"] Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.540690 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.540753 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.540809 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.541690 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c572bfb7b460a06e6b3fd8562f5f892b5fdb169c89eb0df815351c3f11f58b78"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.541768 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://c572bfb7b460a06e6b3fd8562f5f892b5fdb169c89eb0df815351c3f11f58b78" gracePeriod=600 Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.581868 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4ea666a-1e63-421a-a9fe-c085feabb6ff-horizon-secret-key\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.581954 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4ea666a-1e63-421a-a9fe-c085feabb6ff-scripts\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.582039 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ea666a-1e63-421a-a9fe-c085feabb6ff-combined-ca-bundle\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.582069 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ea666a-1e63-421a-a9fe-c085feabb6ff-logs\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.582093 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4ea666a-1e63-421a-a9fe-c085feabb6ff-config-data\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.582140 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crzx2\" (UniqueName: \"kubernetes.io/projected/f4ea666a-1e63-421a-a9fe-c085feabb6ff-kube-api-access-crzx2\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.582433 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ea666a-1e63-421a-a9fe-c085feabb6ff-horizon-tls-certs\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.684046 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4ea666a-1e63-421a-a9fe-c085feabb6ff-scripts\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.684182 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ea666a-1e63-421a-a9fe-c085feabb6ff-combined-ca-bundle\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.684233 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ea666a-1e63-421a-a9fe-c085feabb6ff-logs\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.684273 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4ea666a-1e63-421a-a9fe-c085feabb6ff-config-data\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.684468 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crzx2\" (UniqueName: \"kubernetes.io/projected/f4ea666a-1e63-421a-a9fe-c085feabb6ff-kube-api-access-crzx2\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.684792 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ea666a-1e63-421a-a9fe-c085feabb6ff-horizon-tls-certs\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.684907 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4ea666a-1e63-421a-a9fe-c085feabb6ff-horizon-secret-key\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.685148 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ea666a-1e63-421a-a9fe-c085feabb6ff-logs\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.685419 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4ea666a-1e63-421a-a9fe-c085feabb6ff-scripts\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.690037 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f4ea666a-1e63-421a-a9fe-c085feabb6ff-config-data\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.690994 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f4ea666a-1e63-421a-a9fe-c085feabb6ff-horizon-secret-key\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.691298 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ea666a-1e63-421a-a9fe-c085feabb6ff-horizon-tls-certs\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.694854 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ea666a-1e63-421a-a9fe-c085feabb6ff-combined-ca-bundle\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.710117 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crzx2\" (UniqueName: \"kubernetes.io/projected/f4ea666a-1e63-421a-a9fe-c085feabb6ff-kube-api-access-crzx2\") pod \"horizon-6d9787759d-6l97q\" (UID: \"f4ea666a-1e63-421a-a9fe-c085feabb6ff\") " pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:46 crc kubenswrapper[4978]: I0225 08:51:46.786714 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:47 crc kubenswrapper[4978]: I0225 08:51:47.274682 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d9787759d-6l97q"] Feb 25 08:51:47 crc kubenswrapper[4978]: I0225 08:51:47.338510 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb15c2e7-425b-40cd-8cb8-27733c9fe83f" path="/var/lib/kubelet/pods/eb15c2e7-425b-40cd-8cb8-27733c9fe83f/volumes" Feb 25 08:51:47 crc kubenswrapper[4978]: I0225 08:51:47.537407 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="c572bfb7b460a06e6b3fd8562f5f892b5fdb169c89eb0df815351c3f11f58b78" exitCode=0 Feb 25 08:51:47 crc kubenswrapper[4978]: I0225 08:51:47.537481 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"c572bfb7b460a06e6b3fd8562f5f892b5fdb169c89eb0df815351c3f11f58b78"} Feb 25 08:51:47 crc kubenswrapper[4978]: I0225 08:51:47.537546 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595"} Feb 25 08:51:47 crc kubenswrapper[4978]: I0225 08:51:47.537566 4978 scope.go:117] "RemoveContainer" containerID="67cb837aa8d210f23342f9b8e99a71fca2ca116d569e920939af8d0005a87a9b" Feb 25 08:51:47 crc kubenswrapper[4978]: I0225 08:51:47.539025 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d9787759d-6l97q" event={"ID":"f4ea666a-1e63-421a-a9fe-c085feabb6ff","Type":"ContainerStarted","Data":"5ec53e4b49e0dc4de0eacb1a23b0a8334ae18e8d7addaa5f76231cae4c106976"} Feb 25 08:51:47 crc kubenswrapper[4978]: I0225 08:51:47.539052 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d9787759d-6l97q" event={"ID":"f4ea666a-1e63-421a-a9fe-c085feabb6ff","Type":"ContainerStarted","Data":"55b7a1abd20673c0330c915e59d5d77e63b6ecb58d6a96f42119f67505b7ef33"} Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.017988 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-96h4w"] Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.019885 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-96h4w" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.035215 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-96h4w"] Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.115462 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-operator-scripts\") pod \"heat-db-create-96h4w\" (UID: \"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7\") " pod="openstack/heat-db-create-96h4w" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.115536 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqkgq\" (UniqueName: \"kubernetes.io/projected/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-kube-api-access-dqkgq\") pod \"heat-db-create-96h4w\" (UID: \"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7\") " pod="openstack/heat-db-create-96h4w" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.124159 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-4f6e-account-create-update-ldpx7"] Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.125331 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4f6e-account-create-update-ldpx7" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.132631 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.142983 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-4f6e-account-create-update-ldpx7"] Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.216877 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvqpv\" (UniqueName: \"kubernetes.io/projected/eff4497f-701f-4a54-afd0-b34020a27d89-kube-api-access-fvqpv\") pod \"heat-4f6e-account-create-update-ldpx7\" (UID: \"eff4497f-701f-4a54-afd0-b34020a27d89\") " pod="openstack/heat-4f6e-account-create-update-ldpx7" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.216952 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eff4497f-701f-4a54-afd0-b34020a27d89-operator-scripts\") pod \"heat-4f6e-account-create-update-ldpx7\" (UID: \"eff4497f-701f-4a54-afd0-b34020a27d89\") " pod="openstack/heat-4f6e-account-create-update-ldpx7" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.217109 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-operator-scripts\") pod \"heat-db-create-96h4w\" (UID: \"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7\") " pod="openstack/heat-db-create-96h4w" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.217210 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqkgq\" (UniqueName: \"kubernetes.io/projected/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-kube-api-access-dqkgq\") pod \"heat-db-create-96h4w\" (UID: \"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7\") " pod="openstack/heat-db-create-96h4w" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.217929 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-operator-scripts\") pod \"heat-db-create-96h4w\" (UID: \"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7\") " pod="openstack/heat-db-create-96h4w" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.246195 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqkgq\" (UniqueName: \"kubernetes.io/projected/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-kube-api-access-dqkgq\") pod \"heat-db-create-96h4w\" (UID: \"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7\") " pod="openstack/heat-db-create-96h4w" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.319058 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvqpv\" (UniqueName: \"kubernetes.io/projected/eff4497f-701f-4a54-afd0-b34020a27d89-kube-api-access-fvqpv\") pod \"heat-4f6e-account-create-update-ldpx7\" (UID: \"eff4497f-701f-4a54-afd0-b34020a27d89\") " pod="openstack/heat-4f6e-account-create-update-ldpx7" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.319118 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eff4497f-701f-4a54-afd0-b34020a27d89-operator-scripts\") pod \"heat-4f6e-account-create-update-ldpx7\" (UID: \"eff4497f-701f-4a54-afd0-b34020a27d89\") " pod="openstack/heat-4f6e-account-create-update-ldpx7" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.320289 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eff4497f-701f-4a54-afd0-b34020a27d89-operator-scripts\") pod \"heat-4f6e-account-create-update-ldpx7\" (UID: \"eff4497f-701f-4a54-afd0-b34020a27d89\") " pod="openstack/heat-4f6e-account-create-update-ldpx7" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.346826 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-96h4w" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.347100 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvqpv\" (UniqueName: \"kubernetes.io/projected/eff4497f-701f-4a54-afd0-b34020a27d89-kube-api-access-fvqpv\") pod \"heat-4f6e-account-create-update-ldpx7\" (UID: \"eff4497f-701f-4a54-afd0-b34020a27d89\") " pod="openstack/heat-4f6e-account-create-update-ldpx7" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.440142 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4f6e-account-create-update-ldpx7" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.568874 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d9787759d-6l97q" event={"ID":"f4ea666a-1e63-421a-a9fe-c085feabb6ff","Type":"ContainerStarted","Data":"f3222e4cea3d3eb93b1e262b04aa003235ec1e08b2378cb0d5fdc8660bae72d2"} Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.593776 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d9787759d-6l97q" podStartSLOduration=2.593750892 podStartE2EDuration="2.593750892s" podCreationTimestamp="2026-02-25 08:51:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:51:48.589079656 +0000 UTC m=+7602.028336125" watchObservedRunningTime="2026-02-25 08:51:48.593750892 +0000 UTC m=+7602.033007351" Feb 25 08:51:48 crc kubenswrapper[4978]: I0225 08:51:48.808226 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-96h4w"] Feb 25 08:51:48 crc kubenswrapper[4978]: W0225 08:51:48.823008 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a0d295e_5470_4e5d_aeac_04bfd4acb5a7.slice/crio-7635b3efc45b0ef062a39626e83973242c081186a27c9c8805271329217aa046 WatchSource:0}: Error finding container 7635b3efc45b0ef062a39626e83973242c081186a27c9c8805271329217aa046: Status 404 returned error can't find the container with id 7635b3efc45b0ef062a39626e83973242c081186a27c9c8805271329217aa046 Feb 25 08:51:49 crc kubenswrapper[4978]: W0225 08:51:49.007859 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeff4497f_701f_4a54_afd0_b34020a27d89.slice/crio-52c791e77d8e3f3208984db5047f7504a597cfcd1eddd6451cb309f5acbe6098 WatchSource:0}: Error finding container 52c791e77d8e3f3208984db5047f7504a597cfcd1eddd6451cb309f5acbe6098: Status 404 returned error can't find the container with id 52c791e77d8e3f3208984db5047f7504a597cfcd1eddd6451cb309f5acbe6098 Feb 25 08:51:49 crc kubenswrapper[4978]: I0225 08:51:49.009792 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-4f6e-account-create-update-ldpx7"] Feb 25 08:51:49 crc kubenswrapper[4978]: I0225 08:51:49.583894 4978 generic.go:334] "Generic (PLEG): container finished" podID="eff4497f-701f-4a54-afd0-b34020a27d89" containerID="ad73f201cbe54de15f3284657aad0c76020e7e4c4bf5c97e4612bbc37d4e60c2" exitCode=0 Feb 25 08:51:49 crc kubenswrapper[4978]: I0225 08:51:49.583975 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-4f6e-account-create-update-ldpx7" event={"ID":"eff4497f-701f-4a54-afd0-b34020a27d89","Type":"ContainerDied","Data":"ad73f201cbe54de15f3284657aad0c76020e7e4c4bf5c97e4612bbc37d4e60c2"} Feb 25 08:51:49 crc kubenswrapper[4978]: I0225 08:51:49.584315 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-4f6e-account-create-update-ldpx7" event={"ID":"eff4497f-701f-4a54-afd0-b34020a27d89","Type":"ContainerStarted","Data":"52c791e77d8e3f3208984db5047f7504a597cfcd1eddd6451cb309f5acbe6098"} Feb 25 08:51:49 crc kubenswrapper[4978]: I0225 08:51:49.589859 4978 generic.go:334] "Generic (PLEG): container finished" podID="9a0d295e-5470-4e5d-aeac-04bfd4acb5a7" containerID="7bde35d6ec1dfe86c589a8031091208db9f0cb65836b801550c1e40ddc7b6a88" exitCode=0 Feb 25 08:51:49 crc kubenswrapper[4978]: I0225 08:51:49.590111 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-96h4w" event={"ID":"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7","Type":"ContainerDied","Data":"7bde35d6ec1dfe86c589a8031091208db9f0cb65836b801550c1e40ddc7b6a88"} Feb 25 08:51:49 crc kubenswrapper[4978]: I0225 08:51:49.590262 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-96h4w" event={"ID":"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7","Type":"ContainerStarted","Data":"7635b3efc45b0ef062a39626e83973242c081186a27c9c8805271329217aa046"} Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.175693 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-96h4w" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.216105 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4f6e-account-create-update-ldpx7" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.296745 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvqpv\" (UniqueName: \"kubernetes.io/projected/eff4497f-701f-4a54-afd0-b34020a27d89-kube-api-access-fvqpv\") pod \"eff4497f-701f-4a54-afd0-b34020a27d89\" (UID: \"eff4497f-701f-4a54-afd0-b34020a27d89\") " Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.297826 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqkgq\" (UniqueName: \"kubernetes.io/projected/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-kube-api-access-dqkgq\") pod \"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7\" (UID: \"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7\") " Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.298058 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-operator-scripts\") pod \"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7\" (UID: \"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7\") " Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.298126 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eff4497f-701f-4a54-afd0-b34020a27d89-operator-scripts\") pod \"eff4497f-701f-4a54-afd0-b34020a27d89\" (UID: \"eff4497f-701f-4a54-afd0-b34020a27d89\") " Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.298659 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9a0d295e-5470-4e5d-aeac-04bfd4acb5a7" (UID: "9a0d295e-5470-4e5d-aeac-04bfd4acb5a7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.298681 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eff4497f-701f-4a54-afd0-b34020a27d89-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eff4497f-701f-4a54-afd0-b34020a27d89" (UID: "eff4497f-701f-4a54-afd0-b34020a27d89"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.302105 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-kube-api-access-dqkgq" (OuterVolumeSpecName: "kube-api-access-dqkgq") pod "9a0d295e-5470-4e5d-aeac-04bfd4acb5a7" (UID: "9a0d295e-5470-4e5d-aeac-04bfd4acb5a7"). InnerVolumeSpecName "kube-api-access-dqkgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.302354 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff4497f-701f-4a54-afd0-b34020a27d89-kube-api-access-fvqpv" (OuterVolumeSpecName: "kube-api-access-fvqpv") pod "eff4497f-701f-4a54-afd0-b34020a27d89" (UID: "eff4497f-701f-4a54-afd0-b34020a27d89"). InnerVolumeSpecName "kube-api-access-fvqpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.399905 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqkgq\" (UniqueName: \"kubernetes.io/projected/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-kube-api-access-dqkgq\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.399925 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.399935 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eff4497f-701f-4a54-afd0-b34020a27d89-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.399944 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvqpv\" (UniqueName: \"kubernetes.io/projected/eff4497f-701f-4a54-afd0-b34020a27d89-kube-api-access-fvqpv\") on node \"crc\" DevicePath \"\"" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.615288 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-96h4w" event={"ID":"9a0d295e-5470-4e5d-aeac-04bfd4acb5a7","Type":"ContainerDied","Data":"7635b3efc45b0ef062a39626e83973242c081186a27c9c8805271329217aa046"} Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.615335 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7635b3efc45b0ef062a39626e83973242c081186a27c9c8805271329217aa046" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.615444 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-96h4w" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.617810 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-4f6e-account-create-update-ldpx7" event={"ID":"eff4497f-701f-4a54-afd0-b34020a27d89","Type":"ContainerDied","Data":"52c791e77d8e3f3208984db5047f7504a597cfcd1eddd6451cb309f5acbe6098"} Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.617887 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52c791e77d8e3f3208984db5047f7504a597cfcd1eddd6451cb309f5acbe6098" Feb 25 08:51:51 crc kubenswrapper[4978]: I0225 08:51:51.617880 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-4f6e-account-create-update-ldpx7" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.250761 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-nndrq"] Feb 25 08:51:53 crc kubenswrapper[4978]: E0225 08:51:53.251516 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0d295e-5470-4e5d-aeac-04bfd4acb5a7" containerName="mariadb-database-create" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.251530 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0d295e-5470-4e5d-aeac-04bfd4acb5a7" containerName="mariadb-database-create" Feb 25 08:51:53 crc kubenswrapper[4978]: E0225 08:51:53.251542 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff4497f-701f-4a54-afd0-b34020a27d89" containerName="mariadb-account-create-update" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.251548 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff4497f-701f-4a54-afd0-b34020a27d89" containerName="mariadb-account-create-update" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.251746 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0d295e-5470-4e5d-aeac-04bfd4acb5a7" containerName="mariadb-database-create" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.251772 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff4497f-701f-4a54-afd0-b34020a27d89" containerName="mariadb-account-create-update" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.252664 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.255797 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-v8dpp" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.256091 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.264640 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-nndrq"] Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.342170 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-combined-ca-bundle\") pod \"heat-db-sync-nndrq\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.342246 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz6sp\" (UniqueName: \"kubernetes.io/projected/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-kube-api-access-tz6sp\") pod \"heat-db-sync-nndrq\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.342721 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-config-data\") pod \"heat-db-sync-nndrq\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.445682 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-config-data\") pod \"heat-db-sync-nndrq\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.446013 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-combined-ca-bundle\") pod \"heat-db-sync-nndrq\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.446782 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz6sp\" (UniqueName: \"kubernetes.io/projected/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-kube-api-access-tz6sp\") pod \"heat-db-sync-nndrq\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.451995 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-config-data\") pod \"heat-db-sync-nndrq\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.455418 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-combined-ca-bundle\") pod \"heat-db-sync-nndrq\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.475977 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz6sp\" (UniqueName: \"kubernetes.io/projected/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-kube-api-access-tz6sp\") pod \"heat-db-sync-nndrq\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:53 crc kubenswrapper[4978]: I0225 08:51:53.577049 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-nndrq" Feb 25 08:51:54 crc kubenswrapper[4978]: I0225 08:51:54.066968 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-nndrq"] Feb 25 08:51:54 crc kubenswrapper[4978]: W0225 08:51:54.076095 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fce5bd5_727e_4ee2_8e67_1c1c8d731f47.slice/crio-5398be757a10f8cd0579d53326c953206eb40b6752f00150ef819c74e00516c7 WatchSource:0}: Error finding container 5398be757a10f8cd0579d53326c953206eb40b6752f00150ef819c74e00516c7: Status 404 returned error can't find the container with id 5398be757a10f8cd0579d53326c953206eb40b6752f00150ef819c74e00516c7 Feb 25 08:51:54 crc kubenswrapper[4978]: I0225 08:51:54.663468 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-nndrq" event={"ID":"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47","Type":"ContainerStarted","Data":"5398be757a10f8cd0579d53326c953206eb40b6752f00150ef819c74e00516c7"} Feb 25 08:51:56 crc kubenswrapper[4978]: I0225 08:51:56.787589 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:51:56 crc kubenswrapper[4978]: I0225 08:51:56.788083 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:52:00 crc kubenswrapper[4978]: I0225 08:52:00.142703 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533492-bpxgq"] Feb 25 08:52:00 crc kubenswrapper[4978]: I0225 08:52:00.146804 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533492-bpxgq" Feb 25 08:52:00 crc kubenswrapper[4978]: I0225 08:52:00.149140 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:52:00 crc kubenswrapper[4978]: I0225 08:52:00.151909 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:52:00 crc kubenswrapper[4978]: I0225 08:52:00.152693 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533492-bpxgq"] Feb 25 08:52:00 crc kubenswrapper[4978]: I0225 08:52:00.153320 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:52:00 crc kubenswrapper[4978]: I0225 08:52:00.290975 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbd8l\" (UniqueName: \"kubernetes.io/projected/708a02e1-f441-447f-9f0d-91cf374e76d1-kube-api-access-zbd8l\") pod \"auto-csr-approver-29533492-bpxgq\" (UID: \"708a02e1-f441-447f-9f0d-91cf374e76d1\") " pod="openshift-infra/auto-csr-approver-29533492-bpxgq" Feb 25 08:52:00 crc kubenswrapper[4978]: I0225 08:52:00.393013 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbd8l\" (UniqueName: \"kubernetes.io/projected/708a02e1-f441-447f-9f0d-91cf374e76d1-kube-api-access-zbd8l\") pod \"auto-csr-approver-29533492-bpxgq\" (UID: \"708a02e1-f441-447f-9f0d-91cf374e76d1\") " pod="openshift-infra/auto-csr-approver-29533492-bpxgq" Feb 25 08:52:00 crc kubenswrapper[4978]: I0225 08:52:00.416449 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbd8l\" (UniqueName: \"kubernetes.io/projected/708a02e1-f441-447f-9f0d-91cf374e76d1-kube-api-access-zbd8l\") pod \"auto-csr-approver-29533492-bpxgq\" (UID: \"708a02e1-f441-447f-9f0d-91cf374e76d1\") " pod="openshift-infra/auto-csr-approver-29533492-bpxgq" Feb 25 08:52:00 crc kubenswrapper[4978]: I0225 08:52:00.505560 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533492-bpxgq" Feb 25 08:52:03 crc kubenswrapper[4978]: I0225 08:52:03.295906 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533492-bpxgq"] Feb 25 08:52:03 crc kubenswrapper[4978]: I0225 08:52:03.760909 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533492-bpxgq" event={"ID":"708a02e1-f441-447f-9f0d-91cf374e76d1","Type":"ContainerStarted","Data":"065143cb83d6b85396e82b23da30d06569c19684b6c4149c2f4321471240446e"} Feb 25 08:52:03 crc kubenswrapper[4978]: I0225 08:52:03.763082 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-nndrq" event={"ID":"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47","Type":"ContainerStarted","Data":"9cdd00f31fccb854d98fb2b38f6893d5de7bb91821633da5d292a7dab0404c3b"} Feb 25 08:52:03 crc kubenswrapper[4978]: I0225 08:52:03.789691 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-nndrq" podStartSLOduration=2.056654988 podStartE2EDuration="10.789669685s" podCreationTimestamp="2026-02-25 08:51:53 +0000 UTC" firstStartedPulling="2026-02-25 08:51:54.079466774 +0000 UTC m=+7607.518723253" lastFinishedPulling="2026-02-25 08:52:02.812481481 +0000 UTC m=+7616.251737950" observedRunningTime="2026-02-25 08:52:03.789057996 +0000 UTC m=+7617.228314495" watchObservedRunningTime="2026-02-25 08:52:03.789669685 +0000 UTC m=+7617.228926154" Feb 25 08:52:04 crc kubenswrapper[4978]: I0225 08:52:04.782286 4978 generic.go:334] "Generic (PLEG): container finished" podID="0fce5bd5-727e-4ee2-8e67-1c1c8d731f47" containerID="9cdd00f31fccb854d98fb2b38f6893d5de7bb91821633da5d292a7dab0404c3b" exitCode=0 Feb 25 08:52:04 crc kubenswrapper[4978]: I0225 08:52:04.782356 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-nndrq" event={"ID":"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47","Type":"ContainerDied","Data":"9cdd00f31fccb854d98fb2b38f6893d5de7bb91821633da5d292a7dab0404c3b"} Feb 25 08:52:04 crc kubenswrapper[4978]: I0225 08:52:04.790595 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533492-bpxgq" event={"ID":"708a02e1-f441-447f-9f0d-91cf374e76d1","Type":"ContainerStarted","Data":"c0d820bd7d896cdafac234504b4db3ba443257d9c44c75b6c1a402bc60277a34"} Feb 25 08:52:04 crc kubenswrapper[4978]: I0225 08:52:04.829576 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533492-bpxgq" podStartSLOduration=3.872231032 podStartE2EDuration="4.829552431s" podCreationTimestamp="2026-02-25 08:52:00 +0000 UTC" firstStartedPulling="2026-02-25 08:52:03.302211713 +0000 UTC m=+7616.741468182" lastFinishedPulling="2026-02-25 08:52:04.259533122 +0000 UTC m=+7617.698789581" observedRunningTime="2026-02-25 08:52:04.82467601 +0000 UTC m=+7618.263932509" watchObservedRunningTime="2026-02-25 08:52:04.829552431 +0000 UTC m=+7618.268808900" Feb 25 08:52:05 crc kubenswrapper[4978]: I0225 08:52:05.820939 4978 generic.go:334] "Generic (PLEG): container finished" podID="708a02e1-f441-447f-9f0d-91cf374e76d1" containerID="c0d820bd7d896cdafac234504b4db3ba443257d9c44c75b6c1a402bc60277a34" exitCode=0 Feb 25 08:52:05 crc kubenswrapper[4978]: I0225 08:52:05.821481 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533492-bpxgq" event={"ID":"708a02e1-f441-447f-9f0d-91cf374e76d1","Type":"ContainerDied","Data":"c0d820bd7d896cdafac234504b4db3ba443257d9c44c75b6c1a402bc60277a34"} Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.270418 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-nndrq" Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.436332 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz6sp\" (UniqueName: \"kubernetes.io/projected/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-kube-api-access-tz6sp\") pod \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.436505 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-combined-ca-bundle\") pod \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.436632 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-config-data\") pod \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\" (UID: \"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47\") " Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.447620 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-kube-api-access-tz6sp" (OuterVolumeSpecName: "kube-api-access-tz6sp") pod "0fce5bd5-727e-4ee2-8e67-1c1c8d731f47" (UID: "0fce5bd5-727e-4ee2-8e67-1c1c8d731f47"). InnerVolumeSpecName "kube-api-access-tz6sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.464332 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fce5bd5-727e-4ee2-8e67-1c1c8d731f47" (UID: "0fce5bd5-727e-4ee2-8e67-1c1c8d731f47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.539219 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz6sp\" (UniqueName: \"kubernetes.io/projected/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-kube-api-access-tz6sp\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.539257 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.554717 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-config-data" (OuterVolumeSpecName: "config-data") pod "0fce5bd5-727e-4ee2-8e67-1c1c8d731f47" (UID: "0fce5bd5-727e-4ee2-8e67-1c1c8d731f47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.641939 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.859701 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-nndrq" event={"ID":"0fce5bd5-727e-4ee2-8e67-1c1c8d731f47","Type":"ContainerDied","Data":"5398be757a10f8cd0579d53326c953206eb40b6752f00150ef819c74e00516c7"} Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.859747 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5398be757a10f8cd0579d53326c953206eb40b6752f00150ef819c74e00516c7" Feb 25 08:52:06 crc kubenswrapper[4978]: I0225 08:52:06.859758 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-nndrq" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.145885 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533492-bpxgq" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.253093 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbd8l\" (UniqueName: \"kubernetes.io/projected/708a02e1-f441-447f-9f0d-91cf374e76d1-kube-api-access-zbd8l\") pod \"708a02e1-f441-447f-9f0d-91cf374e76d1\" (UID: \"708a02e1-f441-447f-9f0d-91cf374e76d1\") " Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.273718 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/708a02e1-f441-447f-9f0d-91cf374e76d1-kube-api-access-zbd8l" (OuterVolumeSpecName: "kube-api-access-zbd8l") pod "708a02e1-f441-447f-9f0d-91cf374e76d1" (UID: "708a02e1-f441-447f-9f0d-91cf374e76d1"). InnerVolumeSpecName "kube-api-access-zbd8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.356230 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbd8l\" (UniqueName: \"kubernetes.io/projected/708a02e1-f441-447f-9f0d-91cf374e76d1-kube-api-access-zbd8l\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.885748 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533492-bpxgq" event={"ID":"708a02e1-f441-447f-9f0d-91cf374e76d1","Type":"ContainerDied","Data":"065143cb83d6b85396e82b23da30d06569c19684b6c4149c2f4321471240446e"} Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.886114 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="065143cb83d6b85396e82b23da30d06569c19684b6c4149c2f4321471240446e" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.886187 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533492-bpxgq" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.908451 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533486-qzncn"] Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.916088 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533486-qzncn"] Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.965261 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6c9d95c995-rfhkv"] Feb 25 08:52:07 crc kubenswrapper[4978]: E0225 08:52:07.965659 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="708a02e1-f441-447f-9f0d-91cf374e76d1" containerName="oc" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.965676 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="708a02e1-f441-447f-9f0d-91cf374e76d1" containerName="oc" Feb 25 08:52:07 crc kubenswrapper[4978]: E0225 08:52:07.965697 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fce5bd5-727e-4ee2-8e67-1c1c8d731f47" containerName="heat-db-sync" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.965704 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fce5bd5-727e-4ee2-8e67-1c1c8d731f47" containerName="heat-db-sync" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.965890 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fce5bd5-727e-4ee2-8e67-1c1c8d731f47" containerName="heat-db-sync" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.965924 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="708a02e1-f441-447f-9f0d-91cf374e76d1" containerName="oc" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.966849 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.973449 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.973658 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-v8dpp" Feb 25 08:52:07 crc kubenswrapper[4978]: I0225 08:52:07.973727 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.015351 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6c9d95c995-rfhkv"] Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.070798 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-combined-ca-bundle\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.070908 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.070977 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data-custom\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.071057 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgdbf\" (UniqueName: \"kubernetes.io/projected/1751340a-9d92-4a12-ba1e-297064818a4e-kube-api-access-mgdbf\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.151296 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5bf79684c-745q9"] Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.152637 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.160849 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.173384 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.173440 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data-custom\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.173516 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgdbf\" (UniqueName: \"kubernetes.io/projected/1751340a-9d92-4a12-ba1e-297064818a4e-kube-api-access-mgdbf\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.173543 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-combined-ca-bundle\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.181104 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6784f568f4-l8zd6"] Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.185701 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.187827 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.192083 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.199549 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-combined-ca-bundle\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.197090 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data-custom\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.207899 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgdbf\" (UniqueName: \"kubernetes.io/projected/1751340a-9d92-4a12-ba1e-297064818a4e-kube-api-access-mgdbf\") pod \"heat-engine-6c9d95c995-rfhkv\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.228045 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5bf79684c-745q9"] Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.246788 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6784f568f4-l8zd6"] Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.277219 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.277493 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data-custom\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.277519 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfcc9\" (UniqueName: \"kubernetes.io/projected/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-kube-api-access-pfcc9\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.277537 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-combined-ca-bundle\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.277584 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.277655 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data-custom\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.277674 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjlql\" (UniqueName: \"kubernetes.io/projected/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-kube-api-access-xjlql\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.277692 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-combined-ca-bundle\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.309201 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.379404 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data-custom\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.379443 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjlql\" (UniqueName: \"kubernetes.io/projected/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-kube-api-access-xjlql\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.379468 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-combined-ca-bundle\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.379526 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.379551 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data-custom\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.379576 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfcc9\" (UniqueName: \"kubernetes.io/projected/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-kube-api-access-pfcc9\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.379591 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-combined-ca-bundle\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.379642 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.385841 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-combined-ca-bundle\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.385922 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data-custom\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.391943 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.395045 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data-custom\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.395076 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.397025 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjlql\" (UniqueName: \"kubernetes.io/projected/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-kube-api-access-xjlql\") pod \"heat-api-5bf79684c-745q9\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.397971 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-combined-ca-bundle\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.398899 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfcc9\" (UniqueName: \"kubernetes.io/projected/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-kube-api-access-pfcc9\") pod \"heat-cfnapi-6784f568f4-l8zd6\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.583672 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.592229 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.598101 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.832998 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6c9d95c995-rfhkv"] Feb 25 08:52:08 crc kubenswrapper[4978]: W0225 08:52:08.835138 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1751340a_9d92_4a12_ba1e_297064818a4e.slice/crio-57ab7268ade78f37c5e203a1d99b390be36de62c77db0730ca452e25da9139aa WatchSource:0}: Error finding container 57ab7268ade78f37c5e203a1d99b390be36de62c77db0730ca452e25da9139aa: Status 404 returned error can't find the container with id 57ab7268ade78f37c5e203a1d99b390be36de62c77db0730ca452e25da9139aa Feb 25 08:52:08 crc kubenswrapper[4978]: I0225 08:52:08.898618 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6c9d95c995-rfhkv" event={"ID":"1751340a-9d92-4a12-ba1e-297064818a4e","Type":"ContainerStarted","Data":"57ab7268ade78f37c5e203a1d99b390be36de62c77db0730ca452e25da9139aa"} Feb 25 08:52:09 crc kubenswrapper[4978]: I0225 08:52:09.113864 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5bf79684c-745q9"] Feb 25 08:52:09 crc kubenswrapper[4978]: W0225 08:52:09.117125 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad9b26c4_0fc4_4c3b_a9ea_41b01f4568ec.slice/crio-b682f6c11be0196efd66b5530b183c66ddfed0e0ca5734738f2b4f56ae16a721 WatchSource:0}: Error finding container b682f6c11be0196efd66b5530b183c66ddfed0e0ca5734738f2b4f56ae16a721: Status 404 returned error can't find the container with id b682f6c11be0196efd66b5530b183c66ddfed0e0ca5734738f2b4f56ae16a721 Feb 25 08:52:09 crc kubenswrapper[4978]: I0225 08:52:09.226136 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6784f568f4-l8zd6"] Feb 25 08:52:09 crc kubenswrapper[4978]: I0225 08:52:09.339576 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2378b284-beb2-444b-aaa9-05a96e14c8c1" path="/var/lib/kubelet/pods/2378b284-beb2-444b-aaa9-05a96e14c8c1/volumes" Feb 25 08:52:09 crc kubenswrapper[4978]: I0225 08:52:09.925873 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" event={"ID":"ba8f40d8-0414-4c6d-b90f-3feeadbd0190","Type":"ContainerStarted","Data":"37c9f01f13210650b97d28858edb4e410d71c50cf3b3b05935d0f8f9fd549593"} Feb 25 08:52:09 crc kubenswrapper[4978]: I0225 08:52:09.934105 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5bf79684c-745q9" event={"ID":"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec","Type":"ContainerStarted","Data":"b682f6c11be0196efd66b5530b183c66ddfed0e0ca5734738f2b4f56ae16a721"} Feb 25 08:52:09 crc kubenswrapper[4978]: I0225 08:52:09.942658 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6c9d95c995-rfhkv" event={"ID":"1751340a-9d92-4a12-ba1e-297064818a4e","Type":"ContainerStarted","Data":"a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e"} Feb 25 08:52:09 crc kubenswrapper[4978]: I0225 08:52:09.942927 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:09 crc kubenswrapper[4978]: I0225 08:52:09.963943 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6c9d95c995-rfhkv" podStartSLOduration=2.963923138 podStartE2EDuration="2.963923138s" podCreationTimestamp="2026-02-25 08:52:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:52:09.959292915 +0000 UTC m=+7623.398549374" watchObservedRunningTime="2026-02-25 08:52:09.963923138 +0000 UTC m=+7623.403179587" Feb 25 08:52:10 crc kubenswrapper[4978]: I0225 08:52:10.294050 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6d9787759d-6l97q" Feb 25 08:52:10 crc kubenswrapper[4978]: I0225 08:52:10.356479 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cf5bdb968-klhbw"] Feb 25 08:52:10 crc kubenswrapper[4978]: I0225 08:52:10.356706 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7cf5bdb968-klhbw" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon-log" containerID="cri-o://191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b" gracePeriod=30 Feb 25 08:52:10 crc kubenswrapper[4978]: I0225 08:52:10.357079 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7cf5bdb968-klhbw" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon" containerID="cri-o://1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa" gracePeriod=30 Feb 25 08:52:11 crc kubenswrapper[4978]: I0225 08:52:11.965688 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" event={"ID":"ba8f40d8-0414-4c6d-b90f-3feeadbd0190","Type":"ContainerStarted","Data":"1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def"} Feb 25 08:52:11 crc kubenswrapper[4978]: I0225 08:52:11.966239 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:11 crc kubenswrapper[4978]: I0225 08:52:11.967696 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5bf79684c-745q9" event={"ID":"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec","Type":"ContainerStarted","Data":"62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842"} Feb 25 08:52:11 crc kubenswrapper[4978]: I0225 08:52:11.967862 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:11 crc kubenswrapper[4978]: I0225 08:52:11.981741 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" podStartSLOduration=2.215140092 podStartE2EDuration="3.981720923s" podCreationTimestamp="2026-02-25 08:52:08 +0000 UTC" firstStartedPulling="2026-02-25 08:52:09.254083147 +0000 UTC m=+7622.693339606" lastFinishedPulling="2026-02-25 08:52:11.020663978 +0000 UTC m=+7624.459920437" observedRunningTime="2026-02-25 08:52:11.97680757 +0000 UTC m=+7625.416064029" watchObservedRunningTime="2026-02-25 08:52:11.981720923 +0000 UTC m=+7625.420977392" Feb 25 08:52:11 crc kubenswrapper[4978]: I0225 08:52:11.996888 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5bf79684c-745q9" podStartSLOduration=2.104505033 podStartE2EDuration="3.996871171s" podCreationTimestamp="2026-02-25 08:52:08 +0000 UTC" firstStartedPulling="2026-02-25 08:52:09.123156031 +0000 UTC m=+7622.562412500" lastFinishedPulling="2026-02-25 08:52:11.015522179 +0000 UTC m=+7624.454778638" observedRunningTime="2026-02-25 08:52:11.992022981 +0000 UTC m=+7625.431279480" watchObservedRunningTime="2026-02-25 08:52:11.996871171 +0000 UTC m=+7625.436127630" Feb 25 08:52:13 crc kubenswrapper[4978]: I0225 08:52:13.865772 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7cf5bdb968-klhbw" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.168:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:57488->10.217.1.168:8443: read: connection reset by peer" Feb 25 08:52:13 crc kubenswrapper[4978]: I0225 08:52:13.990704 4978 generic.go:334] "Generic (PLEG): container finished" podID="01581086-b21b-4f8a-8390-df51be3a5999" containerID="1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa" exitCode=0 Feb 25 08:52:13 crc kubenswrapper[4978]: I0225 08:52:13.990753 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cf5bdb968-klhbw" event={"ID":"01581086-b21b-4f8a-8390-df51be3a5999","Type":"ContainerDied","Data":"1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa"} Feb 25 08:52:14 crc kubenswrapper[4978]: I0225 08:52:14.976525 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7d87784d56-fghgj"] Feb 25 08:52:14 crc kubenswrapper[4978]: I0225 08:52:14.977727 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:14 crc kubenswrapper[4978]: I0225 08:52:14.996899 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7d87784d56-fghgj"] Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.059401 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-57f8f4d4bc-ltxm5"] Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.060558 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.073555 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-894b45677-8r2pm"] Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.079311 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.115028 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-57f8f4d4bc-ltxm5"] Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.117877 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-combined-ca-bundle\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.117922 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggkr7\" (UniqueName: \"kubernetes.io/projected/fd0033bc-91e4-498b-9c02-0e3d67123c4a-kube-api-access-ggkr7\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.117952 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-config-data\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.117975 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.117999 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjcbn\" (UniqueName: \"kubernetes.io/projected/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-kube-api-access-kjcbn\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.118054 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.118075 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-776qn\" (UniqueName: \"kubernetes.io/projected/207dc629-cc88-460b-8159-34b8fca288cf-kube-api-access-776qn\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.118099 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data-custom\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.118164 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-config-data-custom\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.118257 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data-custom\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.118337 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-combined-ca-bundle\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.118441 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-combined-ca-bundle\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.144788 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-894b45677-8r2pm"] Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.219833 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-776qn\" (UniqueName: \"kubernetes.io/projected/207dc629-cc88-460b-8159-34b8fca288cf-kube-api-access-776qn\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.219883 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data-custom\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.219949 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-config-data-custom\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.219985 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data-custom\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.220003 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-combined-ca-bundle\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.220077 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-combined-ca-bundle\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.220105 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-combined-ca-bundle\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.220125 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggkr7\" (UniqueName: \"kubernetes.io/projected/fd0033bc-91e4-498b-9c02-0e3d67123c4a-kube-api-access-ggkr7\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.220143 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-config-data\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.220160 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.220179 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjcbn\" (UniqueName: \"kubernetes.io/projected/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-kube-api-access-kjcbn\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.220223 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.226877 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-combined-ca-bundle\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.227952 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data-custom\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.228219 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-combined-ca-bundle\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.228234 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-combined-ca-bundle\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.228723 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.229116 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data-custom\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.230117 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.231155 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-config-data-custom\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.235535 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-config-data\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.239036 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjcbn\" (UniqueName: \"kubernetes.io/projected/7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b-kube-api-access-kjcbn\") pod \"heat-engine-7d87784d56-fghgj\" (UID: \"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b\") " pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.239363 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggkr7\" (UniqueName: \"kubernetes.io/projected/fd0033bc-91e4-498b-9c02-0e3d67123c4a-kube-api-access-ggkr7\") pod \"heat-api-894b45677-8r2pm\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.250014 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-776qn\" (UniqueName: \"kubernetes.io/projected/207dc629-cc88-460b-8159-34b8fca288cf-kube-api-access-776qn\") pod \"heat-cfnapi-57f8f4d4bc-ltxm5\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.295957 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.443791 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.444825 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.713273 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7d87784d56-fghgj"] Feb 25 08:52:15 crc kubenswrapper[4978]: W0225 08:52:15.718105 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ffd8086_b8c0_410d_a4cb_e5d0c2ce634b.slice/crio-06e24338a48648f506f90efb3355bacc2cd013c00c511b2e13ebeb5453c7986a WatchSource:0}: Error finding container 06e24338a48648f506f90efb3355bacc2cd013c00c511b2e13ebeb5453c7986a: Status 404 returned error can't find the container with id 06e24338a48648f506f90efb3355bacc2cd013c00c511b2e13ebeb5453c7986a Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.918649 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-57f8f4d4bc-ltxm5"] Feb 25 08:52:15 crc kubenswrapper[4978]: W0225 08:52:15.925276 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod207dc629_cc88_460b_8159_34b8fca288cf.slice/crio-9b4ffc64005a377358236ef89fc3149e6664b0c823bfa3c0b79b97a40accfdf8 WatchSource:0}: Error finding container 9b4ffc64005a377358236ef89fc3149e6664b0c823bfa3c0b79b97a40accfdf8: Status 404 returned error can't find the container with id 9b4ffc64005a377358236ef89fc3149e6664b0c823bfa3c0b79b97a40accfdf8 Feb 25 08:52:15 crc kubenswrapper[4978]: I0225 08:52:15.926994 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-894b45677-8r2pm"] Feb 25 08:52:15 crc kubenswrapper[4978]: W0225 08:52:15.927100 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd0033bc_91e4_498b_9c02_0e3d67123c4a.slice/crio-94a3f577ca5d1dfe9df581bf724b08a752c8a47edef4e46302242a5f23c1cddf WatchSource:0}: Error finding container 94a3f577ca5d1dfe9df581bf724b08a752c8a47edef4e46302242a5f23c1cddf: Status 404 returned error can't find the container with id 94a3f577ca5d1dfe9df581bf724b08a752c8a47edef4e46302242a5f23c1cddf Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.028201 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-894b45677-8r2pm" event={"ID":"fd0033bc-91e4-498b-9c02-0e3d67123c4a","Type":"ContainerStarted","Data":"94a3f577ca5d1dfe9df581bf724b08a752c8a47edef4e46302242a5f23c1cddf"} Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.029252 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" event={"ID":"207dc629-cc88-460b-8159-34b8fca288cf","Type":"ContainerStarted","Data":"9b4ffc64005a377358236ef89fc3149e6664b0c823bfa3c0b79b97a40accfdf8"} Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.031522 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7d87784d56-fghgj" event={"ID":"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b","Type":"ContainerStarted","Data":"16b0cb585f82de72bdc2460d243fc5056f49047914a41587d68f69bfad4894d4"} Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.031575 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7d87784d56-fghgj" event={"ID":"7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b","Type":"ContainerStarted","Data":"06e24338a48648f506f90efb3355bacc2cd013c00c511b2e13ebeb5453c7986a"} Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.032657 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.059097 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5bf79684c-745q9"] Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.059332 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-5bf79684c-745q9" podUID="ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" containerName="heat-api" containerID="cri-o://62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842" gracePeriod=60 Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.086062 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7d87784d56-fghgj" podStartSLOduration=2.086043998 podStartE2EDuration="2.086043998s" podCreationTimestamp="2026-02-25 08:52:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:52:16.057350579 +0000 UTC m=+7629.496607068" watchObservedRunningTime="2026-02-25 08:52:16.086043998 +0000 UTC m=+7629.525300457" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.087344 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6784f568f4-l8zd6"] Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.087548 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" podUID="ba8f40d8-0414-4c6d-b90f-3feeadbd0190" containerName="heat-cfnapi" containerID="cri-o://1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def" gracePeriod=60 Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.105784 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" podUID="ba8f40d8-0414-4c6d-b90f-3feeadbd0190" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.179:8000/healthcheck\": EOF" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.111875 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7fc7d9db6-jkz4h"] Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.113168 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.119043 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.119188 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.123266 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5b459dcfff-rkn7k"] Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.124675 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.126085 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.126615 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.156429 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7fc7d9db6-jkz4h"] Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.168034 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5b459dcfff-rkn7k"] Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.242131 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5bf79684c-745q9" podUID="ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.178:8004/healthcheck\": read tcp 10.217.0.2:41530->10.217.1.178:8004: read: connection reset by peer" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248339 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl9k8\" (UniqueName: \"kubernetes.io/projected/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-kube-api-access-xl9k8\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248395 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-public-tls-certs\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248424 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-internal-tls-certs\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248441 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-public-tls-certs\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248471 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-config-data-custom\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248507 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-combined-ca-bundle\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248540 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-config-data-custom\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248564 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-internal-tls-certs\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248583 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-combined-ca-bundle\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248625 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-config-data\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248644 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kshg8\" (UniqueName: \"kubernetes.io/projected/128089fe-663c-42b0-80a2-b32e5793d555-kube-api-access-kshg8\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.248694 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-config-data\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.350800 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-config-data\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351097 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl9k8\" (UniqueName: \"kubernetes.io/projected/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-kube-api-access-xl9k8\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351125 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-public-tls-certs\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351145 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-internal-tls-certs\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351163 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-public-tls-certs\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351191 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-config-data-custom\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351227 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-combined-ca-bundle\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351260 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-config-data-custom\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351283 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-internal-tls-certs\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351303 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-combined-ca-bundle\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351344 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-config-data\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.351362 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kshg8\" (UniqueName: \"kubernetes.io/projected/128089fe-663c-42b0-80a2-b32e5793d555-kube-api-access-kshg8\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.359086 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-config-data\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.367356 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-config-data-custom\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.371027 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-public-tls-certs\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.372027 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-internal-tls-certs\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.374196 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-public-tls-certs\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.374216 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kshg8\" (UniqueName: \"kubernetes.io/projected/128089fe-663c-42b0-80a2-b32e5793d555-kube-api-access-kshg8\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.374745 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-combined-ca-bundle\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.375108 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-combined-ca-bundle\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.375362 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-config-data\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.376212 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-config-data-custom\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.381411 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/128089fe-663c-42b0-80a2-b32e5793d555-internal-tls-certs\") pod \"heat-api-7fc7d9db6-jkz4h\" (UID: \"128089fe-663c-42b0-80a2-b32e5793d555\") " pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.383204 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl9k8\" (UniqueName: \"kubernetes.io/projected/e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf-kube-api-access-xl9k8\") pod \"heat-cfnapi-5b459dcfff-rkn7k\" (UID: \"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf\") " pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.444821 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.461464 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.637397 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.768540 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data\") pod \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.768600 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data-custom\") pod \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.768664 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-combined-ca-bundle\") pod \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.768780 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjlql\" (UniqueName: \"kubernetes.io/projected/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-kube-api-access-xjlql\") pod \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\" (UID: \"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec\") " Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.777324 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" (UID: "ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.777406 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-kube-api-access-xjlql" (OuterVolumeSpecName: "kube-api-access-xjlql") pod "ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" (UID: "ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec"). InnerVolumeSpecName "kube-api-access-xjlql". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.800543 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" (UID: "ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.821815 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data" (OuterVolumeSpecName: "config-data") pod "ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" (UID: "ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.871086 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.871135 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.871145 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjlql\" (UniqueName: \"kubernetes.io/projected/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-kube-api-access-xjlql\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.871155 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.974916 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7fc7d9db6-jkz4h"] Feb 25 08:52:16 crc kubenswrapper[4978]: W0225 08:52:16.976993 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod128089fe_663c_42b0_80a2_b32e5793d555.slice/crio-a142b9e3289b2ac3134cffbbfcdfcc0be70f24407213d9d5a4e5e1263effe49d WatchSource:0}: Error finding container a142b9e3289b2ac3134cffbbfcdfcc0be70f24407213d9d5a4e5e1263effe49d: Status 404 returned error can't find the container with id a142b9e3289b2ac3134cffbbfcdfcc0be70f24407213d9d5a4e5e1263effe49d Feb 25 08:52:16 crc kubenswrapper[4978]: I0225 08:52:16.998545 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5b459dcfff-rkn7k"] Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.055343 4978 generic.go:334] "Generic (PLEG): container finished" podID="207dc629-cc88-460b-8159-34b8fca288cf" containerID="3cd5a79a055f71340da4336666349ef246099e0d345aec20d6fc306441f2aa8c" exitCode=1 Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.056402 4978 scope.go:117] "RemoveContainer" containerID="3cd5a79a055f71340da4336666349ef246099e0d345aec20d6fc306441f2aa8c" Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.060718 4978 generic.go:334] "Generic (PLEG): container finished" podID="ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" containerID="62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842" exitCode=0 Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.060858 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5bf79684c-745q9" Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.061074 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" event={"ID":"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf","Type":"ContainerStarted","Data":"c04cf897d0661f1d47140918dee631ddcdccf058f82571b3cd80cb890db7b73c"} Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.061132 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" event={"ID":"207dc629-cc88-460b-8159-34b8fca288cf","Type":"ContainerDied","Data":"3cd5a79a055f71340da4336666349ef246099e0d345aec20d6fc306441f2aa8c"} Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.061152 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5bf79684c-745q9" event={"ID":"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec","Type":"ContainerDied","Data":"62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842"} Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.061170 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5bf79684c-745q9" event={"ID":"ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec","Type":"ContainerDied","Data":"b682f6c11be0196efd66b5530b183c66ddfed0e0ca5734738f2b4f56ae16a721"} Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.061199 4978 scope.go:117] "RemoveContainer" containerID="62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842" Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.064686 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7fc7d9db6-jkz4h" event={"ID":"128089fe-663c-42b0-80a2-b32e5793d555","Type":"ContainerStarted","Data":"a142b9e3289b2ac3134cffbbfcdfcc0be70f24407213d9d5a4e5e1263effe49d"} Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.072351 4978 generic.go:334] "Generic (PLEG): container finished" podID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" containerID="38ddf981a4c93423ed3b83def0087af71caa3251587888f71d4b97cc5b0d94fd" exitCode=1 Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.074325 4978 scope.go:117] "RemoveContainer" containerID="38ddf981a4c93423ed3b83def0087af71caa3251587888f71d4b97cc5b0d94fd" Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.074323 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-894b45677-8r2pm" event={"ID":"fd0033bc-91e4-498b-9c02-0e3d67123c4a","Type":"ContainerDied","Data":"38ddf981a4c93423ed3b83def0087af71caa3251587888f71d4b97cc5b0d94fd"} Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.117549 4978 scope.go:117] "RemoveContainer" containerID="62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842" Feb 25 08:52:17 crc kubenswrapper[4978]: E0225 08:52:17.119862 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842\": container with ID starting with 62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842 not found: ID does not exist" containerID="62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842" Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.120108 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842"} err="failed to get container status \"62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842\": rpc error: code = NotFound desc = could not find container \"62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842\": container with ID starting with 62ca8689b55925e3d89fa729ab7f14df086a68e2d0a82dfd2eecdeed7bc9b842 not found: ID does not exist" Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.152424 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5bf79684c-745q9"] Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.161278 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5bf79684c-745q9"] Feb 25 08:52:17 crc kubenswrapper[4978]: I0225 08:52:17.337678 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" path="/var/lib/kubelet/pods/ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec/volumes" Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.108696 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7fc7d9db6-jkz4h" event={"ID":"128089fe-663c-42b0-80a2-b32e5793d555","Type":"ContainerStarted","Data":"94cc9d2f76db523c730f7106b1bc4d0b23f54afc56c5944f894dac1137bc5b88"} Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.108747 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.114639 4978 generic.go:334] "Generic (PLEG): container finished" podID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" containerID="52f26e1fbbcc8b4f76ab4368557e659aa36a878d983493e9ea230628e569c530" exitCode=1 Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.114741 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-894b45677-8r2pm" event={"ID":"fd0033bc-91e4-498b-9c02-0e3d67123c4a","Type":"ContainerDied","Data":"52f26e1fbbcc8b4f76ab4368557e659aa36a878d983493e9ea230628e569c530"} Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.114808 4978 scope.go:117] "RemoveContainer" containerID="38ddf981a4c93423ed3b83def0087af71caa3251587888f71d4b97cc5b0d94fd" Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.116031 4978 scope.go:117] "RemoveContainer" containerID="52f26e1fbbcc8b4f76ab4368557e659aa36a878d983493e9ea230628e569c530" Feb 25 08:52:18 crc kubenswrapper[4978]: E0225 08:52:18.116489 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-894b45677-8r2pm_openstack(fd0033bc-91e4-498b-9c02-0e3d67123c4a)\"" pod="openstack/heat-api-894b45677-8r2pm" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.123492 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" event={"ID":"e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf","Type":"ContainerStarted","Data":"5d53e78d4611bfc4cd58b8253aa8b81d852acdd588e3b3b572a05e4ed6b0a595"} Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.124263 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.129417 4978 generic.go:334] "Generic (PLEG): container finished" podID="207dc629-cc88-460b-8159-34b8fca288cf" containerID="3924ccbeb1105d7edb99f7895483c66bc1d515f33759ed49e1e9895226b097e8" exitCode=1 Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.129497 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" event={"ID":"207dc629-cc88-460b-8159-34b8fca288cf","Type":"ContainerDied","Data":"3924ccbeb1105d7edb99f7895483c66bc1d515f33759ed49e1e9895226b097e8"} Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.130224 4978 scope.go:117] "RemoveContainer" containerID="3924ccbeb1105d7edb99f7895483c66bc1d515f33759ed49e1e9895226b097e8" Feb 25 08:52:18 crc kubenswrapper[4978]: E0225 08:52:18.130522 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-57f8f4d4bc-ltxm5_openstack(207dc629-cc88-460b-8159-34b8fca288cf)\"" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" podUID="207dc629-cc88-460b-8159-34b8fca288cf" Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.140280 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7fc7d9db6-jkz4h" podStartSLOduration=2.14026347 podStartE2EDuration="2.14026347s" podCreationTimestamp="2026-02-25 08:52:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:52:18.12448255 +0000 UTC m=+7631.563739009" watchObservedRunningTime="2026-02-25 08:52:18.14026347 +0000 UTC m=+7631.579519929" Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.147691 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" podStartSLOduration=2.147674229 podStartE2EDuration="2.147674229s" podCreationTimestamp="2026-02-25 08:52:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:52:18.146120471 +0000 UTC m=+7631.585376940" watchObservedRunningTime="2026-02-25 08:52:18.147674229 +0000 UTC m=+7631.586930688" Feb 25 08:52:18 crc kubenswrapper[4978]: I0225 08:52:18.202148 4978 scope.go:117] "RemoveContainer" containerID="3cd5a79a055f71340da4336666349ef246099e0d345aec20d6fc306441f2aa8c" Feb 25 08:52:19 crc kubenswrapper[4978]: I0225 08:52:19.143230 4978 scope.go:117] "RemoveContainer" containerID="52f26e1fbbcc8b4f76ab4368557e659aa36a878d983493e9ea230628e569c530" Feb 25 08:52:19 crc kubenswrapper[4978]: E0225 08:52:19.143738 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-894b45677-8r2pm_openstack(fd0033bc-91e4-498b-9c02-0e3d67123c4a)\"" pod="openstack/heat-api-894b45677-8r2pm" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" Feb 25 08:52:19 crc kubenswrapper[4978]: I0225 08:52:19.144607 4978 scope.go:117] "RemoveContainer" containerID="3924ccbeb1105d7edb99f7895483c66bc1d515f33759ed49e1e9895226b097e8" Feb 25 08:52:19 crc kubenswrapper[4978]: E0225 08:52:19.144884 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-57f8f4d4bc-ltxm5_openstack(207dc629-cc88-460b-8159-34b8fca288cf)\"" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" podUID="207dc629-cc88-460b-8159-34b8fca288cf" Feb 25 08:52:20 crc kubenswrapper[4978]: I0225 08:52:20.444145 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:20 crc kubenswrapper[4978]: I0225 08:52:20.444961 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:20 crc kubenswrapper[4978]: I0225 08:52:20.445085 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:20 crc kubenswrapper[4978]: I0225 08:52:20.445462 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:20 crc kubenswrapper[4978]: I0225 08:52:20.445889 4978 scope.go:117] "RemoveContainer" containerID="52f26e1fbbcc8b4f76ab4368557e659aa36a878d983493e9ea230628e569c530" Feb 25 08:52:20 crc kubenswrapper[4978]: I0225 08:52:20.446198 4978 scope.go:117] "RemoveContainer" containerID="3924ccbeb1105d7edb99f7895483c66bc1d515f33759ed49e1e9895226b097e8" Feb 25 08:52:20 crc kubenswrapper[4978]: E0225 08:52:20.446261 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-894b45677-8r2pm_openstack(fd0033bc-91e4-498b-9c02-0e3d67123c4a)\"" pod="openstack/heat-api-894b45677-8r2pm" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" Feb 25 08:52:20 crc kubenswrapper[4978]: E0225 08:52:20.446698 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-57f8f4d4bc-ltxm5_openstack(207dc629-cc88-460b-8159-34b8fca288cf)\"" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" podUID="207dc629-cc88-460b-8159-34b8fca288cf" Feb 25 08:52:21 crc kubenswrapper[4978]: I0225 08:52:21.162740 4978 scope.go:117] "RemoveContainer" containerID="3924ccbeb1105d7edb99f7895483c66bc1d515f33759ed49e1e9895226b097e8" Feb 25 08:52:21 crc kubenswrapper[4978]: I0225 08:52:21.163120 4978 scope.go:117] "RemoveContainer" containerID="52f26e1fbbcc8b4f76ab4368557e659aa36a878d983493e9ea230628e569c530" Feb 25 08:52:21 crc kubenswrapper[4978]: E0225 08:52:21.163300 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-57f8f4d4bc-ltxm5_openstack(207dc629-cc88-460b-8159-34b8fca288cf)\"" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" podUID="207dc629-cc88-460b-8159-34b8fca288cf" Feb 25 08:52:21 crc kubenswrapper[4978]: E0225 08:52:21.163351 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-894b45677-8r2pm_openstack(fd0033bc-91e4-498b-9c02-0e3d67123c4a)\"" pod="openstack/heat-api-894b45677-8r2pm" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" Feb 25 08:52:21 crc kubenswrapper[4978]: I0225 08:52:21.477145 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" podUID="ba8f40d8-0414-4c6d-b90f-3feeadbd0190" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.179:8000/healthcheck\": read tcp 10.217.0.2:35186->10.217.1.179:8000: read: connection reset by peer" Feb 25 08:52:21 crc kubenswrapper[4978]: I0225 08:52:21.964960 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7cf5bdb968-klhbw" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.168:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.168:8443: connect: connection refused" Feb 25 08:52:21 crc kubenswrapper[4978]: I0225 08:52:21.989460 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.083321 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data-custom\") pod \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.083448 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-combined-ca-bundle\") pod \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.083687 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfcc9\" (UniqueName: \"kubernetes.io/projected/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-kube-api-access-pfcc9\") pod \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.083759 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data\") pod \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\" (UID: \"ba8f40d8-0414-4c6d-b90f-3feeadbd0190\") " Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.089688 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ba8f40d8-0414-4c6d-b90f-3feeadbd0190" (UID: "ba8f40d8-0414-4c6d-b90f-3feeadbd0190"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.091033 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-kube-api-access-pfcc9" (OuterVolumeSpecName: "kube-api-access-pfcc9") pod "ba8f40d8-0414-4c6d-b90f-3feeadbd0190" (UID: "ba8f40d8-0414-4c6d-b90f-3feeadbd0190"). InnerVolumeSpecName "kube-api-access-pfcc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.134545 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba8f40d8-0414-4c6d-b90f-3feeadbd0190" (UID: "ba8f40d8-0414-4c6d-b90f-3feeadbd0190"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.164218 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data" (OuterVolumeSpecName: "config-data") pod "ba8f40d8-0414-4c6d-b90f-3feeadbd0190" (UID: "ba8f40d8-0414-4c6d-b90f-3feeadbd0190"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.172208 4978 generic.go:334] "Generic (PLEG): container finished" podID="ba8f40d8-0414-4c6d-b90f-3feeadbd0190" containerID="1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def" exitCode=0 Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.172244 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" event={"ID":"ba8f40d8-0414-4c6d-b90f-3feeadbd0190","Type":"ContainerDied","Data":"1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def"} Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.172269 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" event={"ID":"ba8f40d8-0414-4c6d-b90f-3feeadbd0190","Type":"ContainerDied","Data":"37c9f01f13210650b97d28858edb4e410d71c50cf3b3b05935d0f8f9fd549593"} Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.172289 4978 scope.go:117] "RemoveContainer" containerID="1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.172356 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6784f568f4-l8zd6" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.185862 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfcc9\" (UniqueName: \"kubernetes.io/projected/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-kube-api-access-pfcc9\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.185891 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.185907 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.185919 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba8f40d8-0414-4c6d-b90f-3feeadbd0190-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.197642 4978 scope.go:117] "RemoveContainer" containerID="1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def" Feb 25 08:52:22 crc kubenswrapper[4978]: E0225 08:52:22.198174 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def\": container with ID starting with 1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def not found: ID does not exist" containerID="1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.198227 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def"} err="failed to get container status \"1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def\": rpc error: code = NotFound desc = could not find container \"1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def\": container with ID starting with 1e6ec72c6eea6957b555f10b5781fbee987482cdaa801b3411669fa8852f7def not found: ID does not exist" Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.238530 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6784f568f4-l8zd6"] Feb 25 08:52:22 crc kubenswrapper[4978]: I0225 08:52:22.252572 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-6784f568f4-l8zd6"] Feb 25 08:52:23 crc kubenswrapper[4978]: I0225 08:52:23.349974 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba8f40d8-0414-4c6d-b90f-3feeadbd0190" path="/var/lib/kubelet/pods/ba8f40d8-0414-4c6d-b90f-3feeadbd0190/volumes" Feb 25 08:52:25 crc kubenswrapper[4978]: I0225 08:52:25.349886 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7d87784d56-fghgj" Feb 25 08:52:25 crc kubenswrapper[4978]: I0225 08:52:25.410885 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6c9d95c995-rfhkv"] Feb 25 08:52:25 crc kubenswrapper[4978]: I0225 08:52:25.411531 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-6c9d95c995-rfhkv" podUID="1751340a-9d92-4a12-ba1e-297064818a4e" containerName="heat-engine" containerID="cri-o://a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" gracePeriod=60 Feb 25 08:52:25 crc kubenswrapper[4978]: E0225 08:52:25.414512 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 25 08:52:25 crc kubenswrapper[4978]: E0225 08:52:25.416264 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 25 08:52:25 crc kubenswrapper[4978]: E0225 08:52:25.418154 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 25 08:52:25 crc kubenswrapper[4978]: E0225 08:52:25.418222 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6c9d95c995-rfhkv" podUID="1751340a-9d92-4a12-ba1e-297064818a4e" containerName="heat-engine" Feb 25 08:52:27 crc kubenswrapper[4978]: I0225 08:52:27.722886 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5b459dcfff-rkn7k" Feb 25 08:52:27 crc kubenswrapper[4978]: I0225 08:52:27.778088 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-57f8f4d4bc-ltxm5"] Feb 25 08:52:27 crc kubenswrapper[4978]: I0225 08:52:27.785700 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-7fc7d9db6-jkz4h" Feb 25 08:52:27 crc kubenswrapper[4978]: I0225 08:52:27.882927 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-894b45677-8r2pm"] Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.252356 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" event={"ID":"207dc629-cc88-460b-8159-34b8fca288cf","Type":"ContainerDied","Data":"9b4ffc64005a377358236ef89fc3149e6664b0c823bfa3c0b79b97a40accfdf8"} Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.252684 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b4ffc64005a377358236ef89fc3149e6664b0c823bfa3c0b79b97a40accfdf8" Feb 25 08:52:28 crc kubenswrapper[4978]: E0225 08:52:28.312904 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 25 08:52:28 crc kubenswrapper[4978]: E0225 08:52:28.314935 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 25 08:52:28 crc kubenswrapper[4978]: E0225 08:52:28.316413 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 25 08:52:28 crc kubenswrapper[4978]: E0225 08:52:28.316482 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6c9d95c995-rfhkv" podUID="1751340a-9d92-4a12-ba1e-297064818a4e" containerName="heat-engine" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.328256 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.338113 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.424709 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-776qn\" (UniqueName: \"kubernetes.io/projected/207dc629-cc88-460b-8159-34b8fca288cf-kube-api-access-776qn\") pod \"207dc629-cc88-460b-8159-34b8fca288cf\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.424879 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data\") pod \"207dc629-cc88-460b-8159-34b8fca288cf\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.424964 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-combined-ca-bundle\") pod \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.424996 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data-custom\") pod \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.425096 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data-custom\") pod \"207dc629-cc88-460b-8159-34b8fca288cf\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.425130 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data\") pod \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.425165 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggkr7\" (UniqueName: \"kubernetes.io/projected/fd0033bc-91e4-498b-9c02-0e3d67123c4a-kube-api-access-ggkr7\") pod \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\" (UID: \"fd0033bc-91e4-498b-9c02-0e3d67123c4a\") " Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.425197 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-combined-ca-bundle\") pod \"207dc629-cc88-460b-8159-34b8fca288cf\" (UID: \"207dc629-cc88-460b-8159-34b8fca288cf\") " Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.431324 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/207dc629-cc88-460b-8159-34b8fca288cf-kube-api-access-776qn" (OuterVolumeSpecName: "kube-api-access-776qn") pod "207dc629-cc88-460b-8159-34b8fca288cf" (UID: "207dc629-cc88-460b-8159-34b8fca288cf"). InnerVolumeSpecName "kube-api-access-776qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.431993 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "207dc629-cc88-460b-8159-34b8fca288cf" (UID: "207dc629-cc88-460b-8159-34b8fca288cf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.432199 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fd0033bc-91e4-498b-9c02-0e3d67123c4a" (UID: "fd0033bc-91e4-498b-9c02-0e3d67123c4a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.442669 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd0033bc-91e4-498b-9c02-0e3d67123c4a-kube-api-access-ggkr7" (OuterVolumeSpecName: "kube-api-access-ggkr7") pod "fd0033bc-91e4-498b-9c02-0e3d67123c4a" (UID: "fd0033bc-91e4-498b-9c02-0e3d67123c4a"). InnerVolumeSpecName "kube-api-access-ggkr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.472499 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd0033bc-91e4-498b-9c02-0e3d67123c4a" (UID: "fd0033bc-91e4-498b-9c02-0e3d67123c4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.493164 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "207dc629-cc88-460b-8159-34b8fca288cf" (UID: "207dc629-cc88-460b-8159-34b8fca288cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.503803 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data" (OuterVolumeSpecName: "config-data") pod "fd0033bc-91e4-498b-9c02-0e3d67123c4a" (UID: "fd0033bc-91e4-498b-9c02-0e3d67123c4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.519153 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data" (OuterVolumeSpecName: "config-data") pod "207dc629-cc88-460b-8159-34b8fca288cf" (UID: "207dc629-cc88-460b-8159-34b8fca288cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.528246 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.528286 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.528302 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.528315 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggkr7\" (UniqueName: \"kubernetes.io/projected/fd0033bc-91e4-498b-9c02-0e3d67123c4a-kube-api-access-ggkr7\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.528331 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.528342 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-776qn\" (UniqueName: \"kubernetes.io/projected/207dc629-cc88-460b-8159-34b8fca288cf-kube-api-access-776qn\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.528354 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207dc629-cc88-460b-8159-34b8fca288cf-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:28 crc kubenswrapper[4978]: I0225 08:52:28.528368 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0033bc-91e4-498b-9c02-0e3d67123c4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.265882 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-57f8f4d4bc-ltxm5" Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.268700 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-894b45677-8r2pm" Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.268710 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-894b45677-8r2pm" event={"ID":"fd0033bc-91e4-498b-9c02-0e3d67123c4a","Type":"ContainerDied","Data":"94a3f577ca5d1dfe9df581bf724b08a752c8a47edef4e46302242a5f23c1cddf"} Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.268790 4978 scope.go:117] "RemoveContainer" containerID="52f26e1fbbcc8b4f76ab4368557e659aa36a878d983493e9ea230628e569c530" Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.359107 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-894b45677-8r2pm"] Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.368451 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-894b45677-8r2pm"] Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.376356 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-57f8f4d4bc-ltxm5"] Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.384445 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-57f8f4d4bc-ltxm5"] Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.689153 4978 scope.go:117] "RemoveContainer" containerID="ac931080148ca4131204bdaad50dd5c4b10a2fc070727d947b5bf249a6baefce" Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.752204 4978 scope.go:117] "RemoveContainer" containerID="bafeca1f67cde8e70de199fac502d5aa183a599e2c8d87dfe80da5a94b37a495" Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.815792 4978 scope.go:117] "RemoveContainer" containerID="f57e19bdd8f4cc851d980cd420e2ac9670d5ff1fbed8c1e499050f11d05a0cda" Feb 25 08:52:29 crc kubenswrapper[4978]: I0225 08:52:29.884343 4978 scope.go:117] "RemoveContainer" containerID="f2aafde9a69057765fe9dbf66c856cc1f4937c9d15ca0dadbddb7ea6385f0b22" Feb 25 08:52:31 crc kubenswrapper[4978]: I0225 08:52:31.349364 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="207dc629-cc88-460b-8159-34b8fca288cf" path="/var/lib/kubelet/pods/207dc629-cc88-460b-8159-34b8fca288cf/volumes" Feb 25 08:52:31 crc kubenswrapper[4978]: I0225 08:52:31.350934 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" path="/var/lib/kubelet/pods/fd0033bc-91e4-498b-9c02-0e3d67123c4a/volumes" Feb 25 08:52:31 crc kubenswrapper[4978]: I0225 08:52:31.965567 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7cf5bdb968-klhbw" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.168:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.168:8443: connect: connection refused" Feb 25 08:52:31 crc kubenswrapper[4978]: I0225 08:52:31.965687 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.143742 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.313126 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data\") pod \"1751340a-9d92-4a12-ba1e-297064818a4e\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.313198 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgdbf\" (UniqueName: \"kubernetes.io/projected/1751340a-9d92-4a12-ba1e-297064818a4e-kube-api-access-mgdbf\") pod \"1751340a-9d92-4a12-ba1e-297064818a4e\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.313245 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-combined-ca-bundle\") pod \"1751340a-9d92-4a12-ba1e-297064818a4e\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.313660 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data-custom\") pod \"1751340a-9d92-4a12-ba1e-297064818a4e\" (UID: \"1751340a-9d92-4a12-ba1e-297064818a4e\") " Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.318883 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1751340a-9d92-4a12-ba1e-297064818a4e-kube-api-access-mgdbf" (OuterVolumeSpecName: "kube-api-access-mgdbf") pod "1751340a-9d92-4a12-ba1e-297064818a4e" (UID: "1751340a-9d92-4a12-ba1e-297064818a4e"). InnerVolumeSpecName "kube-api-access-mgdbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.319305 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1751340a-9d92-4a12-ba1e-297064818a4e" (UID: "1751340a-9d92-4a12-ba1e-297064818a4e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.350691 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1751340a-9d92-4a12-ba1e-297064818a4e" (UID: "1751340a-9d92-4a12-ba1e-297064818a4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.352352 4978 generic.go:334] "Generic (PLEG): container finished" podID="1751340a-9d92-4a12-ba1e-297064818a4e" containerID="a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" exitCode=0 Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.352404 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6c9d95c995-rfhkv" event={"ID":"1751340a-9d92-4a12-ba1e-297064818a4e","Type":"ContainerDied","Data":"a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e"} Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.352435 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6c9d95c995-rfhkv" event={"ID":"1751340a-9d92-4a12-ba1e-297064818a4e","Type":"ContainerDied","Data":"57ab7268ade78f37c5e203a1d99b390be36de62c77db0730ca452e25da9139aa"} Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.352453 4978 scope.go:117] "RemoveContainer" containerID="a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.352488 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6c9d95c995-rfhkv" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.382955 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data" (OuterVolumeSpecName: "config-data") pod "1751340a-9d92-4a12-ba1e-297064818a4e" (UID: "1751340a-9d92-4a12-ba1e-297064818a4e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.416470 4978 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.416513 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.416530 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgdbf\" (UniqueName: \"kubernetes.io/projected/1751340a-9d92-4a12-ba1e-297064818a4e-kube-api-access-mgdbf\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.416541 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1751340a-9d92-4a12-ba1e-297064818a4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.461564 4978 scope.go:117] "RemoveContainer" containerID="a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" Feb 25 08:52:36 crc kubenswrapper[4978]: E0225 08:52:36.462230 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e\": container with ID starting with a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e not found: ID does not exist" containerID="a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.462283 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e"} err="failed to get container status \"a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e\": rpc error: code = NotFound desc = could not find container \"a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e\": container with ID starting with a78ba9e2ea4017d7392107433ed330ac60efad251d62ea3f8a425c625571273e not found: ID does not exist" Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.720173 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6c9d95c995-rfhkv"] Feb 25 08:52:36 crc kubenswrapper[4978]: I0225 08:52:36.731961 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-6c9d95c995-rfhkv"] Feb 25 08:52:37 crc kubenswrapper[4978]: I0225 08:52:37.351842 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1751340a-9d92-4a12-ba1e-297064818a4e" path="/var/lib/kubelet/pods/1751340a-9d92-4a12-ba1e-297064818a4e/volumes" Feb 25 08:52:40 crc kubenswrapper[4978]: I0225 08:52:40.905154 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.037039 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-secret-key\") pod \"01581086-b21b-4f8a-8390-df51be3a5999\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.037166 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-combined-ca-bundle\") pod \"01581086-b21b-4f8a-8390-df51be3a5999\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.037204 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-config-data\") pod \"01581086-b21b-4f8a-8390-df51be3a5999\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.037262 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-scripts\") pod \"01581086-b21b-4f8a-8390-df51be3a5999\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.037430 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01581086-b21b-4f8a-8390-df51be3a5999-logs\") pod \"01581086-b21b-4f8a-8390-df51be3a5999\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.037550 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpnxb\" (UniqueName: \"kubernetes.io/projected/01581086-b21b-4f8a-8390-df51be3a5999-kube-api-access-tpnxb\") pod \"01581086-b21b-4f8a-8390-df51be3a5999\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.037774 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-tls-certs\") pod \"01581086-b21b-4f8a-8390-df51be3a5999\" (UID: \"01581086-b21b-4f8a-8390-df51be3a5999\") " Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.038085 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01581086-b21b-4f8a-8390-df51be3a5999-logs" (OuterVolumeSpecName: "logs") pod "01581086-b21b-4f8a-8390-df51be3a5999" (UID: "01581086-b21b-4f8a-8390-df51be3a5999"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.038848 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01581086-b21b-4f8a-8390-df51be3a5999-logs\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.043334 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01581086-b21b-4f8a-8390-df51be3a5999-kube-api-access-tpnxb" (OuterVolumeSpecName: "kube-api-access-tpnxb") pod "01581086-b21b-4f8a-8390-df51be3a5999" (UID: "01581086-b21b-4f8a-8390-df51be3a5999"). InnerVolumeSpecName "kube-api-access-tpnxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.043801 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "01581086-b21b-4f8a-8390-df51be3a5999" (UID: "01581086-b21b-4f8a-8390-df51be3a5999"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.065175 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-scripts" (OuterVolumeSpecName: "scripts") pod "01581086-b21b-4f8a-8390-df51be3a5999" (UID: "01581086-b21b-4f8a-8390-df51be3a5999"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.075817 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01581086-b21b-4f8a-8390-df51be3a5999" (UID: "01581086-b21b-4f8a-8390-df51be3a5999"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.081134 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-config-data" (OuterVolumeSpecName: "config-data") pod "01581086-b21b-4f8a-8390-df51be3a5999" (UID: "01581086-b21b-4f8a-8390-df51be3a5999"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.114565 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "01581086-b21b-4f8a-8390-df51be3a5999" (UID: "01581086-b21b-4f8a-8390-df51be3a5999"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.140750 4978 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.140816 4978 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.140841 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01581086-b21b-4f8a-8390-df51be3a5999-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.140867 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.140890 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01581086-b21b-4f8a-8390-df51be3a5999-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.140918 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpnxb\" (UniqueName: \"kubernetes.io/projected/01581086-b21b-4f8a-8390-df51be3a5999-kube-api-access-tpnxb\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.415277 4978 generic.go:334] "Generic (PLEG): container finished" podID="01581086-b21b-4f8a-8390-df51be3a5999" containerID="191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b" exitCode=137 Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.415427 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cf5bdb968-klhbw" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.415409 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cf5bdb968-klhbw" event={"ID":"01581086-b21b-4f8a-8390-df51be3a5999","Type":"ContainerDied","Data":"191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b"} Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.415816 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cf5bdb968-klhbw" event={"ID":"01581086-b21b-4f8a-8390-df51be3a5999","Type":"ContainerDied","Data":"1db00d1aa44876704b44d201dbf9acd246a0d4795ecb68201c60d654a3883f53"} Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.415880 4978 scope.go:117] "RemoveContainer" containerID="1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.468111 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cf5bdb968-klhbw"] Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.482908 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7cf5bdb968-klhbw"] Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.597941 4978 scope.go:117] "RemoveContainer" containerID="191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.621421 4978 scope.go:117] "RemoveContainer" containerID="1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa" Feb 25 08:52:41 crc kubenswrapper[4978]: E0225 08:52:41.621845 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa\": container with ID starting with 1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa not found: ID does not exist" containerID="1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.621876 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa"} err="failed to get container status \"1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa\": rpc error: code = NotFound desc = could not find container \"1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa\": container with ID starting with 1042a1f4f007b5b4efb5ea928c759413fc88bf1765d8d530bb264756fb4ffaaa not found: ID does not exist" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.621899 4978 scope.go:117] "RemoveContainer" containerID="191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b" Feb 25 08:52:41 crc kubenswrapper[4978]: E0225 08:52:41.622119 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b\": container with ID starting with 191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b not found: ID does not exist" containerID="191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b" Feb 25 08:52:41 crc kubenswrapper[4978]: I0225 08:52:41.622140 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b"} err="failed to get container status \"191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b\": rpc error: code = NotFound desc = could not find container \"191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b\": container with ID starting with 191734d832ec940e4733dced55b4700cd27877edc11a4d28589f0e098e70d33b not found: ID does not exist" Feb 25 08:52:43 crc kubenswrapper[4978]: I0225 08:52:43.346483 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01581086-b21b-4f8a-8390-df51be3a5999" path="/var/lib/kubelet/pods/01581086-b21b-4f8a-8390-df51be3a5999/volumes" Feb 25 08:52:48 crc kubenswrapper[4978]: I0225 08:52:48.069685 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-75j64"] Feb 25 08:52:48 crc kubenswrapper[4978]: I0225 08:52:48.082239 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-75j64"] Feb 25 08:52:49 crc kubenswrapper[4978]: I0225 08:52:49.033797 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-51be-account-create-update-xvkcz"] Feb 25 08:52:49 crc kubenswrapper[4978]: I0225 08:52:49.043634 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-51be-account-create-update-xvkcz"] Feb 25 08:52:49 crc kubenswrapper[4978]: I0225 08:52:49.347982 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5" path="/var/lib/kubelet/pods/15956d3a-88d0-4c2f-bfc2-93dd5e79c1c5/volumes" Feb 25 08:52:49 crc kubenswrapper[4978]: I0225 08:52:49.349800 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f" path="/var/lib/kubelet/pods/6fbb51bf-dcc5-47b6-85fd-85b6ed46da1f/volumes" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.920671 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n"] Feb 25 08:52:52 crc kubenswrapper[4978]: E0225 08:52:52.921517 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon-log" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.921540 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon-log" Feb 25 08:52:52 crc kubenswrapper[4978]: E0225 08:52:52.921571 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" containerName="heat-api" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.921582 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" containerName="heat-api" Feb 25 08:52:52 crc kubenswrapper[4978]: E0225 08:52:52.921617 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" containerName="heat-api" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.921628 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" containerName="heat-api" Feb 25 08:52:52 crc kubenswrapper[4978]: E0225 08:52:52.921643 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba8f40d8-0414-4c6d-b90f-3feeadbd0190" containerName="heat-cfnapi" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.921655 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba8f40d8-0414-4c6d-b90f-3feeadbd0190" containerName="heat-cfnapi" Feb 25 08:52:52 crc kubenswrapper[4978]: E0225 08:52:52.921672 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207dc629-cc88-460b-8159-34b8fca288cf" containerName="heat-cfnapi" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.921682 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="207dc629-cc88-460b-8159-34b8fca288cf" containerName="heat-cfnapi" Feb 25 08:52:52 crc kubenswrapper[4978]: E0225 08:52:52.921700 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="207dc629-cc88-460b-8159-34b8fca288cf" containerName="heat-cfnapi" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.921710 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="207dc629-cc88-460b-8159-34b8fca288cf" containerName="heat-cfnapi" Feb 25 08:52:52 crc kubenswrapper[4978]: E0225 08:52:52.921731 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1751340a-9d92-4a12-ba1e-297064818a4e" containerName="heat-engine" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.921740 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1751340a-9d92-4a12-ba1e-297064818a4e" containerName="heat-engine" Feb 25 08:52:52 crc kubenswrapper[4978]: E0225 08:52:52.921765 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.921775 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.922045 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad9b26c4-0fc4-4c3b-a9ea-41b01f4568ec" containerName="heat-api" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.922073 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.922086 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" containerName="heat-api" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.922104 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="207dc629-cc88-460b-8159-34b8fca288cf" containerName="heat-cfnapi" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.922116 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="01581086-b21b-4f8a-8390-df51be3a5999" containerName="horizon-log" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.922130 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" containerName="heat-api" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.922152 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba8f40d8-0414-4c6d-b90f-3feeadbd0190" containerName="heat-cfnapi" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.922177 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1751340a-9d92-4a12-ba1e-297064818a4e" containerName="heat-engine" Feb 25 08:52:52 crc kubenswrapper[4978]: E0225 08:52:52.922473 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" containerName="heat-api" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.922486 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd0033bc-91e4-498b-9c02-0e3d67123c4a" containerName="heat-api" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.922709 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="207dc629-cc88-460b-8159-34b8fca288cf" containerName="heat-cfnapi" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.923946 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.926624 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 25 08:52:52 crc kubenswrapper[4978]: I0225 08:52:52.932650 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n"] Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.018562 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2nlm\" (UniqueName: \"kubernetes.io/projected/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-kube-api-access-z2nlm\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.018702 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.018771 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.120297 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.120402 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.120489 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2nlm\" (UniqueName: \"kubernetes.io/projected/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-kube-api-access-z2nlm\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.120995 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.121280 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.155122 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2nlm\" (UniqueName: \"kubernetes.io/projected/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-kube-api-access-z2nlm\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.256868 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:53 crc kubenswrapper[4978]: I0225 08:52:53.799386 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n"] Feb 25 08:52:54 crc kubenswrapper[4978]: I0225 08:52:54.574254 4978 generic.go:334] "Generic (PLEG): container finished" podID="9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" containerID="97566171fc81869009db071f28cd2208953d5da3a7ca3a03ba8b0076a7bde4eb" exitCode=0 Feb 25 08:52:54 crc kubenswrapper[4978]: I0225 08:52:54.574450 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" event={"ID":"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035","Type":"ContainerDied","Data":"97566171fc81869009db071f28cd2208953d5da3a7ca3a03ba8b0076a7bde4eb"} Feb 25 08:52:54 crc kubenswrapper[4978]: I0225 08:52:54.575344 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" event={"ID":"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035","Type":"ContainerStarted","Data":"b2ee1846d8f5b48400306bba9ae206b89e56be8689bb3a271ce78e701396c74e"} Feb 25 08:52:56 crc kubenswrapper[4978]: I0225 08:52:56.604625 4978 generic.go:334] "Generic (PLEG): container finished" podID="9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" containerID="1a9720761f4b8b87476e8bbe84b50916c8b305c218f3272b79ec5e7cf424c5d9" exitCode=0 Feb 25 08:52:56 crc kubenswrapper[4978]: I0225 08:52:56.604741 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" event={"ID":"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035","Type":"ContainerDied","Data":"1a9720761f4b8b87476e8bbe84b50916c8b305c218f3272b79ec5e7cf424c5d9"} Feb 25 08:52:57 crc kubenswrapper[4978]: I0225 08:52:57.034766 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-f7v86"] Feb 25 08:52:57 crc kubenswrapper[4978]: I0225 08:52:57.043685 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-f7v86"] Feb 25 08:52:57 crc kubenswrapper[4978]: I0225 08:52:57.341514 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d02cc3c-4121-46d0-ae30-7842907c283f" path="/var/lib/kubelet/pods/5d02cc3c-4121-46d0-ae30-7842907c283f/volumes" Feb 25 08:52:57 crc kubenswrapper[4978]: I0225 08:52:57.623699 4978 generic.go:334] "Generic (PLEG): container finished" podID="9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" containerID="f63af1bb6eda105d54596beec1fa77417af4fcb3d4c90665fe7a38469dfb880b" exitCode=0 Feb 25 08:52:57 crc kubenswrapper[4978]: I0225 08:52:57.623758 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" event={"ID":"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035","Type":"ContainerDied","Data":"f63af1bb6eda105d54596beec1fa77417af4fcb3d4c90665fe7a38469dfb880b"} Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.119448 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.257140 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-bundle\") pod \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.257525 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-util\") pod \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.257584 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2nlm\" (UniqueName: \"kubernetes.io/projected/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-kube-api-access-z2nlm\") pod \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\" (UID: \"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035\") " Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.260295 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-bundle" (OuterVolumeSpecName: "bundle") pod "9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" (UID: "9cdf7cf4-2d4a-481c-9d02-0f32e00f6035"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.265526 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-kube-api-access-z2nlm" (OuterVolumeSpecName: "kube-api-access-z2nlm") pod "9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" (UID: "9cdf7cf4-2d4a-481c-9d02-0f32e00f6035"). InnerVolumeSpecName "kube-api-access-z2nlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.274087 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-util" (OuterVolumeSpecName: "util") pod "9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" (UID: "9cdf7cf4-2d4a-481c-9d02-0f32e00f6035"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.359821 4978 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-util\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.359859 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2nlm\" (UniqueName: \"kubernetes.io/projected/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-kube-api-access-z2nlm\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.359873 4978 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9cdf7cf4-2d4a-481c-9d02-0f32e00f6035-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.652841 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" event={"ID":"9cdf7cf4-2d4a-481c-9d02-0f32e00f6035","Type":"ContainerDied","Data":"b2ee1846d8f5b48400306bba9ae206b89e56be8689bb3a271ce78e701396c74e"} Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.652880 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2ee1846d8f5b48400306bba9ae206b89e56be8689bb3a271ce78e701396c74e" Feb 25 08:52:59 crc kubenswrapper[4978]: I0225 08:52:59.652979 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.709734 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts"] Feb 25 08:53:10 crc kubenswrapper[4978]: E0225 08:53:10.710775 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" containerName="extract" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.710791 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" containerName="extract" Feb 25 08:53:10 crc kubenswrapper[4978]: E0225 08:53:10.710809 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" containerName="util" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.710816 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" containerName="util" Feb 25 08:53:10 crc kubenswrapper[4978]: E0225 08:53:10.710829 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" containerName="pull" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.710836 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" containerName="pull" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.711054 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cdf7cf4-2d4a-481c-9d02-0f32e00f6035" containerName="extract" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.711839 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.714165 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-7tlmt" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.714235 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.714449 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.721948 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts"] Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.734668 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw"] Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.735852 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.737812 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-nksk7" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.738350 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.757391 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg"] Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.758616 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.781482 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg"] Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.797462 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bb1b2828-81ee-4bc1-afdb-0306ae16d896-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw\" (UID: \"bb1b2828-81ee-4bc1-afdb-0306ae16d896\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.797564 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg\" (UID: \"dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.797619 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkfb8\" (UniqueName: \"kubernetes.io/projected/7701e422-276b-4919-ae42-2901a90faa26-kube-api-access-gkfb8\") pod \"obo-prometheus-operator-68bc856cb9-vw9ts\" (UID: \"7701e422-276b-4919-ae42-2901a90faa26\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.797668 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bb1b2828-81ee-4bc1-afdb-0306ae16d896-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw\" (UID: \"bb1b2828-81ee-4bc1-afdb-0306ae16d896\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.797689 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg\" (UID: \"dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.818112 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw"] Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.898475 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg\" (UID: \"dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.898562 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkfb8\" (UniqueName: \"kubernetes.io/projected/7701e422-276b-4919-ae42-2901a90faa26-kube-api-access-gkfb8\") pod \"obo-prometheus-operator-68bc856cb9-vw9ts\" (UID: \"7701e422-276b-4919-ae42-2901a90faa26\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.898621 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bb1b2828-81ee-4bc1-afdb-0306ae16d896-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw\" (UID: \"bb1b2828-81ee-4bc1-afdb-0306ae16d896\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.898640 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg\" (UID: \"dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.898677 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bb1b2828-81ee-4bc1-afdb-0306ae16d896-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw\" (UID: \"bb1b2828-81ee-4bc1-afdb-0306ae16d896\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.907251 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bb1b2828-81ee-4bc1-afdb-0306ae16d896-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw\" (UID: \"bb1b2828-81ee-4bc1-afdb-0306ae16d896\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.908605 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg\" (UID: \"dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.915333 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg\" (UID: \"dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.916942 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bb1b2828-81ee-4bc1-afdb-0306ae16d896-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw\" (UID: \"bb1b2828-81ee-4bc1-afdb-0306ae16d896\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.920182 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkfb8\" (UniqueName: \"kubernetes.io/projected/7701e422-276b-4919-ae42-2901a90faa26-kube-api-access-gkfb8\") pod \"obo-prometheus-operator-68bc856cb9-vw9ts\" (UID: \"7701e422-276b-4919-ae42-2901a90faa26\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.937090 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-w8cvq"] Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.938510 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.942813 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-mbrw5" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.943117 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Feb 25 08:53:10 crc kubenswrapper[4978]: I0225 08:53:10.959568 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-w8cvq"] Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.008794 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a1db5e2-0139-4456-9e91-a3f31d7b38fb-observability-operator-tls\") pod \"observability-operator-59bdc8b94-w8cvq\" (UID: \"1a1db5e2-0139-4456-9e91-a3f31d7b38fb\") " pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.008913 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm67b\" (UniqueName: \"kubernetes.io/projected/1a1db5e2-0139-4456-9e91-a3f31d7b38fb-kube-api-access-gm67b\") pod \"observability-operator-59bdc8b94-w8cvq\" (UID: \"1a1db5e2-0139-4456-9e91-a3f31d7b38fb\") " pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.029882 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.055881 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.075986 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.113521 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a1db5e2-0139-4456-9e91-a3f31d7b38fb-observability-operator-tls\") pod \"observability-operator-59bdc8b94-w8cvq\" (UID: \"1a1db5e2-0139-4456-9e91-a3f31d7b38fb\") " pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.113648 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm67b\" (UniqueName: \"kubernetes.io/projected/1a1db5e2-0139-4456-9e91-a3f31d7b38fb-kube-api-access-gm67b\") pod \"observability-operator-59bdc8b94-w8cvq\" (UID: \"1a1db5e2-0139-4456-9e91-a3f31d7b38fb\") " pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.122106 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a1db5e2-0139-4456-9e91-a3f31d7b38fb-observability-operator-tls\") pod \"observability-operator-59bdc8b94-w8cvq\" (UID: \"1a1db5e2-0139-4456-9e91-a3f31d7b38fb\") " pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.122434 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-4vd82"] Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.123893 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-4vd82" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.127537 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-7pc47" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.137161 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-4vd82"] Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.151052 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm67b\" (UniqueName: \"kubernetes.io/projected/1a1db5e2-0139-4456-9e91-a3f31d7b38fb-kube-api-access-gm67b\") pod \"observability-operator-59bdc8b94-w8cvq\" (UID: \"1a1db5e2-0139-4456-9e91-a3f31d7b38fb\") " pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.214874 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcz7b\" (UniqueName: \"kubernetes.io/projected/79e63f5b-296b-4145-b3b9-83f7ddd2d901-kube-api-access-rcz7b\") pod \"perses-operator-5bf474d74f-4vd82\" (UID: \"79e63f5b-296b-4145-b3b9-83f7ddd2d901\") " pod="openshift-operators/perses-operator-5bf474d74f-4vd82" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.214943 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/79e63f5b-296b-4145-b3b9-83f7ddd2d901-openshift-service-ca\") pod \"perses-operator-5bf474d74f-4vd82\" (UID: \"79e63f5b-296b-4145-b3b9-83f7ddd2d901\") " pod="openshift-operators/perses-operator-5bf474d74f-4vd82" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.319607 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcz7b\" (UniqueName: \"kubernetes.io/projected/79e63f5b-296b-4145-b3b9-83f7ddd2d901-kube-api-access-rcz7b\") pod \"perses-operator-5bf474d74f-4vd82\" (UID: \"79e63f5b-296b-4145-b3b9-83f7ddd2d901\") " pod="openshift-operators/perses-operator-5bf474d74f-4vd82" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.319717 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/79e63f5b-296b-4145-b3b9-83f7ddd2d901-openshift-service-ca\") pod \"perses-operator-5bf474d74f-4vd82\" (UID: \"79e63f5b-296b-4145-b3b9-83f7ddd2d901\") " pod="openshift-operators/perses-operator-5bf474d74f-4vd82" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.321095 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/79e63f5b-296b-4145-b3b9-83f7ddd2d901-openshift-service-ca\") pod \"perses-operator-5bf474d74f-4vd82\" (UID: \"79e63f5b-296b-4145-b3b9-83f7ddd2d901\") " pod="openshift-operators/perses-operator-5bf474d74f-4vd82" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.325263 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.366251 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcz7b\" (UniqueName: \"kubernetes.io/projected/79e63f5b-296b-4145-b3b9-83f7ddd2d901-kube-api-access-rcz7b\") pod \"perses-operator-5bf474d74f-4vd82\" (UID: \"79e63f5b-296b-4145-b3b9-83f7ddd2d901\") " pod="openshift-operators/perses-operator-5bf474d74f-4vd82" Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.554693 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-4vd82" Feb 25 08:53:11 crc kubenswrapper[4978]: W0225 08:53:11.867325 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7701e422_276b_4919_ae42_2901a90faa26.slice/crio-1eadc11ae240d9e63d7a23354c5c7428a0fbe7b3a23ef8db2383d6d905dedb09 WatchSource:0}: Error finding container 1eadc11ae240d9e63d7a23354c5c7428a0fbe7b3a23ef8db2383d6d905dedb09: Status 404 returned error can't find the container with id 1eadc11ae240d9e63d7a23354c5c7428a0fbe7b3a23ef8db2383d6d905dedb09 Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.871032 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts"] Feb 25 08:53:11 crc kubenswrapper[4978]: I0225 08:53:11.888436 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw"] Feb 25 08:53:12 crc kubenswrapper[4978]: I0225 08:53:12.059954 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg"] Feb 25 08:53:12 crc kubenswrapper[4978]: W0225 08:53:12.065247 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbf2dd91_faf4_4ef7_86f9_1d29b411dbe5.slice/crio-4613ef962aa93db33b98112a3360690960defd174c0003b2f94f8cda43e2debc WatchSource:0}: Error finding container 4613ef962aa93db33b98112a3360690960defd174c0003b2f94f8cda43e2debc: Status 404 returned error can't find the container with id 4613ef962aa93db33b98112a3360690960defd174c0003b2f94f8cda43e2debc Feb 25 08:53:12 crc kubenswrapper[4978]: I0225 08:53:12.163093 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-4vd82"] Feb 25 08:53:12 crc kubenswrapper[4978]: W0225 08:53:12.167483 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79e63f5b_296b_4145_b3b9_83f7ddd2d901.slice/crio-cba60f10b0edb3f7c5ead26e10e191c1af087017a7bacb9fb4dffa2a7ffc03a3 WatchSource:0}: Error finding container cba60f10b0edb3f7c5ead26e10e191c1af087017a7bacb9fb4dffa2a7ffc03a3: Status 404 returned error can't find the container with id cba60f10b0edb3f7c5ead26e10e191c1af087017a7bacb9fb4dffa2a7ffc03a3 Feb 25 08:53:12 crc kubenswrapper[4978]: W0225 08:53:12.167755 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a1db5e2_0139_4456_9e91_a3f31d7b38fb.slice/crio-0f092659ae3781c0c5bbf7ec4e3c850be00e2eb8dd1b20a7739d67039d462511 WatchSource:0}: Error finding container 0f092659ae3781c0c5bbf7ec4e3c850be00e2eb8dd1b20a7739d67039d462511: Status 404 returned error can't find the container with id 0f092659ae3781c0c5bbf7ec4e3c850be00e2eb8dd1b20a7739d67039d462511 Feb 25 08:53:12 crc kubenswrapper[4978]: I0225 08:53:12.172485 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-w8cvq"] Feb 25 08:53:12 crc kubenswrapper[4978]: I0225 08:53:12.785058 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" event={"ID":"1a1db5e2-0139-4456-9e91-a3f31d7b38fb","Type":"ContainerStarted","Data":"0f092659ae3781c0c5bbf7ec4e3c850be00e2eb8dd1b20a7739d67039d462511"} Feb 25 08:53:12 crc kubenswrapper[4978]: I0225 08:53:12.786629 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts" event={"ID":"7701e422-276b-4919-ae42-2901a90faa26","Type":"ContainerStarted","Data":"1eadc11ae240d9e63d7a23354c5c7428a0fbe7b3a23ef8db2383d6d905dedb09"} Feb 25 08:53:12 crc kubenswrapper[4978]: I0225 08:53:12.787357 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-4vd82" event={"ID":"79e63f5b-296b-4145-b3b9-83f7ddd2d901","Type":"ContainerStarted","Data":"cba60f10b0edb3f7c5ead26e10e191c1af087017a7bacb9fb4dffa2a7ffc03a3"} Feb 25 08:53:12 crc kubenswrapper[4978]: I0225 08:53:12.788413 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" event={"ID":"bb1b2828-81ee-4bc1-afdb-0306ae16d896","Type":"ContainerStarted","Data":"10b2ee9e05efd3c83ad5f38bf0407f453c313b01f3e8a611df2cd1cae31f0542"} Feb 25 08:53:12 crc kubenswrapper[4978]: I0225 08:53:12.794100 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" event={"ID":"dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5","Type":"ContainerStarted","Data":"4613ef962aa93db33b98112a3360690960defd174c0003b2f94f8cda43e2debc"} Feb 25 08:53:24 crc kubenswrapper[4978]: I0225 08:53:24.933244 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" event={"ID":"1a1db5e2-0139-4456-9e91-a3f31d7b38fb","Type":"ContainerStarted","Data":"732bb3f676138e34a5b59703cb946ac1c74a129a9368cb656cd068ba704e76d3"} Feb 25 08:53:24 crc kubenswrapper[4978]: I0225 08:53:24.933699 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" Feb 25 08:53:24 crc kubenswrapper[4978]: I0225 08:53:24.941868 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-4vd82" event={"ID":"79e63f5b-296b-4145-b3b9-83f7ddd2d901","Type":"ContainerStarted","Data":"93fc38a4b5655596e0dec15919690329d245cd492b41be71e825568a67ab27f4"} Feb 25 08:53:24 crc kubenswrapper[4978]: I0225 08:53:24.942393 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-4vd82" Feb 25 08:53:24 crc kubenswrapper[4978]: I0225 08:53:24.949933 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" event={"ID":"bb1b2828-81ee-4bc1-afdb-0306ae16d896","Type":"ContainerStarted","Data":"8968010aebbe49b3497b4d7c59eaf50bbafe8cc494a51bd9dca458935e383155"} Feb 25 08:53:24 crc kubenswrapper[4978]: I0225 08:53:24.952831 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" Feb 25 08:53:24 crc kubenswrapper[4978]: I0225 08:53:24.952927 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" event={"ID":"dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5","Type":"ContainerStarted","Data":"499e30d24186366889f9375d3c8b013c1bafe2bf64c6071e02e491135b1a3d33"} Feb 25 08:53:24 crc kubenswrapper[4978]: I0225 08:53:24.966093 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-w8cvq" podStartSLOduration=2.932612455 podStartE2EDuration="14.966077982s" podCreationTimestamp="2026-02-25 08:53:10 +0000 UTC" firstStartedPulling="2026-02-25 08:53:12.170067831 +0000 UTC m=+7685.609324280" lastFinishedPulling="2026-02-25 08:53:24.203533348 +0000 UTC m=+7697.642789807" observedRunningTime="2026-02-25 08:53:24.960539311 +0000 UTC m=+7698.399795770" watchObservedRunningTime="2026-02-25 08:53:24.966077982 +0000 UTC m=+7698.405334431" Feb 25 08:53:24 crc kubenswrapper[4978]: I0225 08:53:24.988559 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw" podStartSLOduration=2.653187598 podStartE2EDuration="14.988543598s" podCreationTimestamp="2026-02-25 08:53:10 +0000 UTC" firstStartedPulling="2026-02-25 08:53:11.891225982 +0000 UTC m=+7685.330482441" lastFinishedPulling="2026-02-25 08:53:24.226581982 +0000 UTC m=+7697.665838441" observedRunningTime="2026-02-25 08:53:24.985000268 +0000 UTC m=+7698.424256737" watchObservedRunningTime="2026-02-25 08:53:24.988543598 +0000 UTC m=+7698.427800047" Feb 25 08:53:25 crc kubenswrapper[4978]: I0225 08:53:25.049330 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-4vd82" podStartSLOduration=2.013698597 podStartE2EDuration="14.049309571s" podCreationTimestamp="2026-02-25 08:53:11 +0000 UTC" firstStartedPulling="2026-02-25 08:53:12.169741721 +0000 UTC m=+7685.608998180" lastFinishedPulling="2026-02-25 08:53:24.205352695 +0000 UTC m=+7697.644609154" observedRunningTime="2026-02-25 08:53:25.040336203 +0000 UTC m=+7698.479592662" watchObservedRunningTime="2026-02-25 08:53:25.049309571 +0000 UTC m=+7698.488566030" Feb 25 08:53:25 crc kubenswrapper[4978]: I0225 08:53:25.079817 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg" podStartSLOduration=2.946204546 podStartE2EDuration="15.079795845s" podCreationTimestamp="2026-02-25 08:53:10 +0000 UTC" firstStartedPulling="2026-02-25 08:53:12.069856406 +0000 UTC m=+7685.509112865" lastFinishedPulling="2026-02-25 08:53:24.203447715 +0000 UTC m=+7697.642704164" observedRunningTime="2026-02-25 08:53:25.074952315 +0000 UTC m=+7698.514208784" watchObservedRunningTime="2026-02-25 08:53:25.079795845 +0000 UTC m=+7698.519052304" Feb 25 08:53:25 crc kubenswrapper[4978]: I0225 08:53:25.961632 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts" event={"ID":"7701e422-276b-4919-ae42-2901a90faa26","Type":"ContainerStarted","Data":"9976731a0b1727947b37d5a9ac18f1d461ef96d84f1331adc2052b57cac3084a"} Feb 25 08:53:25 crc kubenswrapper[4978]: I0225 08:53:25.977315 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-vw9ts" podStartSLOduration=3.626375049 podStartE2EDuration="15.977297761s" podCreationTimestamp="2026-02-25 08:53:10 +0000 UTC" firstStartedPulling="2026-02-25 08:53:11.870537931 +0000 UTC m=+7685.309794390" lastFinishedPulling="2026-02-25 08:53:24.221460643 +0000 UTC m=+7697.660717102" observedRunningTime="2026-02-25 08:53:25.975886607 +0000 UTC m=+7699.415143066" watchObservedRunningTime="2026-02-25 08:53:25.977297761 +0000 UTC m=+7699.416554220" Feb 25 08:53:30 crc kubenswrapper[4978]: I0225 08:53:30.113103 4978 scope.go:117] "RemoveContainer" containerID="9d3d8fd493ab143cb7a9ab8336c2c8d387260cf743db5bfb345c38c00a8a4783" Feb 25 08:53:30 crc kubenswrapper[4978]: I0225 08:53:30.188587 4978 scope.go:117] "RemoveContainer" containerID="708d5bf1739b5b5ad5f990362a7480ef8b7459894fd84af9395a21209a64bc29" Feb 25 08:53:30 crc kubenswrapper[4978]: I0225 08:53:30.226443 4978 scope.go:117] "RemoveContainer" containerID="deb0808a6b3beb6afba2ae303af557f432a49b862af7aa8bc0c89290061078fd" Feb 25 08:53:31 crc kubenswrapper[4978]: I0225 08:53:31.558934 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-4vd82" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.622711 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.623224 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" containerName="openstackclient" containerID="cri-o://51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0" gracePeriod=2 Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.637462 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.674506 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 25 08:53:33 crc kubenswrapper[4978]: E0225 08:53:33.674877 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" containerName="openstackclient" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.674891 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" containerName="openstackclient" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.675057 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" containerName="openstackclient" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.675699 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.678482 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" podUID="8d351813-bd52-4ed0-8ee5-1356dcdc0158" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.686349 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.724728 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 25 08:53:33 crc kubenswrapper[4978]: E0225 08:53:33.725475 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-qz6fc openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[combined-ca-bundle kube-api-access-qz6fc openstack-config openstack-config-secret]: context canceled" pod="openstack/openstackclient" podUID="8d351813-bd52-4ed0-8ee5-1356dcdc0158" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.742860 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.774495 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.776054 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.782057 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.835436 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.836797 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.847004 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-xbvcm" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.847017 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.947726 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config-secret\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.947803 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.947829 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjwmm\" (UniqueName: \"kubernetes.io/projected/ffa89f03-a075-4645-864c-40f29ace9c09-kube-api-access-fjwmm\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.947849 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:33 crc kubenswrapper[4978]: I0225 08:53:33.947889 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtltc\" (UniqueName: \"kubernetes.io/projected/022f2aaa-0cfc-4441-9cdf-c8b36adad1ff-kube-api-access-gtltc\") pod \"kube-state-metrics-0\" (UID: \"022f2aaa-0cfc-4441-9cdf-c8b36adad1ff\") " pod="openstack/kube-state-metrics-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.043361 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.050584 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config-secret\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.051693 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.051751 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjwmm\" (UniqueName: \"kubernetes.io/projected/ffa89f03-a075-4645-864c-40f29ace9c09-kube-api-access-fjwmm\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.051792 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.051854 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtltc\" (UniqueName: \"kubernetes.io/projected/022f2aaa-0cfc-4441-9cdf-c8b36adad1ff-kube-api-access-gtltc\") pod \"kube-state-metrics-0\" (UID: \"022f2aaa-0cfc-4441-9cdf-c8b36adad1ff\") " pod="openstack/kube-state-metrics-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.052399 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8d351813-bd52-4ed0-8ee5-1356dcdc0158" podUID="ffa89f03-a075-4645-864c-40f29ace9c09" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.053353 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.057621 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.060466 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.067932 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config-secret\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.071778 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtltc\" (UniqueName: \"kubernetes.io/projected/022f2aaa-0cfc-4441-9cdf-c8b36adad1ff-kube-api-access-gtltc\") pod \"kube-state-metrics-0\" (UID: \"022f2aaa-0cfc-4441-9cdf-c8b36adad1ff\") " pod="openstack/kube-state-metrics-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.078216 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjwmm\" (UniqueName: \"kubernetes.io/projected/ffa89f03-a075-4645-864c-40f29ace9c09-kube-api-access-fjwmm\") pod \"openstackclient\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.097751 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.166054 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.461011 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.517021 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.520077 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.535520 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-5rdg9" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.535691 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.535770 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.536559 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.537613 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.670430 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/55251903-0563-4a00-8652-96350cc28320-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.671785 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/55251903-0563-4a00-8652-96350cc28320-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.671926 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/55251903-0563-4a00-8652-96350cc28320-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.672057 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwz9\" (UniqueName: \"kubernetes.io/projected/55251903-0563-4a00-8652-96350cc28320-kube-api-access-djwz9\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.672134 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/55251903-0563-4a00-8652-96350cc28320-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.672258 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/55251903-0563-4a00-8652-96350cc28320-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.672477 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/55251903-0563-4a00-8652-96350cc28320-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.773754 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/55251903-0563-4a00-8652-96350cc28320-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.773857 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/55251903-0563-4a00-8652-96350cc28320-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.773887 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/55251903-0563-4a00-8652-96350cc28320-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.773926 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/55251903-0563-4a00-8652-96350cc28320-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.773963 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djwz9\" (UniqueName: \"kubernetes.io/projected/55251903-0563-4a00-8652-96350cc28320-kube-api-access-djwz9\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.773979 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/55251903-0563-4a00-8652-96350cc28320-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.773996 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/55251903-0563-4a00-8652-96350cc28320-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.780189 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/55251903-0563-4a00-8652-96350cc28320-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.783843 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/55251903-0563-4a00-8652-96350cc28320-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.785364 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/55251903-0563-4a00-8652-96350cc28320-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.787636 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.791942 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/55251903-0563-4a00-8652-96350cc28320-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.794686 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/55251903-0563-4a00-8652-96350cc28320-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.806566 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/55251903-0563-4a00-8652-96350cc28320-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.809094 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djwz9\" (UniqueName: \"kubernetes.io/projected/55251903-0563-4a00-8652-96350cc28320-kube-api-access-djwz9\") pod \"alertmanager-metric-storage-0\" (UID: \"55251903-0563-4a00-8652-96350cc28320\") " pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.853032 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Feb 25 08:53:34 crc kubenswrapper[4978]: I0225 08:53:34.862883 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.108185 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ffa89f03-a075-4645-864c-40f29ace9c09","Type":"ContainerStarted","Data":"ec498bc935615785f33049071a626574fe08cd9a0951e1dcf4c7d216e2482234"} Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.133611 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"022f2aaa-0cfc-4441-9cdf-c8b36adad1ff","Type":"ContainerStarted","Data":"f573bfea8a15fa5b587f87c7a9d52d0604220b550517adabba7c1bb00cd1ab47"} Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.133658 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.137054 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8d351813-bd52-4ed0-8ee5-1356dcdc0158" podUID="ffa89f03-a075-4645-864c-40f29ace9c09" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.149951 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.152358 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.162325 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.165189 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.165517 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.165572 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.165765 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.165917 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.166024 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-hp7ht" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.166036 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.168579 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.286062 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.286173 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.286202 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lxw5\" (UniqueName: \"kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-kube-api-access-5lxw5\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.286222 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-config\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.286251 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.286286 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f4941e0b-0e69-4577-837f-6633db74a609-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.286302 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.286316 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.286347 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.286389 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-02862e46-9f8b-46b3-930d-0200c6093953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.363580 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d351813-bd52-4ed0-8ee5-1356dcdc0158" path="/var/lib/kubelet/pods/8d351813-bd52-4ed0-8ee5-1356dcdc0158/volumes" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.407364 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-config\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.407443 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.407485 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f4941e0b-0e69-4577-837f-6633db74a609-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.407509 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.407528 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.407560 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.407590 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-02862e46-9f8b-46b3-930d-0200c6093953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.407620 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.407699 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.407734 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lxw5\" (UniqueName: \"kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-kube-api-access-5lxw5\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.411155 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.411886 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.412273 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.415523 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.419304 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.420731 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f4941e0b-0e69-4577-837f-6633db74a609-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.425967 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-config\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.426386 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.435764 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.435806 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-02862e46-9f8b-46b3-930d-0200c6093953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4d0b98a232139d7495440a8b884ebd34044c8aae36a3e9220f1b8cefdfc7f1e5/globalmount\"" pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.442066 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lxw5\" (UniqueName: \"kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-kube-api-access-5lxw5\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.493157 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.583632 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-02862e46-9f8b-46b3-930d-0200c6093953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") pod \"prometheus-metric-storage-0\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.807436 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 25 08:53:35 crc kubenswrapper[4978]: I0225 08:53:35.982195 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.129058 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config\") pod \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.129163 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hf22\" (UniqueName: \"kubernetes.io/projected/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-kube-api-access-7hf22\") pod \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.129232 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config-secret\") pod \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.129414 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-combined-ca-bundle\") pod \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\" (UID: \"71c9961c-3b9f-4d2a-b392-abd5eb920a3b\") " Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.148398 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-kube-api-access-7hf22" (OuterVolumeSpecName: "kube-api-access-7hf22") pod "71c9961c-3b9f-4d2a-b392-abd5eb920a3b" (UID: "71c9961c-3b9f-4d2a-b392-abd5eb920a3b"). InnerVolumeSpecName "kube-api-access-7hf22". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.172554 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"022f2aaa-0cfc-4441-9cdf-c8b36adad1ff","Type":"ContainerStarted","Data":"46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9"} Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.172903 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "71c9961c-3b9f-4d2a-b392-abd5eb920a3b" (UID: "71c9961c-3b9f-4d2a-b392-abd5eb920a3b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.172973 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.203870 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"55251903-0563-4a00-8652-96350cc28320","Type":"ContainerStarted","Data":"d78063cf9e121cc189d94338093b0ab5cda3673af4bc2963f6aa2bcebf8ef0b9"} Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.207698 4978 generic.go:334] "Generic (PLEG): container finished" podID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" containerID="51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0" exitCode=137 Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.207767 4978 scope.go:117] "RemoveContainer" containerID="51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.207894 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.208873 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.736096887 podStartE2EDuration="3.208857274s" podCreationTimestamp="2026-02-25 08:53:33 +0000 UTC" firstStartedPulling="2026-02-25 08:53:34.875409902 +0000 UTC m=+7708.314666351" lastFinishedPulling="2026-02-25 08:53:35.348170279 +0000 UTC m=+7708.787426738" observedRunningTime="2026-02-25 08:53:36.203693594 +0000 UTC m=+7709.642950063" watchObservedRunningTime="2026-02-25 08:53:36.208857274 +0000 UTC m=+7709.648113733" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.235659 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.235685 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hf22\" (UniqueName: \"kubernetes.io/projected/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-kube-api-access-7hf22\") on node \"crc\" DevicePath \"\"" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.237343 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ffa89f03-a075-4645-864c-40f29ace9c09","Type":"ContainerStarted","Data":"869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7"} Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.242832 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71c9961c-3b9f-4d2a-b392-abd5eb920a3b" (UID: "71c9961c-3b9f-4d2a-b392-abd5eb920a3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.255931 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "71c9961c-3b9f-4d2a-b392-abd5eb920a3b" (UID: "71c9961c-3b9f-4d2a-b392-abd5eb920a3b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.285640 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.285616452 podStartE2EDuration="3.285616452s" podCreationTimestamp="2026-02-25 08:53:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:53:36.259511943 +0000 UTC m=+7709.698768412" watchObservedRunningTime="2026-02-25 08:53:36.285616452 +0000 UTC m=+7709.724872911" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.349823 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.349860 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/71c9961c-3b9f-4d2a-b392-abd5eb920a3b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.350744 4978 scope.go:117] "RemoveContainer" containerID="51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0" Feb 25 08:53:36 crc kubenswrapper[4978]: E0225 08:53:36.355099 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0\": container with ID starting with 51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0 not found: ID does not exist" containerID="51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.355140 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0"} err="failed to get container status \"51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0\": rpc error: code = NotFound desc = could not find container \"51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0\": container with ID starting with 51b861266bbc85f233e7fced1bb53f8be8f0c0c649cdb4b52f7fac95cabf38a0 not found: ID does not exist" Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.389546 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 25 08:53:36 crc kubenswrapper[4978]: W0225 08:53:36.401788 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4941e0b_0e69_4577_837f_6633db74a609.slice/crio-419669153c7f85582552675c6964dc03bf30645f239fa5d4c4abe7361e42dc67 WatchSource:0}: Error finding container 419669153c7f85582552675c6964dc03bf30645f239fa5d4c4abe7361e42dc67: Status 404 returned error can't find the container with id 419669153c7f85582552675c6964dc03bf30645f239fa5d4c4abe7361e42dc67 Feb 25 08:53:36 crc kubenswrapper[4978]: I0225 08:53:36.523634 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" podUID="ffa89f03-a075-4645-864c-40f29ace9c09" Feb 25 08:53:37 crc kubenswrapper[4978]: I0225 08:53:37.252205 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f4941e0b-0e69-4577-837f-6633db74a609","Type":"ContainerStarted","Data":"419669153c7f85582552675c6964dc03bf30645f239fa5d4c4abe7361e42dc67"} Feb 25 08:53:37 crc kubenswrapper[4978]: I0225 08:53:37.347386 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71c9961c-3b9f-4d2a-b392-abd5eb920a3b" path="/var/lib/kubelet/pods/71c9961c-3b9f-4d2a-b392-abd5eb920a3b/volumes" Feb 25 08:53:41 crc kubenswrapper[4978]: I0225 08:53:41.292947 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"55251903-0563-4a00-8652-96350cc28320","Type":"ContainerStarted","Data":"643098c549ff2cc4aa596550217dc0e7c04d14be25a313bad4af8e071c30f4da"} Feb 25 08:53:41 crc kubenswrapper[4978]: I0225 08:53:41.296881 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f4941e0b-0e69-4577-837f-6633db74a609","Type":"ContainerStarted","Data":"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926"} Feb 25 08:53:44 crc kubenswrapper[4978]: I0225 08:53:44.170946 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 25 08:53:46 crc kubenswrapper[4978]: I0225 08:53:46.540551 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:53:46 crc kubenswrapper[4978]: I0225 08:53:46.540924 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:53:49 crc kubenswrapper[4978]: I0225 08:53:49.391414 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4941e0b-0e69-4577-837f-6633db74a609" containerID="cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926" exitCode=0 Feb 25 08:53:49 crc kubenswrapper[4978]: I0225 08:53:49.391543 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f4941e0b-0e69-4577-837f-6633db74a609","Type":"ContainerDied","Data":"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926"} Feb 25 08:53:50 crc kubenswrapper[4978]: I0225 08:53:50.407039 4978 generic.go:334] "Generic (PLEG): container finished" podID="55251903-0563-4a00-8652-96350cc28320" containerID="643098c549ff2cc4aa596550217dc0e7c04d14be25a313bad4af8e071c30f4da" exitCode=0 Feb 25 08:53:50 crc kubenswrapper[4978]: I0225 08:53:50.407123 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"55251903-0563-4a00-8652-96350cc28320","Type":"ContainerDied","Data":"643098c549ff2cc4aa596550217dc0e7c04d14be25a313bad4af8e071c30f4da"} Feb 25 08:53:56 crc kubenswrapper[4978]: I0225 08:53:56.504030 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"55251903-0563-4a00-8652-96350cc28320","Type":"ContainerStarted","Data":"a926c7679da6baeebcf58cf66d17458f89830dcab11c50e57063bc987f30dcca"} Feb 25 08:53:56 crc kubenswrapper[4978]: I0225 08:53:56.507435 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f4941e0b-0e69-4577-837f-6633db74a609","Type":"ContainerStarted","Data":"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0"} Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.203047 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533494-fqpmk"] Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.213721 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533494-fqpmk"] Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.217562 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533494-fqpmk" Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.220972 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.221435 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.222177 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.330758 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gcpz\" (UniqueName: \"kubernetes.io/projected/95208faf-d12c-48b1-aaab-98bb4c11a953-kube-api-access-5gcpz\") pod \"auto-csr-approver-29533494-fqpmk\" (UID: \"95208faf-d12c-48b1-aaab-98bb4c11a953\") " pod="openshift-infra/auto-csr-approver-29533494-fqpmk" Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.434495 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gcpz\" (UniqueName: \"kubernetes.io/projected/95208faf-d12c-48b1-aaab-98bb4c11a953-kube-api-access-5gcpz\") pod \"auto-csr-approver-29533494-fqpmk\" (UID: \"95208faf-d12c-48b1-aaab-98bb4c11a953\") " pod="openshift-infra/auto-csr-approver-29533494-fqpmk" Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.465912 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gcpz\" (UniqueName: \"kubernetes.io/projected/95208faf-d12c-48b1-aaab-98bb4c11a953-kube-api-access-5gcpz\") pod \"auto-csr-approver-29533494-fqpmk\" (UID: \"95208faf-d12c-48b1-aaab-98bb4c11a953\") " pod="openshift-infra/auto-csr-approver-29533494-fqpmk" Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.558883 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533494-fqpmk" Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.594294 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"55251903-0563-4a00-8652-96350cc28320","Type":"ContainerStarted","Data":"59f12ed33595d45e50b77a017581fe8ece42aa25bbd4580810ed8ea2a3c91e7a"} Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.599280 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f4941e0b-0e69-4577-837f-6633db74a609","Type":"ContainerStarted","Data":"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61"} Feb 25 08:54:00 crc kubenswrapper[4978]: I0225 08:54:00.642017 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.687662812 podStartE2EDuration="26.641990194s" podCreationTimestamp="2026-02-25 08:53:34 +0000 UTC" firstStartedPulling="2026-02-25 08:53:35.50055472 +0000 UTC m=+7708.939811179" lastFinishedPulling="2026-02-25 08:53:55.454882102 +0000 UTC m=+7728.894138561" observedRunningTime="2026-02-25 08:54:00.625445351 +0000 UTC m=+7734.064701840" watchObservedRunningTime="2026-02-25 08:54:00.641990194 +0000 UTC m=+7734.081246693" Feb 25 08:54:01 crc kubenswrapper[4978]: I0225 08:54:01.068138 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533494-fqpmk"] Feb 25 08:54:01 crc kubenswrapper[4978]: I0225 08:54:01.608879 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533494-fqpmk" event={"ID":"95208faf-d12c-48b1-aaab-98bb4c11a953","Type":"ContainerStarted","Data":"6439ddaf676d27d88fd84b76ea5aa9e01bb991dbed6c10d131bfd2a30edbef98"} Feb 25 08:54:01 crc kubenswrapper[4978]: I0225 08:54:01.609340 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Feb 25 08:54:01 crc kubenswrapper[4978]: I0225 08:54:01.612915 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Feb 25 08:54:03 crc kubenswrapper[4978]: I0225 08:54:03.636589 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f4941e0b-0e69-4577-837f-6633db74a609","Type":"ContainerStarted","Data":"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5"} Feb 25 08:54:03 crc kubenswrapper[4978]: I0225 08:54:03.640424 4978 generic.go:334] "Generic (PLEG): container finished" podID="95208faf-d12c-48b1-aaab-98bb4c11a953" containerID="14d1938a85c40947dd515537c458414c5216a230ed0fb5efbf541226d287f404" exitCode=0 Feb 25 08:54:03 crc kubenswrapper[4978]: I0225 08:54:03.640481 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533494-fqpmk" event={"ID":"95208faf-d12c-48b1-aaab-98bb4c11a953","Type":"ContainerDied","Data":"14d1938a85c40947dd515537c458414c5216a230ed0fb5efbf541226d287f404"} Feb 25 08:54:03 crc kubenswrapper[4978]: I0225 08:54:03.678447 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.4276782949999998 podStartE2EDuration="29.678364254s" podCreationTimestamp="2026-02-25 08:53:34 +0000 UTC" firstStartedPulling="2026-02-25 08:53:36.41015107 +0000 UTC m=+7709.849407529" lastFinishedPulling="2026-02-25 08:54:02.660836999 +0000 UTC m=+7736.100093488" observedRunningTime="2026-02-25 08:54:03.677183127 +0000 UTC m=+7737.116439646" watchObservedRunningTime="2026-02-25 08:54:03.678364254 +0000 UTC m=+7737.117620753" Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.048581 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-2cvxh"] Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.063105 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-2cvxh"] Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.162593 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533494-fqpmk" Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.345179 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gcpz\" (UniqueName: \"kubernetes.io/projected/95208faf-d12c-48b1-aaab-98bb4c11a953-kube-api-access-5gcpz\") pod \"95208faf-d12c-48b1-aaab-98bb4c11a953\" (UID: \"95208faf-d12c-48b1-aaab-98bb4c11a953\") " Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.345647 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8280b490-fc6f-4ff5-a758-cda686b2483d" path="/var/lib/kubelet/pods/8280b490-fc6f-4ff5-a758-cda686b2483d/volumes" Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.356356 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95208faf-d12c-48b1-aaab-98bb4c11a953-kube-api-access-5gcpz" (OuterVolumeSpecName: "kube-api-access-5gcpz") pod "95208faf-d12c-48b1-aaab-98bb4c11a953" (UID: "95208faf-d12c-48b1-aaab-98bb4c11a953"). InnerVolumeSpecName "kube-api-access-5gcpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.447991 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gcpz\" (UniqueName: \"kubernetes.io/projected/95208faf-d12c-48b1-aaab-98bb4c11a953-kube-api-access-5gcpz\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.667504 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533494-fqpmk" event={"ID":"95208faf-d12c-48b1-aaab-98bb4c11a953","Type":"ContainerDied","Data":"6439ddaf676d27d88fd84b76ea5aa9e01bb991dbed6c10d131bfd2a30edbef98"} Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.667573 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533494-fqpmk" Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.667588 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6439ddaf676d27d88fd84b76ea5aa9e01bb991dbed6c10d131bfd2a30edbef98" Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.808255 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.808324 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:05 crc kubenswrapper[4978]: I0225 08:54:05.811530 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:06 crc kubenswrapper[4978]: I0225 08:54:06.029999 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-da55-account-create-update-zksjb"] Feb 25 08:54:06 crc kubenswrapper[4978]: I0225 08:54:06.039533 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-da55-account-create-update-zksjb"] Feb 25 08:54:06 crc kubenswrapper[4978]: I0225 08:54:06.238047 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533488-pfcfp"] Feb 25 08:54:06 crc kubenswrapper[4978]: I0225 08:54:06.249678 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533488-pfcfp"] Feb 25 08:54:06 crc kubenswrapper[4978]: I0225 08:54:06.683205 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:07 crc kubenswrapper[4978]: I0225 08:54:07.349296 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cb4a363-4e1d-4b95-bacd-ab271457785f" path="/var/lib/kubelet/pods/8cb4a363-4e1d-4b95-bacd-ab271457785f/volumes" Feb 25 08:54:07 crc kubenswrapper[4978]: I0225 08:54:07.351344 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7c59b20-c701-4e2c-8383-03f6d93b5551" path="/var/lib/kubelet/pods/f7c59b20-c701-4e2c-8383-03f6d93b5551/volumes" Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.831196 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.832068 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="ffa89f03-a075-4645-864c-40f29ace9c09" containerName="openstackclient" containerID="cri-o://869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7" gracePeriod=2 Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.845262 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.882058 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 25 08:54:08 crc kubenswrapper[4978]: E0225 08:54:08.882478 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95208faf-d12c-48b1-aaab-98bb4c11a953" containerName="oc" Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.882499 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="95208faf-d12c-48b1-aaab-98bb4c11a953" containerName="oc" Feb 25 08:54:08 crc kubenswrapper[4978]: E0225 08:54:08.882525 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa89f03-a075-4645-864c-40f29ace9c09" containerName="openstackclient" Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.882533 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa89f03-a075-4645-864c-40f29ace9c09" containerName="openstackclient" Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.882704 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="95208faf-d12c-48b1-aaab-98bb4c11a953" containerName="oc" Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.882731 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa89f03-a075-4645-864c-40f29ace9c09" containerName="openstackclient" Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.883448 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.909949 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ffa89f03-a075-4645-864c-40f29ace9c09" podUID="16cdf7b2-0453-4b38-aba4-c44a59e2eee4" Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.926608 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.946272 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Feb 25 08:54:08 crc kubenswrapper[4978]: E0225 08:54:08.946607 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-g7prb openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="16cdf7b2-0453-4b38-aba4-c44a59e2eee4" Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.961182 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.976406 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.977699 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:54:08 crc kubenswrapper[4978]: I0225 08:54:08.987533 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.036702 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.036753 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7prb\" (UniqueName: \"kubernetes.io/projected/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-kube-api-access-g7prb\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.036822 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.036873 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config-secret\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.138011 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.138067 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnz2m\" (UniqueName: \"kubernetes.io/projected/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-kube-api-access-dnz2m\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.138103 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.138124 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-openstack-config\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.138152 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7prb\" (UniqueName: \"kubernetes.io/projected/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-kube-api-access-g7prb\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.138340 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-openstack-config-secret\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.138415 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.138522 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config-secret\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.139424 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: E0225 08:54:09.140776 4978 projected.go:194] Error preparing data for projected volume kube-api-access-g7prb for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (16cdf7b2-0453-4b38-aba4-c44a59e2eee4) does not match the UID in record. The object might have been deleted and then recreated Feb 25 08:54:09 crc kubenswrapper[4978]: E0225 08:54:09.140856 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-kube-api-access-g7prb podName:16cdf7b2-0453-4b38-aba4-c44a59e2eee4 nodeName:}" failed. No retries permitted until 2026-02-25 08:54:09.640832536 +0000 UTC m=+7743.080089095 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-g7prb" (UniqueName: "kubernetes.io/projected/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-kube-api-access-g7prb") pod "openstackclient" (UID: "16cdf7b2-0453-4b38-aba4-c44a59e2eee4") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (16cdf7b2-0453-4b38-aba4-c44a59e2eee4) does not match the UID in record. The object might have been deleted and then recreated Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.143678 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.146937 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config-secret\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.240229 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.240291 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnz2m\" (UniqueName: \"kubernetes.io/projected/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-kube-api-access-dnz2m\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.240330 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-openstack-config\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.240438 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-openstack-config-secret\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.241268 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-openstack-config\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.243449 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-openstack-config-secret\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.243894 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.258173 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnz2m\" (UniqueName: \"kubernetes.io/projected/1a5e0a02-7fb2-4745-b49d-fd5b358c3122-kube-api-access-dnz2m\") pod \"openstackclient\" (UID: \"1a5e0a02-7fb2-4745-b49d-fd5b358c3122\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.294981 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.651236 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7prb\" (UniqueName: \"kubernetes.io/projected/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-kube-api-access-g7prb\") pod \"openstackclient\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: E0225 08:54:09.653307 4978 projected.go:194] Error preparing data for projected volume kube-api-access-g7prb for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (16cdf7b2-0453-4b38-aba4-c44a59e2eee4) does not match the UID in record. The object might have been deleted and then recreated Feb 25 08:54:09 crc kubenswrapper[4978]: E0225 08:54:09.653386 4978 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-kube-api-access-g7prb podName:16cdf7b2-0453-4b38-aba4-c44a59e2eee4 nodeName:}" failed. No retries permitted until 2026-02-25 08:54:10.653353345 +0000 UTC m=+7744.092609804 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-g7prb" (UniqueName: "kubernetes.io/projected/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-kube-api-access-g7prb") pod "openstackclient" (UID: "16cdf7b2-0453-4b38-aba4-c44a59e2eee4") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (16cdf7b2-0453-4b38-aba4-c44a59e2eee4) does not match the UID in record. The object might have been deleted and then recreated Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.718517 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.734295 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.737588 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="16cdf7b2-0453-4b38-aba4-c44a59e2eee4" podUID="1a5e0a02-7fb2-4745-b49d-fd5b358c3122" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.855478 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-combined-ca-bundle\") pod \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.855744 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config-secret\") pod \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.855842 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config\") pod \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\" (UID: \"16cdf7b2-0453-4b38-aba4-c44a59e2eee4\") " Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.856558 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7prb\" (UniqueName: \"kubernetes.io/projected/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-kube-api-access-g7prb\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.857469 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "16cdf7b2-0453-4b38-aba4-c44a59e2eee4" (UID: "16cdf7b2-0453-4b38-aba4-c44a59e2eee4"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.860127 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16cdf7b2-0453-4b38-aba4-c44a59e2eee4" (UID: "16cdf7b2-0453-4b38-aba4-c44a59e2eee4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.860552 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.863217 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "16cdf7b2-0453-4b38-aba4-c44a59e2eee4" (UID: "16cdf7b2-0453-4b38-aba4-c44a59e2eee4"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:09 crc kubenswrapper[4978]: W0225 08:54:09.885500 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a5e0a02_7fb2_4745_b49d_fd5b358c3122.slice/crio-1ad11d2dc64ae4aeb0243e4abe42eb773873f18be72a9ee28a2227f6f1d557c1 WatchSource:0}: Error finding container 1ad11d2dc64ae4aeb0243e4abe42eb773873f18be72a9ee28a2227f6f1d557c1: Status 404 returned error can't find the container with id 1ad11d2dc64ae4aeb0243e4abe42eb773873f18be72a9ee28a2227f6f1d557c1 Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.958514 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.958549 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:09 crc kubenswrapper[4978]: I0225 08:54:09.958560 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/16cdf7b2-0453-4b38-aba4-c44a59e2eee4-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:10 crc kubenswrapper[4978]: I0225 08:54:10.633004 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 25 08:54:10 crc kubenswrapper[4978]: I0225 08:54:10.633807 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="prometheus" containerID="cri-o://1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0" gracePeriod=600 Feb 25 08:54:10 crc kubenswrapper[4978]: I0225 08:54:10.633884 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="thanos-sidecar" containerID="cri-o://e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5" gracePeriod=600 Feb 25 08:54:10 crc kubenswrapper[4978]: I0225 08:54:10.633882 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="config-reloader" containerID="cri-o://9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61" gracePeriod=600 Feb 25 08:54:10 crc kubenswrapper[4978]: I0225 08:54:10.728160 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:54:10 crc kubenswrapper[4978]: I0225 08:54:10.731513 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1a5e0a02-7fb2-4745-b49d-fd5b358c3122","Type":"ContainerStarted","Data":"892be1f31923fabc8d0509703bb1a2235d1b83a5b36f8ec9e81fd0387abaed4c"} Feb 25 08:54:10 crc kubenswrapper[4978]: I0225 08:54:10.731567 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1a5e0a02-7fb2-4745-b49d-fd5b358c3122","Type":"ContainerStarted","Data":"1ad11d2dc64ae4aeb0243e4abe42eb773873f18be72a9ee28a2227f6f1d557c1"} Feb 25 08:54:10 crc kubenswrapper[4978]: I0225 08:54:10.756041 4978 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="16cdf7b2-0453-4b38-aba4-c44a59e2eee4" podUID="1a5e0a02-7fb2-4745-b49d-fd5b358c3122" Feb 25 08:54:10 crc kubenswrapper[4978]: I0225 08:54:10.757408 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.757386408 podStartE2EDuration="2.757386408s" podCreationTimestamp="2026-02-25 08:54:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:54:10.750815385 +0000 UTC m=+7744.190071844" watchObservedRunningTime="2026-02-25 08:54:10.757386408 +0000 UTC m=+7744.196642867" Feb 25 08:54:10 crc kubenswrapper[4978]: I0225 08:54:10.808008 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.1.195:9090/-/ready\": dial tcp 10.217.1.195:9090: connect: connection refused" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.125714 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.279202 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjwmm\" (UniqueName: \"kubernetes.io/projected/ffa89f03-a075-4645-864c-40f29ace9c09-kube-api-access-fjwmm\") pod \"ffa89f03-a075-4645-864c-40f29ace9c09\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.279686 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config-secret\") pod \"ffa89f03-a075-4645-864c-40f29ace9c09\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.279729 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-combined-ca-bundle\") pod \"ffa89f03-a075-4645-864c-40f29ace9c09\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.279830 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config\") pod \"ffa89f03-a075-4645-864c-40f29ace9c09\" (UID: \"ffa89f03-a075-4645-864c-40f29ace9c09\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.288597 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffa89f03-a075-4645-864c-40f29ace9c09-kube-api-access-fjwmm" (OuterVolumeSpecName: "kube-api-access-fjwmm") pod "ffa89f03-a075-4645-864c-40f29ace9c09" (UID: "ffa89f03-a075-4645-864c-40f29ace9c09"). InnerVolumeSpecName "kube-api-access-fjwmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.317182 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ffa89f03-a075-4645-864c-40f29ace9c09" (UID: "ffa89f03-a075-4645-864c-40f29ace9c09"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.346525 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ffa89f03-a075-4645-864c-40f29ace9c09" (UID: "ffa89f03-a075-4645-864c-40f29ace9c09"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.351335 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16cdf7b2-0453-4b38-aba4-c44a59e2eee4" path="/var/lib/kubelet/pods/16cdf7b2-0453-4b38-aba4-c44a59e2eee4/volumes" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.354624 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ffa89f03-a075-4645-864c-40f29ace9c09" (UID: "ffa89f03-a075-4645-864c-40f29ace9c09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.381640 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.381671 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffa89f03-a075-4645-864c-40f29ace9c09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.381680 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ffa89f03-a075-4645-864c-40f29ace9c09-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.381691 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjwmm\" (UniqueName: \"kubernetes.io/projected/ffa89f03-a075-4645-864c-40f29ace9c09-kube-api-access-fjwmm\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.426428 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.584582 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-tls-assets\") pod \"f4941e0b-0e69-4577-837f-6633db74a609\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.584657 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-web-config\") pod \"f4941e0b-0e69-4577-837f-6633db74a609\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.584728 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-config\") pod \"f4941e0b-0e69-4577-837f-6633db74a609\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.584863 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-0\") pod \"f4941e0b-0e69-4577-837f-6633db74a609\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.584911 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-2\") pod \"f4941e0b-0e69-4577-837f-6633db74a609\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.584951 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f4941e0b-0e69-4577-837f-6633db74a609-config-out\") pod \"f4941e0b-0e69-4577-837f-6633db74a609\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.584980 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-1\") pod \"f4941e0b-0e69-4577-837f-6633db74a609\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.585073 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-thanos-prometheus-http-client-file\") pod \"f4941e0b-0e69-4577-837f-6633db74a609\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.585231 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lxw5\" (UniqueName: \"kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-kube-api-access-5lxw5\") pod \"f4941e0b-0e69-4577-837f-6633db74a609\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.585482 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") pod \"f4941e0b-0e69-4577-837f-6633db74a609\" (UID: \"f4941e0b-0e69-4577-837f-6633db74a609\") " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.592643 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4941e0b-0e69-4577-837f-6633db74a609-config-out" (OuterVolumeSpecName: "config-out") pod "f4941e0b-0e69-4577-837f-6633db74a609" (UID: "f4941e0b-0e69-4577-837f-6633db74a609"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.593186 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "f4941e0b-0e69-4577-837f-6633db74a609" (UID: "f4941e0b-0e69-4577-837f-6633db74a609"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.593488 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "f4941e0b-0e69-4577-837f-6633db74a609" (UID: "f4941e0b-0e69-4577-837f-6633db74a609"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.594822 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "f4941e0b-0e69-4577-837f-6633db74a609" (UID: "f4941e0b-0e69-4577-837f-6633db74a609"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.599687 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "f4941e0b-0e69-4577-837f-6633db74a609" (UID: "f4941e0b-0e69-4577-837f-6633db74a609"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.603034 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-config" (OuterVolumeSpecName: "config") pod "f4941e0b-0e69-4577-837f-6633db74a609" (UID: "f4941e0b-0e69-4577-837f-6633db74a609"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.604830 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-kube-api-access-5lxw5" (OuterVolumeSpecName: "kube-api-access-5lxw5") pod "f4941e0b-0e69-4577-837f-6633db74a609" (UID: "f4941e0b-0e69-4577-837f-6633db74a609"). InnerVolumeSpecName "kube-api-access-5lxw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.611877 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "f4941e0b-0e69-4577-837f-6633db74a609" (UID: "f4941e0b-0e69-4577-837f-6633db74a609"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.626528 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-web-config" (OuterVolumeSpecName: "web-config") pod "f4941e0b-0e69-4577-837f-6633db74a609" (UID: "f4941e0b-0e69-4577-837f-6633db74a609"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.689974 4978 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-tls-assets\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.690000 4978 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-web-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.690008 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.690017 4978 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.690026 4978 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.690036 4978 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f4941e0b-0e69-4577-837f-6633db74a609-config-out\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.690046 4978 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/f4941e0b-0e69-4577-837f-6633db74a609-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.690054 4978 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f4941e0b-0e69-4577-837f-6633db74a609-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.690063 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lxw5\" (UniqueName: \"kubernetes.io/projected/f4941e0b-0e69-4577-837f-6633db74a609-kube-api-access-5lxw5\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.748515 4978 generic.go:334] "Generic (PLEG): container finished" podID="ffa89f03-a075-4645-864c-40f29ace9c09" containerID="869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7" exitCode=137 Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.748588 4978 scope.go:117] "RemoveContainer" containerID="869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.748697 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.776216 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4941e0b-0e69-4577-837f-6633db74a609" containerID="e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5" exitCode=0 Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.776247 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4941e0b-0e69-4577-837f-6633db74a609" containerID="9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61" exitCode=0 Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.776257 4978 generic.go:334] "Generic (PLEG): container finished" podID="f4941e0b-0e69-4577-837f-6633db74a609" containerID="1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0" exitCode=0 Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.777188 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.777446 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f4941e0b-0e69-4577-837f-6633db74a609","Type":"ContainerDied","Data":"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5"} Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.777506 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f4941e0b-0e69-4577-837f-6633db74a609","Type":"ContainerDied","Data":"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61"} Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.777519 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f4941e0b-0e69-4577-837f-6633db74a609","Type":"ContainerDied","Data":"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0"} Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.777530 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f4941e0b-0e69-4577-837f-6633db74a609","Type":"ContainerDied","Data":"419669153c7f85582552675c6964dc03bf30645f239fa5d4c4abe7361e42dc67"} Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.809649 4978 scope.go:117] "RemoveContainer" containerID="869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7" Feb 25 08:54:11 crc kubenswrapper[4978]: E0225 08:54:11.810171 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7\": container with ID starting with 869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7 not found: ID does not exist" containerID="869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.810220 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7"} err="failed to get container status \"869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7\": rpc error: code = NotFound desc = could not find container \"869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7\": container with ID starting with 869b074d2e8dd7e0812ae1d845afb379b439336a635b7ddc882bb0b49d1e10c7 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.810247 4978 scope.go:117] "RemoveContainer" containerID="e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.850327 4978 scope.go:117] "RemoveContainer" containerID="9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.865845 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "f4941e0b-0e69-4577-837f-6633db74a609" (UID: "f4941e0b-0e69-4577-837f-6633db74a609"). InnerVolumeSpecName "pvc-02862e46-9f8b-46b3-930d-0200c6093953". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.872504 4978 scope.go:117] "RemoveContainer" containerID="1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.891544 4978 scope.go:117] "RemoveContainer" containerID="cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.895399 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-02862e46-9f8b-46b3-930d-0200c6093953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") on node \"crc\" " Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.914262 4978 scope.go:117] "RemoveContainer" containerID="e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5" Feb 25 08:54:11 crc kubenswrapper[4978]: E0225 08:54:11.914800 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5\": container with ID starting with e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5 not found: ID does not exist" containerID="e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.914834 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5"} err="failed to get container status \"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5\": rpc error: code = NotFound desc = could not find container \"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5\": container with ID starting with e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.914862 4978 scope.go:117] "RemoveContainer" containerID="9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61" Feb 25 08:54:11 crc kubenswrapper[4978]: E0225 08:54:11.915641 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61\": container with ID starting with 9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61 not found: ID does not exist" containerID="9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.915685 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61"} err="failed to get container status \"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61\": rpc error: code = NotFound desc = could not find container \"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61\": container with ID starting with 9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.915713 4978 scope.go:117] "RemoveContainer" containerID="1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0" Feb 25 08:54:11 crc kubenswrapper[4978]: E0225 08:54:11.915978 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0\": container with ID starting with 1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0 not found: ID does not exist" containerID="1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.915998 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0"} err="failed to get container status \"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0\": rpc error: code = NotFound desc = could not find container \"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0\": container with ID starting with 1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.916011 4978 scope.go:117] "RemoveContainer" containerID="cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926" Feb 25 08:54:11 crc kubenswrapper[4978]: E0225 08:54:11.916220 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926\": container with ID starting with cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926 not found: ID does not exist" containerID="cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.916242 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926"} err="failed to get container status \"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926\": rpc error: code = NotFound desc = could not find container \"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926\": container with ID starting with cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.916256 4978 scope.go:117] "RemoveContainer" containerID="e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.916952 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5"} err="failed to get container status \"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5\": rpc error: code = NotFound desc = could not find container \"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5\": container with ID starting with e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.917050 4978 scope.go:117] "RemoveContainer" containerID="9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.917322 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61"} err="failed to get container status \"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61\": rpc error: code = NotFound desc = could not find container \"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61\": container with ID starting with 9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.917344 4978 scope.go:117] "RemoveContainer" containerID="1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.917549 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0"} err="failed to get container status \"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0\": rpc error: code = NotFound desc = could not find container \"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0\": container with ID starting with 1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.917572 4978 scope.go:117] "RemoveContainer" containerID="cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.917887 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926"} err="failed to get container status \"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926\": rpc error: code = NotFound desc = could not find container \"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926\": container with ID starting with cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.917907 4978 scope.go:117] "RemoveContainer" containerID="e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.918152 4978 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.918234 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5"} err="failed to get container status \"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5\": rpc error: code = NotFound desc = could not find container \"e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5\": container with ID starting with e97a11b723f4c0b4d6487353e85a72a081a77177594393c8340011b20be9cab5 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.918274 4978 scope.go:117] "RemoveContainer" containerID="9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.918290 4978 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-02862e46-9f8b-46b3-930d-0200c6093953" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953") on node "crc" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.918524 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61"} err="failed to get container status \"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61\": rpc error: code = NotFound desc = could not find container \"9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61\": container with ID starting with 9a5fd377173dfc374fe5e2f9c747d6171b1643cb00779fe9f9a9a971ab167e61 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.918547 4978 scope.go:117] "RemoveContainer" containerID="1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.918923 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0"} err="failed to get container status \"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0\": rpc error: code = NotFound desc = could not find container \"1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0\": container with ID starting with 1dca1f17378486d6afcacda617b427629fb081857e5b15f7d49c0eacc75d3aa0 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.918942 4978 scope.go:117] "RemoveContainer" containerID="cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.919286 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926"} err="failed to get container status \"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926\": rpc error: code = NotFound desc = could not find container \"cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926\": container with ID starting with cf9cc5d5aea3e7b32e933c125cc1c37cde3ef971758d5f4055890746ba387926 not found: ID does not exist" Feb 25 08:54:11 crc kubenswrapper[4978]: I0225 08:54:11.998310 4978 reconciler_common.go:293] "Volume detached for volume \"pvc-02862e46-9f8b-46b3-930d-0200c6093953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.137058 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.145447 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.178856 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 25 08:54:12 crc kubenswrapper[4978]: E0225 08:54:12.179224 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="config-reloader" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.179244 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="config-reloader" Feb 25 08:54:12 crc kubenswrapper[4978]: E0225 08:54:12.179266 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="prometheus" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.179275 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="prometheus" Feb 25 08:54:12 crc kubenswrapper[4978]: E0225 08:54:12.179292 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="init-config-reloader" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.179300 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="init-config-reloader" Feb 25 08:54:12 crc kubenswrapper[4978]: E0225 08:54:12.179341 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="thanos-sidecar" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.179350 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="thanos-sidecar" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.179569 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="config-reloader" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.179595 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="prometheus" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.179623 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4941e0b-0e69-4577-837f-6633db74a609" containerName="thanos-sidecar" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.181845 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.184178 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.184300 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-hp7ht" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.186129 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.186132 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.186325 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.186452 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.186480 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.186804 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.195389 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.204431 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304016 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304072 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304152 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3e3249bd-3d1b-41a2-9132-976e11161c91-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304183 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3e3249bd-3d1b-41a2-9132-976e11161c91-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304215 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/3e3249bd-3d1b-41a2-9132-976e11161c91-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304287 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304322 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304342 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-02862e46-9f8b-46b3-930d-0200c6093953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304379 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304402 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304423 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzh52\" (UniqueName: \"kubernetes.io/projected/3e3249bd-3d1b-41a2-9132-976e11161c91-kube-api-access-tzh52\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304438 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3e3249bd-3d1b-41a2-9132-976e11161c91-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.304460 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/3e3249bd-3d1b-41a2-9132-976e11161c91-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405517 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/3e3249bd-3d1b-41a2-9132-976e11161c91-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405604 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405664 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405689 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-02862e46-9f8b-46b3-930d-0200c6093953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405713 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405734 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405756 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzh52\" (UniqueName: \"kubernetes.io/projected/3e3249bd-3d1b-41a2-9132-976e11161c91-kube-api-access-tzh52\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405772 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3e3249bd-3d1b-41a2-9132-976e11161c91-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405798 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/3e3249bd-3d1b-41a2-9132-976e11161c91-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405843 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405874 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405907 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3e3249bd-3d1b-41a2-9132-976e11161c91-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.405922 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3e3249bd-3d1b-41a2-9132-976e11161c91-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.408005 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/3e3249bd-3d1b-41a2-9132-976e11161c91-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.409296 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3e3249bd-3d1b-41a2-9132-976e11161c91-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.410084 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.410121 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-02862e46-9f8b-46b3-930d-0200c6093953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4d0b98a232139d7495440a8b884ebd34044c8aae36a3e9220f1b8cefdfc7f1e5/globalmount\"" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.411131 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.411909 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.412137 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/3e3249bd-3d1b-41a2-9132-976e11161c91-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.412529 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.412998 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.413686 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3e3249bd-3d1b-41a2-9132-976e11161c91-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.414904 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.416574 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3e3249bd-3d1b-41a2-9132-976e11161c91-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.427064 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/3e3249bd-3d1b-41a2-9132-976e11161c91-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.427307 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzh52\" (UniqueName: \"kubernetes.io/projected/3e3249bd-3d1b-41a2-9132-976e11161c91-kube-api-access-tzh52\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.470808 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-02862e46-9f8b-46b3-930d-0200c6093953\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-02862e46-9f8b-46b3-930d-0200c6093953\") pod \"prometheus-metric-storage-0\" (UID: \"3e3249bd-3d1b-41a2-9132-976e11161c91\") " pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.549800 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:12 crc kubenswrapper[4978]: I0225 08:54:12.988820 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Feb 25 08:54:12 crc kubenswrapper[4978]: W0225 08:54:12.992775 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e3249bd_3d1b_41a2_9132_976e11161c91.slice/crio-09d02fbce1a92657a84c500b32c521ebfba7728d338e7527d8a9866744f2adec WatchSource:0}: Error finding container 09d02fbce1a92657a84c500b32c521ebfba7728d338e7527d8a9866744f2adec: Status 404 returned error can't find the container with id 09d02fbce1a92657a84c500b32c521ebfba7728d338e7527d8a9866744f2adec Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.346296 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4941e0b-0e69-4577-837f-6633db74a609" path="/var/lib/kubelet/pods/f4941e0b-0e69-4577-837f-6633db74a609/volumes" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.347745 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffa89f03-a075-4645-864c-40f29ace9c09" path="/var/lib/kubelet/pods/ffa89f03-a075-4645-864c-40f29ace9c09/volumes" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.348383 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.352773 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.353083 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.355798 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.355952 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.422785 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-log-httpd\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.422858 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-config-data\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.422884 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-run-httpd\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.422945 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-scripts\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.423056 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.423197 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.423314 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlcwg\" (UniqueName: \"kubernetes.io/projected/60eef5db-6892-460b-bda6-6037affe3fdf-kube-api-access-dlcwg\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.525249 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-log-httpd\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.525302 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-config-data\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.525320 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-run-httpd\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.525354 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-scripts\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.525381 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.525415 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.525450 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlcwg\" (UniqueName: \"kubernetes.io/projected/60eef5db-6892-460b-bda6-6037affe3fdf-kube-api-access-dlcwg\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.525991 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-log-httpd\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.526358 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-run-httpd\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.531476 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.532734 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-scripts\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.532811 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.533678 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-config-data\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.566747 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlcwg\" (UniqueName: \"kubernetes.io/projected/60eef5db-6892-460b-bda6-6037affe3fdf-kube-api-access-dlcwg\") pod \"ceilometer-0\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.674684 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.808092 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e3249bd-3d1b-41a2-9132-976e11161c91","Type":"ContainerStarted","Data":"09d02fbce1a92657a84c500b32c521ebfba7728d338e7527d8a9866744f2adec"} Feb 25 08:54:13 crc kubenswrapper[4978]: I0225 08:54:13.977782 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:54:13 crc kubenswrapper[4978]: W0225 08:54:13.980072 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60eef5db_6892_460b_bda6_6037affe3fdf.slice/crio-125da5e39ed952073aa4a3acb3d63d657de81ae6de51b8cb7f8e940ec61908a7 WatchSource:0}: Error finding container 125da5e39ed952073aa4a3acb3d63d657de81ae6de51b8cb7f8e940ec61908a7: Status 404 returned error can't find the container with id 125da5e39ed952073aa4a3acb3d63d657de81ae6de51b8cb7f8e940ec61908a7 Feb 25 08:54:14 crc kubenswrapper[4978]: I0225 08:54:14.822411 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60eef5db-6892-460b-bda6-6037affe3fdf","Type":"ContainerStarted","Data":"125da5e39ed952073aa4a3acb3d63d657de81ae6de51b8cb7f8e940ec61908a7"} Feb 25 08:54:16 crc kubenswrapper[4978]: I0225 08:54:16.540896 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:54:16 crc kubenswrapper[4978]: I0225 08:54:16.541189 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:54:16 crc kubenswrapper[4978]: I0225 08:54:16.847799 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e3249bd-3d1b-41a2-9132-976e11161c91","Type":"ContainerStarted","Data":"ffbb46fd2dc293968f9bb85d3bdb897cd1482cbc00732b1eeb24cad27e4c730f"} Feb 25 08:54:18 crc kubenswrapper[4978]: I0225 08:54:18.871817 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60eef5db-6892-460b-bda6-6037affe3fdf","Type":"ContainerStarted","Data":"9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47"} Feb 25 08:54:19 crc kubenswrapper[4978]: I0225 08:54:19.884498 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60eef5db-6892-460b-bda6-6037affe3fdf","Type":"ContainerStarted","Data":"4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27"} Feb 25 08:54:19 crc kubenswrapper[4978]: I0225 08:54:19.884857 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60eef5db-6892-460b-bda6-6037affe3fdf","Type":"ContainerStarted","Data":"88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a"} Feb 25 08:54:21 crc kubenswrapper[4978]: I0225 08:54:21.917214 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60eef5db-6892-460b-bda6-6037affe3fdf","Type":"ContainerStarted","Data":"5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2"} Feb 25 08:54:21 crc kubenswrapper[4978]: I0225 08:54:21.917848 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 08:54:21 crc kubenswrapper[4978]: I0225 08:54:21.950027 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.511957338 podStartE2EDuration="8.949987805s" podCreationTimestamp="2026-02-25 08:54:13 +0000 UTC" firstStartedPulling="2026-02-25 08:54:13.981858325 +0000 UTC m=+7747.421114774" lastFinishedPulling="2026-02-25 08:54:21.419888772 +0000 UTC m=+7754.859145241" observedRunningTime="2026-02-25 08:54:21.944559716 +0000 UTC m=+7755.383816195" watchObservedRunningTime="2026-02-25 08:54:21.949987805 +0000 UTC m=+7755.389244294" Feb 25 08:54:24 crc kubenswrapper[4978]: I0225 08:54:24.950603 4978 generic.go:334] "Generic (PLEG): container finished" podID="3e3249bd-3d1b-41a2-9132-976e11161c91" containerID="ffbb46fd2dc293968f9bb85d3bdb897cd1482cbc00732b1eeb24cad27e4c730f" exitCode=0 Feb 25 08:54:24 crc kubenswrapper[4978]: I0225 08:54:24.951051 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e3249bd-3d1b-41a2-9132-976e11161c91","Type":"ContainerDied","Data":"ffbb46fd2dc293968f9bb85d3bdb897cd1482cbc00732b1eeb24cad27e4c730f"} Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.688783 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-9ssj7"] Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.690396 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9ssj7" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.708062 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-9ssj7"] Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.794736 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-072e-account-create-update-g24l2"] Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.796413 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-072e-account-create-update-g24l2" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.798522 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.813133 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-072e-account-create-update-g24l2"] Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.830997 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hghk8\" (UniqueName: \"kubernetes.io/projected/a86f5c5d-c083-48de-802a-14e448d14f86-kube-api-access-hghk8\") pod \"aodh-db-create-9ssj7\" (UID: \"a86f5c5d-c083-48de-802a-14e448d14f86\") " pod="openstack/aodh-db-create-9ssj7" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.831261 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a86f5c5d-c083-48de-802a-14e448d14f86-operator-scripts\") pod \"aodh-db-create-9ssj7\" (UID: \"a86f5c5d-c083-48de-802a-14e448d14f86\") " pod="openstack/aodh-db-create-9ssj7" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.933245 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a86f5c5d-c083-48de-802a-14e448d14f86-operator-scripts\") pod \"aodh-db-create-9ssj7\" (UID: \"a86f5c5d-c083-48de-802a-14e448d14f86\") " pod="openstack/aodh-db-create-9ssj7" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.933576 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-operator-scripts\") pod \"aodh-072e-account-create-update-g24l2\" (UID: \"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9\") " pod="openstack/aodh-072e-account-create-update-g24l2" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.933613 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hghk8\" (UniqueName: \"kubernetes.io/projected/a86f5c5d-c083-48de-802a-14e448d14f86-kube-api-access-hghk8\") pod \"aodh-db-create-9ssj7\" (UID: \"a86f5c5d-c083-48de-802a-14e448d14f86\") " pod="openstack/aodh-db-create-9ssj7" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.933990 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a86f5c5d-c083-48de-802a-14e448d14f86-operator-scripts\") pod \"aodh-db-create-9ssj7\" (UID: \"a86f5c5d-c083-48de-802a-14e448d14f86\") " pod="openstack/aodh-db-create-9ssj7" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.934013 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gghl6\" (UniqueName: \"kubernetes.io/projected/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-kube-api-access-gghl6\") pod \"aodh-072e-account-create-update-g24l2\" (UID: \"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9\") " pod="openstack/aodh-072e-account-create-update-g24l2" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.954332 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hghk8\" (UniqueName: \"kubernetes.io/projected/a86f5c5d-c083-48de-802a-14e448d14f86-kube-api-access-hghk8\") pod \"aodh-db-create-9ssj7\" (UID: \"a86f5c5d-c083-48de-802a-14e448d14f86\") " pod="openstack/aodh-db-create-9ssj7" Feb 25 08:54:25 crc kubenswrapper[4978]: I0225 08:54:25.960609 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e3249bd-3d1b-41a2-9132-976e11161c91","Type":"ContainerStarted","Data":"5ea1506af31c70d2945359f46dc52f8dce98fe66f9d2c84eea6fc8fe97904b47"} Feb 25 08:54:26 crc kubenswrapper[4978]: I0225 08:54:26.011065 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9ssj7" Feb 25 08:54:26 crc kubenswrapper[4978]: I0225 08:54:26.036779 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gghl6\" (UniqueName: \"kubernetes.io/projected/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-kube-api-access-gghl6\") pod \"aodh-072e-account-create-update-g24l2\" (UID: \"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9\") " pod="openstack/aodh-072e-account-create-update-g24l2" Feb 25 08:54:26 crc kubenswrapper[4978]: I0225 08:54:26.036939 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-operator-scripts\") pod \"aodh-072e-account-create-update-g24l2\" (UID: \"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9\") " pod="openstack/aodh-072e-account-create-update-g24l2" Feb 25 08:54:26 crc kubenswrapper[4978]: I0225 08:54:26.039336 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-operator-scripts\") pod \"aodh-072e-account-create-update-g24l2\" (UID: \"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9\") " pod="openstack/aodh-072e-account-create-update-g24l2" Feb 25 08:54:26 crc kubenswrapper[4978]: I0225 08:54:26.055880 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gghl6\" (UniqueName: \"kubernetes.io/projected/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-kube-api-access-gghl6\") pod \"aodh-072e-account-create-update-g24l2\" (UID: \"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9\") " pod="openstack/aodh-072e-account-create-update-g24l2" Feb 25 08:54:26 crc kubenswrapper[4978]: I0225 08:54:26.125221 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-072e-account-create-update-g24l2" Feb 25 08:54:26 crc kubenswrapper[4978]: W0225 08:54:26.521970 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda86f5c5d_c083_48de_802a_14e448d14f86.slice/crio-a14d83c148deaf05471faab82926a664b56f9674fd94735964ab4e51632173fd WatchSource:0}: Error finding container a14d83c148deaf05471faab82926a664b56f9674fd94735964ab4e51632173fd: Status 404 returned error can't find the container with id a14d83c148deaf05471faab82926a664b56f9674fd94735964ab4e51632173fd Feb 25 08:54:26 crc kubenswrapper[4978]: I0225 08:54:26.523797 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-9ssj7"] Feb 25 08:54:26 crc kubenswrapper[4978]: I0225 08:54:26.673018 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-072e-account-create-update-g24l2"] Feb 25 08:54:26 crc kubenswrapper[4978]: W0225 08:54:26.676713 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3bd6215_9ff7_4e0a_b137_eb7f358f5ac9.slice/crio-5db041f4429e112b77c8abea2a19fa363ff457580dabaa524882b26402a26ffc WatchSource:0}: Error finding container 5db041f4429e112b77c8abea2a19fa363ff457580dabaa524882b26402a26ffc: Status 404 returned error can't find the container with id 5db041f4429e112b77c8abea2a19fa363ff457580dabaa524882b26402a26ffc Feb 25 08:54:26 crc kubenswrapper[4978]: I0225 08:54:26.969320 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-072e-account-create-update-g24l2" event={"ID":"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9","Type":"ContainerStarted","Data":"5db041f4429e112b77c8abea2a19fa363ff457580dabaa524882b26402a26ffc"} Feb 25 08:54:26 crc kubenswrapper[4978]: I0225 08:54:26.970556 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9ssj7" event={"ID":"a86f5c5d-c083-48de-802a-14e448d14f86","Type":"ContainerStarted","Data":"a14d83c148deaf05471faab82926a664b56f9674fd94735964ab4e51632173fd"} Feb 25 08:54:28 crc kubenswrapper[4978]: I0225 08:54:28.996156 4978 generic.go:334] "Generic (PLEG): container finished" podID="b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9" containerID="9161200378d1f79de6f0c779c891966368613ba5aef3d369d83ba1e49285975c" exitCode=0 Feb 25 08:54:28 crc kubenswrapper[4978]: I0225 08:54:28.996807 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-072e-account-create-update-g24l2" event={"ID":"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9","Type":"ContainerDied","Data":"9161200378d1f79de6f0c779c891966368613ba5aef3d369d83ba1e49285975c"} Feb 25 08:54:29 crc kubenswrapper[4978]: I0225 08:54:29.012586 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e3249bd-3d1b-41a2-9132-976e11161c91","Type":"ContainerStarted","Data":"a8ff703b621166a12c35dc9f2edb31949f40ebb523e02ac17c8943bef767c250"} Feb 25 08:54:29 crc kubenswrapper[4978]: I0225 08:54:29.012655 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e3249bd-3d1b-41a2-9132-976e11161c91","Type":"ContainerStarted","Data":"3ff03bf2346754e405adca9b348124baf1d29e3965eb3b6c01471f9269c387b8"} Feb 25 08:54:29 crc kubenswrapper[4978]: I0225 08:54:29.015825 4978 generic.go:334] "Generic (PLEG): container finished" podID="a86f5c5d-c083-48de-802a-14e448d14f86" containerID="f778e1e7fd9c03303a600b3f9c819c27f3da07697e4eef6b9ee347e1e708d3af" exitCode=0 Feb 25 08:54:29 crc kubenswrapper[4978]: I0225 08:54:29.016085 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9ssj7" event={"ID":"a86f5c5d-c083-48de-802a-14e448d14f86","Type":"ContainerDied","Data":"f778e1e7fd9c03303a600b3f9c819c27f3da07697e4eef6b9ee347e1e708d3af"} Feb 25 08:54:29 crc kubenswrapper[4978]: I0225 08:54:29.067799 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.067776811 podStartE2EDuration="17.067776811s" podCreationTimestamp="2026-02-25 08:54:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:54:29.05935581 +0000 UTC m=+7762.498612279" watchObservedRunningTime="2026-02-25 08:54:29.067776811 +0000 UTC m=+7762.507033270" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.361579 4978 scope.go:117] "RemoveContainer" containerID="2856f9525243e46948863bde61e1a085af07ea2bf627e25e0b1b1a6ee119a68d" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.392282 4978 scope.go:117] "RemoveContainer" containerID="662c76f3f813c122197536b2b51356488bce382ec99be2af22cab57a70a7432a" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.559590 4978 scope.go:117] "RemoveContainer" containerID="98927fb5e3c62da34936eccdd2ce919109bc411a5349a3650951c61f82284c69" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.583712 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9ssj7" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.590578 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-072e-account-create-update-g24l2" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.634392 4978 scope.go:117] "RemoveContainer" containerID="6cc0f8b5f7058efef4006d3e810c23c9b3121b26c989cf6949a622f33dfb40a7" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.687037 4978 scope.go:117] "RemoveContainer" containerID="d79740afa03266125c91e4c3d98e3e7f2d069644235f21b99695051739bc9195" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.697185 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gghl6\" (UniqueName: \"kubernetes.io/projected/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-kube-api-access-gghl6\") pod \"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9\" (UID: \"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9\") " Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.697243 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hghk8\" (UniqueName: \"kubernetes.io/projected/a86f5c5d-c083-48de-802a-14e448d14f86-kube-api-access-hghk8\") pod \"a86f5c5d-c083-48de-802a-14e448d14f86\" (UID: \"a86f5c5d-c083-48de-802a-14e448d14f86\") " Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.697411 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-operator-scripts\") pod \"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9\" (UID: \"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9\") " Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.697464 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a86f5c5d-c083-48de-802a-14e448d14f86-operator-scripts\") pod \"a86f5c5d-c083-48de-802a-14e448d14f86\" (UID: \"a86f5c5d-c083-48de-802a-14e448d14f86\") " Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.698385 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a86f5c5d-c083-48de-802a-14e448d14f86-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a86f5c5d-c083-48de-802a-14e448d14f86" (UID: "a86f5c5d-c083-48de-802a-14e448d14f86"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.698533 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9" (UID: "b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.703204 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a86f5c5d-c083-48de-802a-14e448d14f86-kube-api-access-hghk8" (OuterVolumeSpecName: "kube-api-access-hghk8") pod "a86f5c5d-c083-48de-802a-14e448d14f86" (UID: "a86f5c5d-c083-48de-802a-14e448d14f86"). InnerVolumeSpecName "kube-api-access-hghk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.703261 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-kube-api-access-gghl6" (OuterVolumeSpecName: "kube-api-access-gghl6") pod "b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9" (UID: "b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9"). InnerVolumeSpecName "kube-api-access-gghl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.710314 4978 scope.go:117] "RemoveContainer" containerID="6c8d45c68e3b713679a3be8b8b70d0e574a9c5c04b9e8ccf0e07b2c867369d79" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.732679 4978 scope.go:117] "RemoveContainer" containerID="597623758b07b4c8c035ef722b65ea7046000a79638c143d94ef7e1e89de4336" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.765985 4978 scope.go:117] "RemoveContainer" containerID="d222424a73352f3f33aeca1f9d171cf20976192c64c4991a9a728d355c919149" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.785840 4978 scope.go:117] "RemoveContainer" containerID="9e36a0cfa136b732c763f3b9118af15510d98e98d67ea2917b0b99b05ab61ded" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.800705 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gghl6\" (UniqueName: \"kubernetes.io/projected/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-kube-api-access-gghl6\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.800764 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hghk8\" (UniqueName: \"kubernetes.io/projected/a86f5c5d-c083-48de-802a-14e448d14f86-kube-api-access-hghk8\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.800795 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:30 crc kubenswrapper[4978]: I0225 08:54:30.800827 4978 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a86f5c5d-c083-48de-802a-14e448d14f86-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:31 crc kubenswrapper[4978]: I0225 08:54:31.039310 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-9ssj7" event={"ID":"a86f5c5d-c083-48de-802a-14e448d14f86","Type":"ContainerDied","Data":"a14d83c148deaf05471faab82926a664b56f9674fd94735964ab4e51632173fd"} Feb 25 08:54:31 crc kubenswrapper[4978]: I0225 08:54:31.039353 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a14d83c148deaf05471faab82926a664b56f9674fd94735964ab4e51632173fd" Feb 25 08:54:31 crc kubenswrapper[4978]: I0225 08:54:31.039408 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-9ssj7" Feb 25 08:54:31 crc kubenswrapper[4978]: I0225 08:54:31.041650 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-072e-account-create-update-g24l2" event={"ID":"b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9","Type":"ContainerDied","Data":"5db041f4429e112b77c8abea2a19fa363ff457580dabaa524882b26402a26ffc"} Feb 25 08:54:31 crc kubenswrapper[4978]: I0225 08:54:31.041686 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5db041f4429e112b77c8abea2a19fa363ff457580dabaa524882b26402a26ffc" Feb 25 08:54:31 crc kubenswrapper[4978]: I0225 08:54:31.041748 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-072e-account-create-update-g24l2" Feb 25 08:54:32 crc kubenswrapper[4978]: I0225 08:54:32.551027 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:33 crc kubenswrapper[4978]: I0225 08:54:33.092922 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-426w4"] Feb 25 08:54:33 crc kubenswrapper[4978]: I0225 08:54:33.104671 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-426w4"] Feb 25 08:54:33 crc kubenswrapper[4978]: I0225 08:54:33.343672 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09d826f1-229c-4a05-bfac-68f2c3f66c70" path="/var/lib/kubelet/pods/09d826f1-229c-4a05-bfac-68f2c3f66c70/volumes" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.299326 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-2jmrd"] Feb 25 08:54:36 crc kubenswrapper[4978]: E0225 08:54:36.300324 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9" containerName="mariadb-account-create-update" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.300348 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9" containerName="mariadb-account-create-update" Feb 25 08:54:36 crc kubenswrapper[4978]: E0225 08:54:36.300400 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86f5c5d-c083-48de-802a-14e448d14f86" containerName="mariadb-database-create" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.300412 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86f5c5d-c083-48de-802a-14e448d14f86" containerName="mariadb-database-create" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.300713 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a86f5c5d-c083-48de-802a-14e448d14f86" containerName="mariadb-database-create" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.300731 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9" containerName="mariadb-account-create-update" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.302290 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.310177 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.310220 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.310220 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-7bmkx" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.310309 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.317218 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-2jmrd"] Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.436176 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-combined-ca-bundle\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.436334 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-config-data\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.436427 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9mxw\" (UniqueName: \"kubernetes.io/projected/67524ebc-c6e6-4d7e-a175-e939f659a952-kube-api-access-g9mxw\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.436598 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-scripts\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.538268 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-scripts\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.538618 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-combined-ca-bundle\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.538735 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-config-data\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.538822 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9mxw\" (UniqueName: \"kubernetes.io/projected/67524ebc-c6e6-4d7e-a175-e939f659a952-kube-api-access-g9mxw\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.545401 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-scripts\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.548599 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-config-data\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.553158 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-combined-ca-bundle\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.560275 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9mxw\" (UniqueName: \"kubernetes.io/projected/67524ebc-c6e6-4d7e-a175-e939f659a952-kube-api-access-g9mxw\") pod \"aodh-db-sync-2jmrd\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:36 crc kubenswrapper[4978]: I0225 08:54:36.642788 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:37 crc kubenswrapper[4978]: I0225 08:54:37.114864 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-2jmrd"] Feb 25 08:54:37 crc kubenswrapper[4978]: I0225 08:54:37.142223 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2jmrd" event={"ID":"67524ebc-c6e6-4d7e-a175-e939f659a952","Type":"ContainerStarted","Data":"76a46e728cfee90985b84cf849ba206b95e335b365d2771774d09d6b95b6edec"} Feb 25 08:54:42 crc kubenswrapper[4978]: I0225 08:54:42.195849 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2jmrd" event={"ID":"67524ebc-c6e6-4d7e-a175-e939f659a952","Type":"ContainerStarted","Data":"66755eae1e385188325158cac2f0a047e582d3529a37909f220e5032796601ce"} Feb 25 08:54:42 crc kubenswrapper[4978]: I0225 08:54:42.214273 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-2jmrd" podStartSLOduration=1.500130522 podStartE2EDuration="6.21425319s" podCreationTimestamp="2026-02-25 08:54:36 +0000 UTC" firstStartedPulling="2026-02-25 08:54:37.116194138 +0000 UTC m=+7770.555450597" lastFinishedPulling="2026-02-25 08:54:41.830316786 +0000 UTC m=+7775.269573265" observedRunningTime="2026-02-25 08:54:42.212928929 +0000 UTC m=+7775.652185408" watchObservedRunningTime="2026-02-25 08:54:42.21425319 +0000 UTC m=+7775.653509659" Feb 25 08:54:42 crc kubenswrapper[4978]: I0225 08:54:42.550987 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:42 crc kubenswrapper[4978]: I0225 08:54:42.559460 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:43 crc kubenswrapper[4978]: I0225 08:54:43.223926 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Feb 25 08:54:43 crc kubenswrapper[4978]: I0225 08:54:43.686593 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 25 08:54:45 crc kubenswrapper[4978]: I0225 08:54:45.237347 4978 generic.go:334] "Generic (PLEG): container finished" podID="67524ebc-c6e6-4d7e-a175-e939f659a952" containerID="66755eae1e385188325158cac2f0a047e582d3529a37909f220e5032796601ce" exitCode=0 Feb 25 08:54:45 crc kubenswrapper[4978]: I0225 08:54:45.237411 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2jmrd" event={"ID":"67524ebc-c6e6-4d7e-a175-e939f659a952","Type":"ContainerDied","Data":"66755eae1e385188325158cac2f0a047e582d3529a37909f220e5032796601ce"} Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.548471 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.548690 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.548737 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.549392 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.549435 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" gracePeriod=600 Feb 25 08:54:46 crc kubenswrapper[4978]: E0225 08:54:46.678164 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.747944 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.851772 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9mxw\" (UniqueName: \"kubernetes.io/projected/67524ebc-c6e6-4d7e-a175-e939f659a952-kube-api-access-g9mxw\") pod \"67524ebc-c6e6-4d7e-a175-e939f659a952\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.851859 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-combined-ca-bundle\") pod \"67524ebc-c6e6-4d7e-a175-e939f659a952\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.852058 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-config-data\") pod \"67524ebc-c6e6-4d7e-a175-e939f659a952\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.852152 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-scripts\") pod \"67524ebc-c6e6-4d7e-a175-e939f659a952\" (UID: \"67524ebc-c6e6-4d7e-a175-e939f659a952\") " Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.856753 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-scripts" (OuterVolumeSpecName: "scripts") pod "67524ebc-c6e6-4d7e-a175-e939f659a952" (UID: "67524ebc-c6e6-4d7e-a175-e939f659a952"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.865646 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67524ebc-c6e6-4d7e-a175-e939f659a952-kube-api-access-g9mxw" (OuterVolumeSpecName: "kube-api-access-g9mxw") pod "67524ebc-c6e6-4d7e-a175-e939f659a952" (UID: "67524ebc-c6e6-4d7e-a175-e939f659a952"). InnerVolumeSpecName "kube-api-access-g9mxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.877529 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67524ebc-c6e6-4d7e-a175-e939f659a952" (UID: "67524ebc-c6e6-4d7e-a175-e939f659a952"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.884506 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-config-data" (OuterVolumeSpecName: "config-data") pod "67524ebc-c6e6-4d7e-a175-e939f659a952" (UID: "67524ebc-c6e6-4d7e-a175-e939f659a952"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.954312 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9mxw\" (UniqueName: \"kubernetes.io/projected/67524ebc-c6e6-4d7e-a175-e939f659a952-kube-api-access-g9mxw\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.954351 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.954362 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:46 crc kubenswrapper[4978]: I0225 08:54:46.954390 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67524ebc-c6e6-4d7e-a175-e939f659a952-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:47 crc kubenswrapper[4978]: I0225 08:54:47.266219 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" exitCode=0 Feb 25 08:54:47 crc kubenswrapper[4978]: I0225 08:54:47.266324 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595"} Feb 25 08:54:47 crc kubenswrapper[4978]: I0225 08:54:47.266437 4978 scope.go:117] "RemoveContainer" containerID="c572bfb7b460a06e6b3fd8562f5f892b5fdb169c89eb0df815351c3f11f58b78" Feb 25 08:54:47 crc kubenswrapper[4978]: I0225 08:54:47.267331 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:54:47 crc kubenswrapper[4978]: E0225 08:54:47.269200 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:54:47 crc kubenswrapper[4978]: I0225 08:54:47.273401 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2jmrd" event={"ID":"67524ebc-c6e6-4d7e-a175-e939f659a952","Type":"ContainerDied","Data":"76a46e728cfee90985b84cf849ba206b95e335b365d2771774d09d6b95b6edec"} Feb 25 08:54:47 crc kubenswrapper[4978]: I0225 08:54:47.273766 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76a46e728cfee90985b84cf849ba206b95e335b365d2771774d09d6b95b6edec" Feb 25 08:54:47 crc kubenswrapper[4978]: I0225 08:54:47.273427 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2jmrd" Feb 25 08:54:48 crc kubenswrapper[4978]: I0225 08:54:48.194358 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 08:54:48 crc kubenswrapper[4978]: I0225 08:54:48.194617 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="022f2aaa-0cfc-4441-9cdf-c8b36adad1ff" containerName="kube-state-metrics" containerID="cri-o://46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9" gracePeriod=30 Feb 25 08:54:48 crc kubenswrapper[4978]: I0225 08:54:48.699756 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 08:54:48 crc kubenswrapper[4978]: I0225 08:54:48.795209 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtltc\" (UniqueName: \"kubernetes.io/projected/022f2aaa-0cfc-4441-9cdf-c8b36adad1ff-kube-api-access-gtltc\") pod \"022f2aaa-0cfc-4441-9cdf-c8b36adad1ff\" (UID: \"022f2aaa-0cfc-4441-9cdf-c8b36adad1ff\") " Feb 25 08:54:48 crc kubenswrapper[4978]: I0225 08:54:48.800908 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/022f2aaa-0cfc-4441-9cdf-c8b36adad1ff-kube-api-access-gtltc" (OuterVolumeSpecName: "kube-api-access-gtltc") pod "022f2aaa-0cfc-4441-9cdf-c8b36adad1ff" (UID: "022f2aaa-0cfc-4441-9cdf-c8b36adad1ff"). InnerVolumeSpecName "kube-api-access-gtltc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:54:48 crc kubenswrapper[4978]: I0225 08:54:48.898451 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtltc\" (UniqueName: \"kubernetes.io/projected/022f2aaa-0cfc-4441-9cdf-c8b36adad1ff-kube-api-access-gtltc\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.297976 4978 generic.go:334] "Generic (PLEG): container finished" podID="022f2aaa-0cfc-4441-9cdf-c8b36adad1ff" containerID="46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9" exitCode=2 Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.298160 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"022f2aaa-0cfc-4441-9cdf-c8b36adad1ff","Type":"ContainerDied","Data":"46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9"} Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.298326 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"022f2aaa-0cfc-4441-9cdf-c8b36adad1ff","Type":"ContainerDied","Data":"f573bfea8a15fa5b587f87c7a9d52d0604220b550517adabba7c1bb00cd1ab47"} Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.298348 4978 scope.go:117] "RemoveContainer" containerID="46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.298237 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.332087 4978 scope.go:117] "RemoveContainer" containerID="46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9" Feb 25 08:54:49 crc kubenswrapper[4978]: E0225 08:54:49.334408 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9\": container with ID starting with 46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9 not found: ID does not exist" containerID="46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.334499 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9"} err="failed to get container status \"46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9\": rpc error: code = NotFound desc = could not find container \"46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9\": container with ID starting with 46a4bcc7fa59fb49ff0b8ce98fb09d6c22f5b6c0f933e1585218ec9d8583d5b9 not found: ID does not exist" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.379593 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.402759 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.414229 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 08:54:49 crc kubenswrapper[4978]: E0225 08:54:49.414728 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="022f2aaa-0cfc-4441-9cdf-c8b36adad1ff" containerName="kube-state-metrics" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.414747 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="022f2aaa-0cfc-4441-9cdf-c8b36adad1ff" containerName="kube-state-metrics" Feb 25 08:54:49 crc kubenswrapper[4978]: E0225 08:54:49.414776 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67524ebc-c6e6-4d7e-a175-e939f659a952" containerName="aodh-db-sync" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.414782 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="67524ebc-c6e6-4d7e-a175-e939f659a952" containerName="aodh-db-sync" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.414977 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="67524ebc-c6e6-4d7e-a175-e939f659a952" containerName="aodh-db-sync" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.415000 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="022f2aaa-0cfc-4441-9cdf-c8b36adad1ff" containerName="kube-state-metrics" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.415994 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.421343 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.421547 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.424041 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.509545 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvbdm\" (UniqueName: \"kubernetes.io/projected/902de2fe-e738-4665-a008-a9a9ed92c7c1-kube-api-access-cvbdm\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.509607 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902de2fe-e738-4665-a008-a9a9ed92c7c1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.509651 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/902de2fe-e738-4665-a008-a9a9ed92c7c1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.509858 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/902de2fe-e738-4665-a008-a9a9ed92c7c1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.612291 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/902de2fe-e738-4665-a008-a9a9ed92c7c1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.612424 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/902de2fe-e738-4665-a008-a9a9ed92c7c1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.612534 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvbdm\" (UniqueName: \"kubernetes.io/projected/902de2fe-e738-4665-a008-a9a9ed92c7c1-kube-api-access-cvbdm\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.612576 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902de2fe-e738-4665-a008-a9a9ed92c7c1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.627140 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/902de2fe-e738-4665-a008-a9a9ed92c7c1-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.627229 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902de2fe-e738-4665-a008-a9a9ed92c7c1-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.631673 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvbdm\" (UniqueName: \"kubernetes.io/projected/902de2fe-e738-4665-a008-a9a9ed92c7c1-kube-api-access-cvbdm\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.637330 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/902de2fe-e738-4665-a008-a9a9ed92c7c1-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"902de2fe-e738-4665-a008-a9a9ed92c7c1\") " pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.740586 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.988330 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.988917 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="proxy-httpd" containerID="cri-o://5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2" gracePeriod=30 Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.989001 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="sg-core" containerID="cri-o://4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27" gracePeriod=30 Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.989122 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="ceilometer-notification-agent" containerID="cri-o://88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a" gracePeriod=30 Feb 25 08:54:49 crc kubenswrapper[4978]: I0225 08:54:49.988881 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="ceilometer-central-agent" containerID="cri-o://9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47" gracePeriod=30 Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.189161 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.315063 4978 generic.go:334] "Generic (PLEG): container finished" podID="60eef5db-6892-460b-bda6-6037affe3fdf" containerID="5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2" exitCode=0 Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.315764 4978 generic.go:334] "Generic (PLEG): container finished" podID="60eef5db-6892-460b-bda6-6037affe3fdf" containerID="4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27" exitCode=2 Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.315265 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60eef5db-6892-460b-bda6-6037affe3fdf","Type":"ContainerDied","Data":"5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2"} Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.315850 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60eef5db-6892-460b-bda6-6037affe3fdf","Type":"ContainerDied","Data":"4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27"} Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.322032 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"902de2fe-e738-4665-a008-a9a9ed92c7c1","Type":"ContainerStarted","Data":"97edff3b93b214fcca334999fafcd3590cfc1b5701cb085a6bdfcaef5a7d2eb4"} Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.852442 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.855017 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.859070 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.859311 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-7bmkx" Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.869164 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.873227 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.940852 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.940962 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tblqx\" (UniqueName: \"kubernetes.io/projected/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-kube-api-access-tblqx\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.941036 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-config-data\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:50 crc kubenswrapper[4978]: I0225 08:54:50.941079 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-scripts\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.043155 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tblqx\" (UniqueName: \"kubernetes.io/projected/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-kube-api-access-tblqx\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.043253 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-config-data\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.043302 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-scripts\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.043329 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.047974 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-config-data\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.049517 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-scripts\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.059085 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tblqx\" (UniqueName: \"kubernetes.io/projected/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-kube-api-access-tblqx\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.060507 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " pod="openstack/aodh-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.186666 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.320832 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.350582 4978 generic.go:334] "Generic (PLEG): container finished" podID="60eef5db-6892-460b-bda6-6037affe3fdf" containerID="88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a" exitCode=0 Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.350617 4978 generic.go:334] "Generic (PLEG): container finished" podID="60eef5db-6892-460b-bda6-6037affe3fdf" containerID="9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47" exitCode=0 Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.350733 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.352406 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="022f2aaa-0cfc-4441-9cdf-c8b36adad1ff" path="/var/lib/kubelet/pods/022f2aaa-0cfc-4441-9cdf-c8b36adad1ff/volumes" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.366108 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60eef5db-6892-460b-bda6-6037affe3fdf","Type":"ContainerDied","Data":"88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a"} Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.366138 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60eef5db-6892-460b-bda6-6037affe3fdf","Type":"ContainerDied","Data":"9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47"} Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.366155 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"60eef5db-6892-460b-bda6-6037affe3fdf","Type":"ContainerDied","Data":"125da5e39ed952073aa4a3acb3d63d657de81ae6de51b8cb7f8e940ec61908a7"} Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.366173 4978 scope.go:117] "RemoveContainer" containerID="5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.382378 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"902de2fe-e738-4665-a008-a9a9ed92c7c1","Type":"ContainerStarted","Data":"93b1db0c90a7dd6a1c065a3d00ee8093103d90951c6f28f39aa3b05b2f6fdf82"} Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.384589 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.411892 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.015815009 podStartE2EDuration="2.41187623s" podCreationTimestamp="2026-02-25 08:54:49 +0000 UTC" firstStartedPulling="2026-02-25 08:54:50.1888871 +0000 UTC m=+7783.628143559" lastFinishedPulling="2026-02-25 08:54:50.584948321 +0000 UTC m=+7784.024204780" observedRunningTime="2026-02-25 08:54:51.403348976 +0000 UTC m=+7784.842605435" watchObservedRunningTime="2026-02-25 08:54:51.41187623 +0000 UTC m=+7784.851132679" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.431576 4978 scope.go:117] "RemoveContainer" containerID="4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.450322 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-run-httpd\") pod \"60eef5db-6892-460b-bda6-6037affe3fdf\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.450427 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-sg-core-conf-yaml\") pod \"60eef5db-6892-460b-bda6-6037affe3fdf\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.450507 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-config-data\") pod \"60eef5db-6892-460b-bda6-6037affe3fdf\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.450648 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-log-httpd\") pod \"60eef5db-6892-460b-bda6-6037affe3fdf\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.450684 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlcwg\" (UniqueName: \"kubernetes.io/projected/60eef5db-6892-460b-bda6-6037affe3fdf-kube-api-access-dlcwg\") pod \"60eef5db-6892-460b-bda6-6037affe3fdf\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.450728 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-scripts\") pod \"60eef5db-6892-460b-bda6-6037affe3fdf\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.450723 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "60eef5db-6892-460b-bda6-6037affe3fdf" (UID: "60eef5db-6892-460b-bda6-6037affe3fdf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.450772 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-combined-ca-bundle\") pod \"60eef5db-6892-460b-bda6-6037affe3fdf\" (UID: \"60eef5db-6892-460b-bda6-6037affe3fdf\") " Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.451497 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.453173 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "60eef5db-6892-460b-bda6-6037affe3fdf" (UID: "60eef5db-6892-460b-bda6-6037affe3fdf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.460939 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60eef5db-6892-460b-bda6-6037affe3fdf-kube-api-access-dlcwg" (OuterVolumeSpecName: "kube-api-access-dlcwg") pod "60eef5db-6892-460b-bda6-6037affe3fdf" (UID: "60eef5db-6892-460b-bda6-6037affe3fdf"). InnerVolumeSpecName "kube-api-access-dlcwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.462607 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-scripts" (OuterVolumeSpecName: "scripts") pod "60eef5db-6892-460b-bda6-6037affe3fdf" (UID: "60eef5db-6892-460b-bda6-6037affe3fdf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.473858 4978 scope.go:117] "RemoveContainer" containerID="88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.482135 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "60eef5db-6892-460b-bda6-6037affe3fdf" (UID: "60eef5db-6892-460b-bda6-6037affe3fdf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.500693 4978 scope.go:117] "RemoveContainer" containerID="9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.552790 4978 scope.go:117] "RemoveContainer" containerID="5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2" Feb 25 08:54:51 crc kubenswrapper[4978]: E0225 08:54:51.566414 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2\": container with ID starting with 5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2 not found: ID does not exist" containerID="5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.566449 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2"} err="failed to get container status \"5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2\": rpc error: code = NotFound desc = could not find container \"5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2\": container with ID starting with 5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2 not found: ID does not exist" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.566474 4978 scope.go:117] "RemoveContainer" containerID="4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27" Feb 25 08:54:51 crc kubenswrapper[4978]: E0225 08:54:51.566821 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27\": container with ID starting with 4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27 not found: ID does not exist" containerID="4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.566851 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27"} err="failed to get container status \"4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27\": rpc error: code = NotFound desc = could not find container \"4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27\": container with ID starting with 4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27 not found: ID does not exist" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.566878 4978 scope.go:117] "RemoveContainer" containerID="88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.567065 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/60eef5db-6892-460b-bda6-6037affe3fdf-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.567089 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlcwg\" (UniqueName: \"kubernetes.io/projected/60eef5db-6892-460b-bda6-6037affe3fdf-kube-api-access-dlcwg\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.567098 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.567107 4978 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:51 crc kubenswrapper[4978]: E0225 08:54:51.567415 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a\": container with ID starting with 88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a not found: ID does not exist" containerID="88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.567433 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a"} err="failed to get container status \"88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a\": rpc error: code = NotFound desc = could not find container \"88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a\": container with ID starting with 88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a not found: ID does not exist" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.567447 4978 scope.go:117] "RemoveContainer" containerID="9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47" Feb 25 08:54:51 crc kubenswrapper[4978]: E0225 08:54:51.568190 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47\": container with ID starting with 9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47 not found: ID does not exist" containerID="9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.568216 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47"} err="failed to get container status \"9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47\": rpc error: code = NotFound desc = could not find container \"9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47\": container with ID starting with 9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47 not found: ID does not exist" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.568229 4978 scope.go:117] "RemoveContainer" containerID="5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.568483 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2"} err="failed to get container status \"5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2\": rpc error: code = NotFound desc = could not find container \"5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2\": container with ID starting with 5c15a1a671970cf752661fd224f966f11bc97f9f9c075bb73e33648b0e8669b2 not found: ID does not exist" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.568502 4978 scope.go:117] "RemoveContainer" containerID="4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.568816 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27"} err="failed to get container status \"4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27\": rpc error: code = NotFound desc = could not find container \"4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27\": container with ID starting with 4abf24740f526bae7c1c657b23c010ea1d80fc725402b860687a38c2bb744a27 not found: ID does not exist" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.568856 4978 scope.go:117] "RemoveContainer" containerID="88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.569206 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a"} err="failed to get container status \"88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a\": rpc error: code = NotFound desc = could not find container \"88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a\": container with ID starting with 88e02e624091221365a6217e510f26c75e786c80a3dd26689efc58c28103b85a not found: ID does not exist" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.569229 4978 scope.go:117] "RemoveContainer" containerID="9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.572672 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47"} err="failed to get container status \"9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47\": rpc error: code = NotFound desc = could not find container \"9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47\": container with ID starting with 9593041167dee44001d2655670393cf4cdb2dac34a665333672aab8d1929bc47 not found: ID does not exist" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.581175 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60eef5db-6892-460b-bda6-6037affe3fdf" (UID: "60eef5db-6892-460b-bda6-6037affe3fdf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.606359 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-config-data" (OuterVolumeSpecName: "config-data") pod "60eef5db-6892-460b-bda6-6037affe3fdf" (UID: "60eef5db-6892-460b-bda6-6037affe3fdf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.668452 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.668477 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60eef5db-6892-460b-bda6-6037affe3fdf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.732052 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.741052 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:54:51 crc kubenswrapper[4978]: W0225 08:54:51.755112 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1e9a596_1d6c_4f92_9867_c8a4f644f2c5.slice/crio-71c418bc015e506bf42a350238bac3e8cb0c95d90f69960e4416a96560dc7803 WatchSource:0}: Error finding container 71c418bc015e506bf42a350238bac3e8cb0c95d90f69960e4416a96560dc7803: Status 404 returned error can't find the container with id 71c418bc015e506bf42a350238bac3e8cb0c95d90f69960e4416a96560dc7803 Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.757972 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:54:51 crc kubenswrapper[4978]: E0225 08:54:51.758317 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="ceilometer-notification-agent" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.758334 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="ceilometer-notification-agent" Feb 25 08:54:51 crc kubenswrapper[4978]: E0225 08:54:51.758347 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="sg-core" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.758354 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="sg-core" Feb 25 08:54:51 crc kubenswrapper[4978]: E0225 08:54:51.758382 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="ceilometer-central-agent" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.758389 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="ceilometer-central-agent" Feb 25 08:54:51 crc kubenswrapper[4978]: E0225 08:54:51.758400 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="proxy-httpd" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.758405 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="proxy-httpd" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.758605 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="ceilometer-central-agent" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.758621 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="proxy-httpd" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.758632 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="sg-core" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.758643 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" containerName="ceilometer-notification-agent" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.760151 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.762507 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.762689 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.762799 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.769889 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.780029 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.872135 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.872190 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dzsk\" (UniqueName: \"kubernetes.io/projected/398e1e22-b429-4ab8-928a-4ac595b4830c-kube-api-access-8dzsk\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.872342 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.872429 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.872528 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-log-httpd\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.872622 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-scripts\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.872695 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-config-data\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.872923 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-run-httpd\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.975133 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-scripts\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.975201 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-config-data\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.975305 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-run-httpd\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.975397 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.975422 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dzsk\" (UniqueName: \"kubernetes.io/projected/398e1e22-b429-4ab8-928a-4ac595b4830c-kube-api-access-8dzsk\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.975470 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.975507 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.975545 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-log-httpd\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.976058 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-log-httpd\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.977685 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-run-httpd\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.980212 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-scripts\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.981125 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-config-data\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.983879 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.984140 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:51 crc kubenswrapper[4978]: I0225 08:54:51.985118 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:52 crc kubenswrapper[4978]: I0225 08:54:52.011687 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dzsk\" (UniqueName: \"kubernetes.io/projected/398e1e22-b429-4ab8-928a-4ac595b4830c-kube-api-access-8dzsk\") pod \"ceilometer-0\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " pod="openstack/ceilometer-0" Feb 25 08:54:52 crc kubenswrapper[4978]: I0225 08:54:52.107765 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 08:54:52 crc kubenswrapper[4978]: I0225 08:54:52.416017 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5","Type":"ContainerStarted","Data":"15e1faa4ea8831012f680f41a64d9a7afa74764825e1e9777e66ff9dbb0d9409"} Feb 25 08:54:52 crc kubenswrapper[4978]: I0225 08:54:52.416335 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5","Type":"ContainerStarted","Data":"71c418bc015e506bf42a350238bac3e8cb0c95d90f69960e4416a96560dc7803"} Feb 25 08:54:52 crc kubenswrapper[4978]: I0225 08:54:52.747970 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:54:52 crc kubenswrapper[4978]: W0225 08:54:52.768474 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod398e1e22_b429_4ab8_928a_4ac595b4830c.slice/crio-ab931798d9118a78eccb57d9684ec7bae14b0bc0f21b72d2257172f54ef42366 WatchSource:0}: Error finding container ab931798d9118a78eccb57d9684ec7bae14b0bc0f21b72d2257172f54ef42366: Status 404 returned error can't find the container with id ab931798d9118a78eccb57d9684ec7bae14b0bc0f21b72d2257172f54ef42366 Feb 25 08:54:53 crc kubenswrapper[4978]: I0225 08:54:53.344042 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60eef5db-6892-460b-bda6-6037affe3fdf" path="/var/lib/kubelet/pods/60eef5db-6892-460b-bda6-6037affe3fdf/volumes" Feb 25 08:54:53 crc kubenswrapper[4978]: I0225 08:54:53.449993 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398e1e22-b429-4ab8-928a-4ac595b4830c","Type":"ContainerStarted","Data":"ab931798d9118a78eccb57d9684ec7bae14b0bc0f21b72d2257172f54ef42366"} Feb 25 08:54:54 crc kubenswrapper[4978]: I0225 08:54:54.449690 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Feb 25 08:54:54 crc kubenswrapper[4978]: I0225 08:54:54.459261 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5","Type":"ContainerStarted","Data":"c06d036c4537043204215edfd958423d057a4d0c5d8be8199ab455f29c014677"} Feb 25 08:54:54 crc kubenswrapper[4978]: I0225 08:54:54.463786 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398e1e22-b429-4ab8-928a-4ac595b4830c","Type":"ContainerStarted","Data":"37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3"} Feb 25 08:54:54 crc kubenswrapper[4978]: I0225 08:54:54.463826 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398e1e22-b429-4ab8-928a-4ac595b4830c","Type":"ContainerStarted","Data":"9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d"} Feb 25 08:54:55 crc kubenswrapper[4978]: I0225 08:54:55.010505 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:54:55 crc kubenswrapper[4978]: I0225 08:54:55.496006 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398e1e22-b429-4ab8-928a-4ac595b4830c","Type":"ContainerStarted","Data":"be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a"} Feb 25 08:54:55 crc kubenswrapper[4978]: I0225 08:54:55.513991 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5","Type":"ContainerStarted","Data":"c4de36a782c0753857e4120b568d80b9f8c0597f57044467377c109bd4e5a34e"} Feb 25 08:54:57 crc kubenswrapper[4978]: I0225 08:54:57.538419 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5","Type":"ContainerStarted","Data":"9d71cdfa3942fbf42fbefd4d49580e727d46616fc58658c7fd8050e718cd61d5"} Feb 25 08:54:57 crc kubenswrapper[4978]: I0225 08:54:57.538819 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-api" containerID="cri-o://15e1faa4ea8831012f680f41a64d9a7afa74764825e1e9777e66ff9dbb0d9409" gracePeriod=30 Feb 25 08:54:57 crc kubenswrapper[4978]: I0225 08:54:57.539420 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-listener" containerID="cri-o://9d71cdfa3942fbf42fbefd4d49580e727d46616fc58658c7fd8050e718cd61d5" gracePeriod=30 Feb 25 08:54:57 crc kubenswrapper[4978]: I0225 08:54:57.539483 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-notifier" containerID="cri-o://c4de36a782c0753857e4120b568d80b9f8c0597f57044467377c109bd4e5a34e" gracePeriod=30 Feb 25 08:54:57 crc kubenswrapper[4978]: I0225 08:54:57.539527 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-evaluator" containerID="cri-o://c06d036c4537043204215edfd958423d057a4d0c5d8be8199ab455f29c014677" gracePeriod=30 Feb 25 08:54:57 crc kubenswrapper[4978]: I0225 08:54:57.574091 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.982359394 podStartE2EDuration="7.57407251s" podCreationTimestamp="2026-02-25 08:54:50 +0000 UTC" firstStartedPulling="2026-02-25 08:54:51.757621001 +0000 UTC m=+7785.196877450" lastFinishedPulling="2026-02-25 08:54:56.349334107 +0000 UTC m=+7789.788590566" observedRunningTime="2026-02-25 08:54:57.56373949 +0000 UTC m=+7791.002995959" watchObservedRunningTime="2026-02-25 08:54:57.57407251 +0000 UTC m=+7791.013328969" Feb 25 08:54:58 crc kubenswrapper[4978]: I0225 08:54:58.560052 4978 generic.go:334] "Generic (PLEG): container finished" podID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerID="c4de36a782c0753857e4120b568d80b9f8c0597f57044467377c109bd4e5a34e" exitCode=0 Feb 25 08:54:58 crc kubenswrapper[4978]: I0225 08:54:58.560253 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5","Type":"ContainerDied","Data":"c4de36a782c0753857e4120b568d80b9f8c0597f57044467377c109bd4e5a34e"} Feb 25 08:54:58 crc kubenswrapper[4978]: I0225 08:54:58.560282 4978 generic.go:334] "Generic (PLEG): container finished" podID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerID="c06d036c4537043204215edfd958423d057a4d0c5d8be8199ab455f29c014677" exitCode=0 Feb 25 08:54:58 crc kubenswrapper[4978]: I0225 08:54:58.560293 4978 generic.go:334] "Generic (PLEG): container finished" podID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerID="15e1faa4ea8831012f680f41a64d9a7afa74764825e1e9777e66ff9dbb0d9409" exitCode=0 Feb 25 08:54:58 crc kubenswrapper[4978]: I0225 08:54:58.560309 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5","Type":"ContainerDied","Data":"c06d036c4537043204215edfd958423d057a4d0c5d8be8199ab455f29c014677"} Feb 25 08:54:58 crc kubenswrapper[4978]: I0225 08:54:58.560319 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5","Type":"ContainerDied","Data":"15e1faa4ea8831012f680f41a64d9a7afa74764825e1e9777e66ff9dbb0d9409"} Feb 25 08:54:59 crc kubenswrapper[4978]: I0225 08:54:59.574782 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398e1e22-b429-4ab8-928a-4ac595b4830c","Type":"ContainerStarted","Data":"f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8"} Feb 25 08:54:59 crc kubenswrapper[4978]: I0225 08:54:59.574979 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="ceilometer-central-agent" containerID="cri-o://9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d" gracePeriod=30 Feb 25 08:54:59 crc kubenswrapper[4978]: I0225 08:54:59.575086 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="ceilometer-notification-agent" containerID="cri-o://37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3" gracePeriod=30 Feb 25 08:54:59 crc kubenswrapper[4978]: I0225 08:54:59.575033 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="sg-core" containerID="cri-o://be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a" gracePeriod=30 Feb 25 08:54:59 crc kubenswrapper[4978]: I0225 08:54:59.575033 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="proxy-httpd" containerID="cri-o://f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8" gracePeriod=30 Feb 25 08:54:59 crc kubenswrapper[4978]: I0225 08:54:59.575296 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 08:54:59 crc kubenswrapper[4978]: I0225 08:54:59.611271 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.56580694 podStartE2EDuration="8.611254184s" podCreationTimestamp="2026-02-25 08:54:51 +0000 UTC" firstStartedPulling="2026-02-25 08:54:52.776985262 +0000 UTC m=+7786.216241721" lastFinishedPulling="2026-02-25 08:54:58.822432506 +0000 UTC m=+7792.261688965" observedRunningTime="2026-02-25 08:54:59.609191041 +0000 UTC m=+7793.048447510" watchObservedRunningTime="2026-02-25 08:54:59.611254184 +0000 UTC m=+7793.050510643" Feb 25 08:54:59 crc kubenswrapper[4978]: I0225 08:54:59.761822 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 25 08:55:00 crc kubenswrapper[4978]: I0225 08:55:00.592018 4978 generic.go:334] "Generic (PLEG): container finished" podID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerID="f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8" exitCode=0 Feb 25 08:55:00 crc kubenswrapper[4978]: I0225 08:55:00.592052 4978 generic.go:334] "Generic (PLEG): container finished" podID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerID="be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a" exitCode=2 Feb 25 08:55:00 crc kubenswrapper[4978]: I0225 08:55:00.592080 4978 generic.go:334] "Generic (PLEG): container finished" podID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerID="37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3" exitCode=0 Feb 25 08:55:00 crc kubenswrapper[4978]: I0225 08:55:00.592101 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398e1e22-b429-4ab8-928a-4ac595b4830c","Type":"ContainerDied","Data":"f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8"} Feb 25 08:55:00 crc kubenswrapper[4978]: I0225 08:55:00.592126 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398e1e22-b429-4ab8-928a-4ac595b4830c","Type":"ContainerDied","Data":"be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a"} Feb 25 08:55:00 crc kubenswrapper[4978]: I0225 08:55:00.592158 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398e1e22-b429-4ab8-928a-4ac595b4830c","Type":"ContainerDied","Data":"37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3"} Feb 25 08:55:01 crc kubenswrapper[4978]: I0225 08:55:01.327547 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:55:01 crc kubenswrapper[4978]: E0225 08:55:01.328287 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.619011 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.622109 4978 generic.go:334] "Generic (PLEG): container finished" podID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerID="9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d" exitCode=0 Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.622206 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398e1e22-b429-4ab8-928a-4ac595b4830c","Type":"ContainerDied","Data":"9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d"} Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.622501 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"398e1e22-b429-4ab8-928a-4ac595b4830c","Type":"ContainerDied","Data":"ab931798d9118a78eccb57d9684ec7bae14b0bc0f21b72d2257172f54ef42366"} Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.622610 4978 scope.go:117] "RemoveContainer" containerID="f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.662900 4978 scope.go:117] "RemoveContainer" containerID="be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.687682 4978 scope.go:117] "RemoveContainer" containerID="37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.724487 4978 scope.go:117] "RemoveContainer" containerID="9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.743797 4978 scope.go:117] "RemoveContainer" containerID="f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8" Feb 25 08:55:03 crc kubenswrapper[4978]: E0225 08:55:03.744322 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8\": container with ID starting with f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8 not found: ID does not exist" containerID="f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.744394 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8"} err="failed to get container status \"f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8\": rpc error: code = NotFound desc = could not find container \"f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8\": container with ID starting with f6b6584993260e4c407c5c12ad3dd7246b2a3c0890a8d186a59b0bd755488da8 not found: ID does not exist" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.744421 4978 scope.go:117] "RemoveContainer" containerID="be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a" Feb 25 08:55:03 crc kubenswrapper[4978]: E0225 08:55:03.744802 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a\": container with ID starting with be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a not found: ID does not exist" containerID="be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.744835 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a"} err="failed to get container status \"be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a\": rpc error: code = NotFound desc = could not find container \"be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a\": container with ID starting with be7419c544ee37db0821bea689e1117d4968fcdf079f52dce2cc0dabebc0203a not found: ID does not exist" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.744858 4978 scope.go:117] "RemoveContainer" containerID="37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3" Feb 25 08:55:03 crc kubenswrapper[4978]: E0225 08:55:03.745234 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3\": container with ID starting with 37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3 not found: ID does not exist" containerID="37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.745256 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3"} err="failed to get container status \"37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3\": rpc error: code = NotFound desc = could not find container \"37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3\": container with ID starting with 37550e46bcc32262aebca0c153c1dddcaf2b1af9e49ada9a9d73a122e7068fc3 not found: ID does not exist" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.745268 4978 scope.go:117] "RemoveContainer" containerID="9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d" Feb 25 08:55:03 crc kubenswrapper[4978]: E0225 08:55:03.745636 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d\": container with ID starting with 9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d not found: ID does not exist" containerID="9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.745685 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d"} err="failed to get container status \"9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d\": rpc error: code = NotFound desc = could not find container \"9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d\": container with ID starting with 9a7a17630c72a61239f341ec6065ec71689fd80438dfd5dc5f00a87ae99c5e3d not found: ID does not exist" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.769633 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-config-data\") pod \"398e1e22-b429-4ab8-928a-4ac595b4830c\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.769720 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-scripts\") pod \"398e1e22-b429-4ab8-928a-4ac595b4830c\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.769785 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-sg-core-conf-yaml\") pod \"398e1e22-b429-4ab8-928a-4ac595b4830c\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.769818 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dzsk\" (UniqueName: \"kubernetes.io/projected/398e1e22-b429-4ab8-928a-4ac595b4830c-kube-api-access-8dzsk\") pod \"398e1e22-b429-4ab8-928a-4ac595b4830c\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.769866 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-log-httpd\") pod \"398e1e22-b429-4ab8-928a-4ac595b4830c\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.769942 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-combined-ca-bundle\") pod \"398e1e22-b429-4ab8-928a-4ac595b4830c\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.769976 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-ceilometer-tls-certs\") pod \"398e1e22-b429-4ab8-928a-4ac595b4830c\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.769998 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-run-httpd\") pod \"398e1e22-b429-4ab8-928a-4ac595b4830c\" (UID: \"398e1e22-b429-4ab8-928a-4ac595b4830c\") " Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.770704 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "398e1e22-b429-4ab8-928a-4ac595b4830c" (UID: "398e1e22-b429-4ab8-928a-4ac595b4830c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.770723 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "398e1e22-b429-4ab8-928a-4ac595b4830c" (UID: "398e1e22-b429-4ab8-928a-4ac595b4830c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.771596 4978 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.771659 4978 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/398e1e22-b429-4ab8-928a-4ac595b4830c-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.775304 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/398e1e22-b429-4ab8-928a-4ac595b4830c-kube-api-access-8dzsk" (OuterVolumeSpecName: "kube-api-access-8dzsk") pod "398e1e22-b429-4ab8-928a-4ac595b4830c" (UID: "398e1e22-b429-4ab8-928a-4ac595b4830c"). InnerVolumeSpecName "kube-api-access-8dzsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.784205 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-scripts" (OuterVolumeSpecName: "scripts") pod "398e1e22-b429-4ab8-928a-4ac595b4830c" (UID: "398e1e22-b429-4ab8-928a-4ac595b4830c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.818585 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "398e1e22-b429-4ab8-928a-4ac595b4830c" (UID: "398e1e22-b429-4ab8-928a-4ac595b4830c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.870617 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "398e1e22-b429-4ab8-928a-4ac595b4830c" (UID: "398e1e22-b429-4ab8-928a-4ac595b4830c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.871713 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "398e1e22-b429-4ab8-928a-4ac595b4830c" (UID: "398e1e22-b429-4ab8-928a-4ac595b4830c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.873062 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.873082 4978 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.873092 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dzsk\" (UniqueName: \"kubernetes.io/projected/398e1e22-b429-4ab8-928a-4ac595b4830c-kube-api-access-8dzsk\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.873101 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.873111 4978 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.912241 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-config-data" (OuterVolumeSpecName: "config-data") pod "398e1e22-b429-4ab8-928a-4ac595b4830c" (UID: "398e1e22-b429-4ab8-928a-4ac595b4830c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:55:03 crc kubenswrapper[4978]: I0225 08:55:03.975096 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/398e1e22-b429-4ab8-928a-4ac595b4830c-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.635656 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.691304 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.707181 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.735175 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:55:04 crc kubenswrapper[4978]: E0225 08:55:04.735820 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="sg-core" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.735856 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="sg-core" Feb 25 08:55:04 crc kubenswrapper[4978]: E0225 08:55:04.735898 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="ceilometer-notification-agent" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.735912 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="ceilometer-notification-agent" Feb 25 08:55:04 crc kubenswrapper[4978]: E0225 08:55:04.735951 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="ceilometer-central-agent" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.735963 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="ceilometer-central-agent" Feb 25 08:55:04 crc kubenswrapper[4978]: E0225 08:55:04.736001 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="proxy-httpd" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.736012 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="proxy-httpd" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.736301 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="sg-core" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.736332 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="ceilometer-central-agent" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.736361 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="ceilometer-notification-agent" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.737249 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" containerName="proxy-httpd" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.740181 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.744408 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.744722 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.744976 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.752490 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.898073 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41cbe834-6422-4dc3-b32f-f97042673369-log-httpd\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.898200 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-scripts\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.898263 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fqh8\" (UniqueName: \"kubernetes.io/projected/41cbe834-6422-4dc3-b32f-f97042673369-kube-api-access-5fqh8\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.898321 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-config-data\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.898469 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.898499 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.898568 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:04 crc kubenswrapper[4978]: I0225 08:55:04.898646 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41cbe834-6422-4dc3-b32f-f97042673369-run-httpd\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.000765 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.000807 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.000836 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.000889 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41cbe834-6422-4dc3-b32f-f97042673369-run-httpd\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.000923 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41cbe834-6422-4dc3-b32f-f97042673369-log-httpd\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.000985 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-scripts\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.001022 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fqh8\" (UniqueName: \"kubernetes.io/projected/41cbe834-6422-4dc3-b32f-f97042673369-kube-api-access-5fqh8\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.001060 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-config-data\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.001653 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41cbe834-6422-4dc3-b32f-f97042673369-run-httpd\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.001664 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41cbe834-6422-4dc3-b32f-f97042673369-log-httpd\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.007082 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.007431 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.007984 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.009482 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-scripts\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.022892 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41cbe834-6422-4dc3-b32f-f97042673369-config-data\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.031540 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fqh8\" (UniqueName: \"kubernetes.io/projected/41cbe834-6422-4dc3-b32f-f97042673369-kube-api-access-5fqh8\") pod \"ceilometer-0\" (UID: \"41cbe834-6422-4dc3-b32f-f97042673369\") " pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.097851 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.341445 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="398e1e22-b429-4ab8-928a-4ac595b4830c" path="/var/lib/kubelet/pods/398e1e22-b429-4ab8-928a-4ac595b4830c/volumes" Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.629117 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.632496 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 25 08:55:05 crc kubenswrapper[4978]: I0225 08:55:05.649206 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41cbe834-6422-4dc3-b32f-f97042673369","Type":"ContainerStarted","Data":"e538bd04f5f1e9921bc923c96846b5c9f88d3d299f1d836a8b62c290a6aa3ae2"} Feb 25 08:55:06 crc kubenswrapper[4978]: I0225 08:55:06.678009 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41cbe834-6422-4dc3-b32f-f97042673369","Type":"ContainerStarted","Data":"1b25085093794dcfbe13ccb7f651c2f3a07b8833eff4911ce294472e72bb597c"} Feb 25 08:55:06 crc kubenswrapper[4978]: I0225 08:55:06.678348 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41cbe834-6422-4dc3-b32f-f97042673369","Type":"ContainerStarted","Data":"f21336c1eb3b30b0249ccc8dc2419287382e0dfc554a530722bbbb52aa50b489"} Feb 25 08:55:07 crc kubenswrapper[4978]: I0225 08:55:07.692648 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41cbe834-6422-4dc3-b32f-f97042673369","Type":"ContainerStarted","Data":"9cec705fcf4c6444538b8cbff3bb7701d87105bad18b2017e2211cd29b730c38"} Feb 25 08:55:09 crc kubenswrapper[4978]: I0225 08:55:09.716000 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41cbe834-6422-4dc3-b32f-f97042673369","Type":"ContainerStarted","Data":"f4d155da66c0d111b9d9e9b7520efe293471a6c6dda31edb4f34149f0a64bedb"} Feb 25 08:55:09 crc kubenswrapper[4978]: I0225 08:55:09.718014 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 25 08:55:09 crc kubenswrapper[4978]: I0225 08:55:09.743901 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.133101306 podStartE2EDuration="5.743882882s" podCreationTimestamp="2026-02-25 08:55:04 +0000 UTC" firstStartedPulling="2026-02-25 08:55:05.628837264 +0000 UTC m=+7799.068093723" lastFinishedPulling="2026-02-25 08:55:09.2396188 +0000 UTC m=+7802.678875299" observedRunningTime="2026-02-25 08:55:09.742361925 +0000 UTC m=+7803.181618384" watchObservedRunningTime="2026-02-25 08:55:09.743882882 +0000 UTC m=+7803.183139341" Feb 25 08:55:15 crc kubenswrapper[4978]: I0225 08:55:15.330565 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:55:15 crc kubenswrapper[4978]: E0225 08:55:15.331740 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:55:27 crc kubenswrapper[4978]: E0225 08:55:27.857149 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1e9a596_1d6c_4f92_9867_c8a4f644f2c5.slice/crio-conmon-9d71cdfa3942fbf42fbefd4d49580e727d46616fc58658c7fd8050e718cd61d5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1e9a596_1d6c_4f92_9867_c8a4f644f2c5.slice/crio-9d71cdfa3942fbf42fbefd4d49580e727d46616fc58658c7fd8050e718cd61d5.scope\": RecentStats: unable to find data in memory cache]" Feb 25 08:55:27 crc kubenswrapper[4978]: I0225 08:55:27.972307 4978 generic.go:334] "Generic (PLEG): container finished" podID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerID="9d71cdfa3942fbf42fbefd4d49580e727d46616fc58658c7fd8050e718cd61d5" exitCode=137 Feb 25 08:55:27 crc kubenswrapper[4978]: I0225 08:55:27.972452 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5","Type":"ContainerDied","Data":"9d71cdfa3942fbf42fbefd4d49580e727d46616fc58658c7fd8050e718cd61d5"} Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.328073 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:55:28 crc kubenswrapper[4978]: E0225 08:55:28.329033 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.534637 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.554525 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-scripts\") pod \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.554836 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-combined-ca-bundle\") pod \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.555044 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tblqx\" (UniqueName: \"kubernetes.io/projected/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-kube-api-access-tblqx\") pod \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.555453 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-config-data\") pod \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\" (UID: \"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5\") " Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.605599 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-scripts" (OuterVolumeSpecName: "scripts") pod "c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" (UID: "c1e9a596-1d6c-4f92-9867-c8a4f644f2c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.605804 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-kube-api-access-tblqx" (OuterVolumeSpecName: "kube-api-access-tblqx") pod "c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" (UID: "c1e9a596-1d6c-4f92-9867-c8a4f644f2c5"). InnerVolumeSpecName "kube-api-access-tblqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.657584 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.657633 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tblqx\" (UniqueName: \"kubernetes.io/projected/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-kube-api-access-tblqx\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.717234 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" (UID: "c1e9a596-1d6c-4f92-9867-c8a4f644f2c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.727479 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-config-data" (OuterVolumeSpecName: "config-data") pod "c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" (UID: "c1e9a596-1d6c-4f92-9867-c8a4f644f2c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.758277 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.758512 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.988540 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c1e9a596-1d6c-4f92-9867-c8a4f644f2c5","Type":"ContainerDied","Data":"71c418bc015e506bf42a350238bac3e8cb0c95d90f69960e4416a96560dc7803"} Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.988625 4978 scope.go:117] "RemoveContainer" containerID="9d71cdfa3942fbf42fbefd4d49580e727d46616fc58658c7fd8050e718cd61d5" Feb 25 08:55:28 crc kubenswrapper[4978]: I0225 08:55:28.988643 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.057676 4978 scope.go:117] "RemoveContainer" containerID="c4de36a782c0753857e4120b568d80b9f8c0597f57044467377c109bd4e5a34e" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.060612 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.073039 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.087108 4978 scope.go:117] "RemoveContainer" containerID="c06d036c4537043204215edfd958423d057a4d0c5d8be8199ab455f29c014677" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.094673 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Feb 25 08:55:29 crc kubenswrapper[4978]: E0225 08:55:29.095158 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-api" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.095174 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-api" Feb 25 08:55:29 crc kubenswrapper[4978]: E0225 08:55:29.095195 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-notifier" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.095204 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-notifier" Feb 25 08:55:29 crc kubenswrapper[4978]: E0225 08:55:29.095224 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-listener" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.095233 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-listener" Feb 25 08:55:29 crc kubenswrapper[4978]: E0225 08:55:29.095261 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-evaluator" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.095270 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-evaluator" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.095516 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-notifier" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.095541 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-api" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.095558 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-evaluator" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.095574 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" containerName="aodh-listener" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.099952 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.102801 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.103500 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.103716 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.103866 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-7bmkx" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.111143 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.125013 4978 scope.go:117] "RemoveContainer" containerID="15e1faa4ea8831012f680f41a64d9a7afa74764825e1e9777e66ff9dbb0d9409" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.134440 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.168194 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-public-tls-certs\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.168396 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-config-data\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.168511 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52f9q\" (UniqueName: \"kubernetes.io/projected/9e08daef-04f0-4303-80a4-a8b21bf0da3d-kube-api-access-52f9q\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.168788 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-internal-tls-certs\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.168926 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.169107 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-scripts\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.271408 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-internal-tls-certs\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.271491 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.271574 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-scripts\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.271663 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-public-tls-certs\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.271700 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52f9q\" (UniqueName: \"kubernetes.io/projected/9e08daef-04f0-4303-80a4-a8b21bf0da3d-kube-api-access-52f9q\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.272621 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-config-data\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.275579 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-scripts\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.276006 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.276102 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-public-tls-certs\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.277192 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-config-data\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.277649 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e08daef-04f0-4303-80a4-a8b21bf0da3d-internal-tls-certs\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.288646 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52f9q\" (UniqueName: \"kubernetes.io/projected/9e08daef-04f0-4303-80a4-a8b21bf0da3d-kube-api-access-52f9q\") pod \"aodh-0\" (UID: \"9e08daef-04f0-4303-80a4-a8b21bf0da3d\") " pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.342454 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1e9a596-1d6c-4f92-9867-c8a4f644f2c5" path="/var/lib/kubelet/pods/c1e9a596-1d6c-4f92-9867-c8a4f644f2c5/volumes" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.429148 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Feb 25 08:55:29 crc kubenswrapper[4978]: I0225 08:55:29.779624 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Feb 25 08:55:30 crc kubenswrapper[4978]: I0225 08:55:30.003191 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"9e08daef-04f0-4303-80a4-a8b21bf0da3d","Type":"ContainerStarted","Data":"6f4ae66e13b35b1bf94c3f85cc1fdc2eea8b279815b87d0b46fcd125673cdb8c"} Feb 25 08:55:30 crc kubenswrapper[4978]: I0225 08:55:30.977200 4978 scope.go:117] "RemoveContainer" containerID="aaccef42f90917448f09edc3c29b0e8dc12daba8eb38d20a87e3a4184d0b8445" Feb 25 08:55:31 crc kubenswrapper[4978]: I0225 08:55:31.023128 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"9e08daef-04f0-4303-80a4-a8b21bf0da3d","Type":"ContainerStarted","Data":"f24917041c01703f2ed77f5924b34288276f2585bd73c223dfa36a909ddf18c6"} Feb 25 08:55:31 crc kubenswrapper[4978]: I0225 08:55:31.023763 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"9e08daef-04f0-4303-80a4-a8b21bf0da3d","Type":"ContainerStarted","Data":"268cff224d7839f4d600ccd62b52144c626b09cced41ed9c08ac389ead239b4f"} Feb 25 08:55:32 crc kubenswrapper[4978]: I0225 08:55:32.034761 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"9e08daef-04f0-4303-80a4-a8b21bf0da3d","Type":"ContainerStarted","Data":"1a995cb2db19bc98cc77741259a0246ba4b02012a5f991995aa1a400e7161506"} Feb 25 08:55:32 crc kubenswrapper[4978]: I0225 08:55:32.035280 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"9e08daef-04f0-4303-80a4-a8b21bf0da3d","Type":"ContainerStarted","Data":"e257af1095f10f36e67bd6d92dba444c9d4f045148dd5a79c5ed7cebaaffca1d"} Feb 25 08:55:32 crc kubenswrapper[4978]: I0225 08:55:32.063717 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.6012845310000001 podStartE2EDuration="3.063694458s" podCreationTimestamp="2026-02-25 08:55:29 +0000 UTC" firstStartedPulling="2026-02-25 08:55:29.776269732 +0000 UTC m=+7823.215526201" lastFinishedPulling="2026-02-25 08:55:31.238679659 +0000 UTC m=+7824.677936128" observedRunningTime="2026-02-25 08:55:32.058711554 +0000 UTC m=+7825.497968053" watchObservedRunningTime="2026-02-25 08:55:32.063694458 +0000 UTC m=+7825.502950917" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.118424 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.218833 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mmg7b"] Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.227445 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.244111 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mmg7b"] Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.308483 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n6gw\" (UniqueName: \"kubernetes.io/projected/baa2b0f4-c923-4d74-a112-bf7e5cae275e-kube-api-access-5n6gw\") pod \"redhat-operators-mmg7b\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.308688 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-catalog-content\") pod \"redhat-operators-mmg7b\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.308828 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-utilities\") pod \"redhat-operators-mmg7b\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.410876 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n6gw\" (UniqueName: \"kubernetes.io/projected/baa2b0f4-c923-4d74-a112-bf7e5cae275e-kube-api-access-5n6gw\") pod \"redhat-operators-mmg7b\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.411253 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-catalog-content\") pod \"redhat-operators-mmg7b\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.411306 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-utilities\") pod \"redhat-operators-mmg7b\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.413673 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-catalog-content\") pod \"redhat-operators-mmg7b\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.413894 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-utilities\") pod \"redhat-operators-mmg7b\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.435635 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n6gw\" (UniqueName: \"kubernetes.io/projected/baa2b0f4-c923-4d74-a112-bf7e5cae275e-kube-api-access-5n6gw\") pod \"redhat-operators-mmg7b\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:35 crc kubenswrapper[4978]: I0225 08:55:35.585934 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:36 crc kubenswrapper[4978]: I0225 08:55:36.141135 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mmg7b"] Feb 25 08:55:37 crc kubenswrapper[4978]: I0225 08:55:37.088658 4978 generic.go:334] "Generic (PLEG): container finished" podID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerID="9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e" exitCode=0 Feb 25 08:55:37 crc kubenswrapper[4978]: I0225 08:55:37.088712 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmg7b" event={"ID":"baa2b0f4-c923-4d74-a112-bf7e5cae275e","Type":"ContainerDied","Data":"9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e"} Feb 25 08:55:37 crc kubenswrapper[4978]: I0225 08:55:37.088776 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmg7b" event={"ID":"baa2b0f4-c923-4d74-a112-bf7e5cae275e","Type":"ContainerStarted","Data":"876380a0b430b9054869e8abaffa57b9ff3c18221b67b3affb7ddc75cfcfb68c"} Feb 25 08:55:38 crc kubenswrapper[4978]: I0225 08:55:38.101921 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmg7b" event={"ID":"baa2b0f4-c923-4d74-a112-bf7e5cae275e","Type":"ContainerStarted","Data":"a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3"} Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.098138 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fb685cd79-hs5wv"] Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.100643 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.104793 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.119439 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fb685cd79-hs5wv"] Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.160604 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-config\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.160718 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-openstack-cell1\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.160757 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-sb\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.160781 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-dns-svc\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.160824 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f96fw\" (UniqueName: \"kubernetes.io/projected/eeeb7261-b809-4530-9ac1-90e20926374f-kube-api-access-f96fw\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.160862 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-nb\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.214424 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fb685cd79-hs5wv"] Feb 25 08:55:42 crc kubenswrapper[4978]: E0225 08:55:42.215309 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-f96fw openstack-cell1 ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" podUID="eeeb7261-b809-4530-9ac1-90e20926374f" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.246008 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7dd877b77-rgkct"] Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.248168 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.254782 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-networker" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.256120 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7dd877b77-rgkct"] Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.263090 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-config\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.263184 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-openstack-cell1\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.263216 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-sb\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.263233 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-dns-svc\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.263275 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f96fw\" (UniqueName: \"kubernetes.io/projected/eeeb7261-b809-4530-9ac1-90e20926374f-kube-api-access-f96fw\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.263303 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-nb\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.264403 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-nb\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.265043 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-config\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.265730 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-openstack-cell1\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.266431 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-sb\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.267147 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-dns-svc\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.285945 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f96fw\" (UniqueName: \"kubernetes.io/projected/eeeb7261-b809-4530-9ac1-90e20926374f-kube-api-access-f96fw\") pod \"dnsmasq-dns-5fb685cd79-hs5wv\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.328110 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:55:42 crc kubenswrapper[4978]: E0225 08:55:42.328655 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.364854 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvljv\" (UniqueName: \"kubernetes.io/projected/8867ff4e-e020-4d2c-a851-40279f46b8fa-kube-api-access-pvljv\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.364967 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-networker\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.365067 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-cell1\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.365088 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-config\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.365121 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-dns-svc\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.365144 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-sb\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.365187 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-nb\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.467511 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-dns-svc\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.467589 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-sb\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.467671 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-nb\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.467713 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvljv\" (UniqueName: \"kubernetes.io/projected/8867ff4e-e020-4d2c-a851-40279f46b8fa-kube-api-access-pvljv\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.467943 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-networker\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.468110 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-cell1\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.468204 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-config\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.468575 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-dns-svc\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.468597 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-nb\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.468924 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-cell1\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.469468 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-networker\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.469716 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-config\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.470066 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-sb\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.486357 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvljv\" (UniqueName: \"kubernetes.io/projected/8867ff4e-e020-4d2c-a851-40279f46b8fa-kube-api-access-pvljv\") pod \"dnsmasq-dns-7dd877b77-rgkct\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:42 crc kubenswrapper[4978]: I0225 08:55:42.639705 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.181714 4978 generic.go:334] "Generic (PLEG): container finished" podID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerID="a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3" exitCode=0 Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.181774 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmg7b" event={"ID":"baa2b0f4-c923-4d74-a112-bf7e5cae275e","Type":"ContainerDied","Data":"a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3"} Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.182139 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.183019 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7dd877b77-rgkct"] Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.227854 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.284270 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-nb\") pod \"eeeb7261-b809-4530-9ac1-90e20926374f\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.284344 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-dns-svc\") pod \"eeeb7261-b809-4530-9ac1-90e20926374f\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.284493 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f96fw\" (UniqueName: \"kubernetes.io/projected/eeeb7261-b809-4530-9ac1-90e20926374f-kube-api-access-f96fw\") pod \"eeeb7261-b809-4530-9ac1-90e20926374f\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.284590 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-openstack-cell1\") pod \"eeeb7261-b809-4530-9ac1-90e20926374f\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.284611 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-config\") pod \"eeeb7261-b809-4530-9ac1-90e20926374f\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.284655 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-sb\") pod \"eeeb7261-b809-4530-9ac1-90e20926374f\" (UID: \"eeeb7261-b809-4530-9ac1-90e20926374f\") " Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.284928 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eeeb7261-b809-4530-9ac1-90e20926374f" (UID: "eeeb7261-b809-4530-9ac1-90e20926374f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.285287 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-config" (OuterVolumeSpecName: "config") pod "eeeb7261-b809-4530-9ac1-90e20926374f" (UID: "eeeb7261-b809-4530-9ac1-90e20926374f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.285318 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.284943 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eeeb7261-b809-4530-9ac1-90e20926374f" (UID: "eeeb7261-b809-4530-9ac1-90e20926374f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.285733 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eeeb7261-b809-4530-9ac1-90e20926374f" (UID: "eeeb7261-b809-4530-9ac1-90e20926374f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.285731 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "eeeb7261-b809-4530-9ac1-90e20926374f" (UID: "eeeb7261-b809-4530-9ac1-90e20926374f"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.289617 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeeb7261-b809-4530-9ac1-90e20926374f-kube-api-access-f96fw" (OuterVolumeSpecName: "kube-api-access-f96fw") pod "eeeb7261-b809-4530-9ac1-90e20926374f" (UID: "eeeb7261-b809-4530-9ac1-90e20926374f"). InnerVolumeSpecName "kube-api-access-f96fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.387966 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f96fw\" (UniqueName: \"kubernetes.io/projected/eeeb7261-b809-4530-9ac1-90e20926374f-kube-api-access-f96fw\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.388005 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.388017 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.388030 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:43 crc kubenswrapper[4978]: I0225 08:55:43.388039 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eeeb7261-b809-4530-9ac1-90e20926374f-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.037694 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-rkqhs"] Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.047175 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b014-account-create-update-m2qtz"] Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.056978 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-rkqhs"] Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.065178 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b014-account-create-update-m2qtz"] Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.191450 4978 generic.go:334] "Generic (PLEG): container finished" podID="8867ff4e-e020-4d2c-a851-40279f46b8fa" containerID="0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4" exitCode=0 Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.191541 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" event={"ID":"8867ff4e-e020-4d2c-a851-40279f46b8fa","Type":"ContainerDied","Data":"0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4"} Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.191580 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" event={"ID":"8867ff4e-e020-4d2c-a851-40279f46b8fa","Type":"ContainerStarted","Data":"883dcfe1568492dbd63317fd51ff982e67fe8e22cc0aa282f87bd766226af6f2"} Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.199502 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmg7b" event={"ID":"baa2b0f4-c923-4d74-a112-bf7e5cae275e","Type":"ContainerStarted","Data":"cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe"} Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.199535 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fb685cd79-hs5wv" Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.245338 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mmg7b" podStartSLOduration=2.635055144 podStartE2EDuration="9.245321327s" podCreationTimestamp="2026-02-25 08:55:35 +0000 UTC" firstStartedPulling="2026-02-25 08:55:37.090794742 +0000 UTC m=+7830.530051241" lastFinishedPulling="2026-02-25 08:55:43.701060955 +0000 UTC m=+7837.140317424" observedRunningTime="2026-02-25 08:55:44.241430706 +0000 UTC m=+7837.680687185" watchObservedRunningTime="2026-02-25 08:55:44.245321327 +0000 UTC m=+7837.684577786" Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.446463 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5fb685cd79-hs5wv"] Feb 25 08:55:44 crc kubenswrapper[4978]: I0225 08:55:44.456488 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5fb685cd79-hs5wv"] Feb 25 08:55:45 crc kubenswrapper[4978]: I0225 08:55:45.216643 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" event={"ID":"8867ff4e-e020-4d2c-a851-40279f46b8fa","Type":"ContainerStarted","Data":"059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff"} Feb 25 08:55:45 crc kubenswrapper[4978]: I0225 08:55:45.216843 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:45 crc kubenswrapper[4978]: I0225 08:55:45.254790 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" podStartSLOduration=3.25476757 podStartE2EDuration="3.25476757s" podCreationTimestamp="2026-02-25 08:55:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:55:45.245124551 +0000 UTC m=+7838.684381050" watchObservedRunningTime="2026-02-25 08:55:45.25476757 +0000 UTC m=+7838.694024059" Feb 25 08:55:45 crc kubenswrapper[4978]: I0225 08:55:45.348786 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34c1a11b-9cf0-4f8a-beb0-14739a3b79c2" path="/var/lib/kubelet/pods/34c1a11b-9cf0-4f8a-beb0-14739a3b79c2/volumes" Feb 25 08:55:45 crc kubenswrapper[4978]: I0225 08:55:45.352720 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a891b63-df50-4203-b1e6-c7b86991fec3" path="/var/lib/kubelet/pods/9a891b63-df50-4203-b1e6-c7b86991fec3/volumes" Feb 25 08:55:45 crc kubenswrapper[4978]: I0225 08:55:45.357355 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeeb7261-b809-4530-9ac1-90e20926374f" path="/var/lib/kubelet/pods/eeeb7261-b809-4530-9ac1-90e20926374f/volumes" Feb 25 08:55:45 crc kubenswrapper[4978]: I0225 08:55:45.587258 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:45 crc kubenswrapper[4978]: I0225 08:55:45.587339 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:55:46 crc kubenswrapper[4978]: I0225 08:55:46.656708 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mmg7b" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerName="registry-server" probeResult="failure" output=< Feb 25 08:55:46 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 08:55:46 crc kubenswrapper[4978]: > Feb 25 08:55:52 crc kubenswrapper[4978]: I0225 08:55:52.642557 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:55:52 crc kubenswrapper[4978]: I0225 08:55:52.727323 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797c8cd5-csvwh"] Feb 25 08:55:52 crc kubenswrapper[4978]: I0225 08:55:52.727860 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" podUID="5714c95c-8a9f-4885-b450-3867dc4ce904" containerName="dnsmasq-dns" containerID="cri-o://2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4" gracePeriod=10 Feb 25 08:55:52 crc kubenswrapper[4978]: I0225 08:55:52.984222 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7799fc56d5-cbjnr"] Feb 25 08:55:52 crc kubenswrapper[4978]: I0225 08:55:52.988297 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.000732 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7799fc56d5-cbjnr"] Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.132010 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-config\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.132378 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-networker\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.132533 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79rnq\" (UniqueName: \"kubernetes.io/projected/e3d28a89-7825-49d2-bb3d-230edc0a067c-kube-api-access-79rnq\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.132571 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-sb\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.132776 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-nb\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.132893 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-dns-svc\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.133005 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-cell1\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.234251 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-dns-svc\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.234325 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-cell1\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.234381 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-config\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.234412 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-networker\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.234467 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79rnq\" (UniqueName: \"kubernetes.io/projected/e3d28a89-7825-49d2-bb3d-230edc0a067c-kube-api-access-79rnq\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.234490 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-sb\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.234532 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-nb\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.235220 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-nb\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.235305 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-dns-svc\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.235754 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-cell1\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.236126 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-networker\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.236504 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-sb\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.236909 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-config\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.236972 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.259910 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79rnq\" (UniqueName: \"kubernetes.io/projected/e3d28a89-7825-49d2-bb3d-230edc0a067c-kube-api-access-79rnq\") pod \"dnsmasq-dns-7799fc56d5-cbjnr\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.309663 4978 generic.go:334] "Generic (PLEG): container finished" podID="5714c95c-8a9f-4885-b450-3867dc4ce904" containerID="2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4" exitCode=0 Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.309729 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" event={"ID":"5714c95c-8a9f-4885-b450-3867dc4ce904","Type":"ContainerDied","Data":"2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4"} Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.309784 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.309815 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797c8cd5-csvwh" event={"ID":"5714c95c-8a9f-4885-b450-3867dc4ce904","Type":"ContainerDied","Data":"54bebd2292834da68bbe0ddbb37867737e5747c9d3ef13214f983162572b288b"} Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.309849 4978 scope.go:117] "RemoveContainer" containerID="2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.313085 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.328362 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:55:53 crc kubenswrapper[4978]: E0225 08:55:53.328640 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.335684 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-config\") pod \"5714c95c-8a9f-4885-b450-3867dc4ce904\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.336101 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-sb\") pod \"5714c95c-8a9f-4885-b450-3867dc4ce904\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.336183 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-nb\") pod \"5714c95c-8a9f-4885-b450-3867dc4ce904\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.336222 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-dns-svc\") pod \"5714c95c-8a9f-4885-b450-3867dc4ce904\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.336256 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xmqk\" (UniqueName: \"kubernetes.io/projected/5714c95c-8a9f-4885-b450-3867dc4ce904-kube-api-access-9xmqk\") pod \"5714c95c-8a9f-4885-b450-3867dc4ce904\" (UID: \"5714c95c-8a9f-4885-b450-3867dc4ce904\") " Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.339287 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5714c95c-8a9f-4885-b450-3867dc4ce904-kube-api-access-9xmqk" (OuterVolumeSpecName: "kube-api-access-9xmqk") pod "5714c95c-8a9f-4885-b450-3867dc4ce904" (UID: "5714c95c-8a9f-4885-b450-3867dc4ce904"). InnerVolumeSpecName "kube-api-access-9xmqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.346255 4978 scope.go:117] "RemoveContainer" containerID="798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.381701 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-config" (OuterVolumeSpecName: "config") pod "5714c95c-8a9f-4885-b450-3867dc4ce904" (UID: "5714c95c-8a9f-4885-b450-3867dc4ce904"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.381850 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5714c95c-8a9f-4885-b450-3867dc4ce904" (UID: "5714c95c-8a9f-4885-b450-3867dc4ce904"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.387033 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5714c95c-8a9f-4885-b450-3867dc4ce904" (UID: "5714c95c-8a9f-4885-b450-3867dc4ce904"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.395813 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5714c95c-8a9f-4885-b450-3867dc4ce904" (UID: "5714c95c-8a9f-4885-b450-3867dc4ce904"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.407650 4978 scope.go:117] "RemoveContainer" containerID="2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4" Feb 25 08:55:53 crc kubenswrapper[4978]: E0225 08:55:53.408092 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4\": container with ID starting with 2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4 not found: ID does not exist" containerID="2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.408136 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4"} err="failed to get container status \"2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4\": rpc error: code = NotFound desc = could not find container \"2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4\": container with ID starting with 2ea46078b8458112eab378172d41c14499c1328122df654909d37ca33994abb4 not found: ID does not exist" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.408164 4978 scope.go:117] "RemoveContainer" containerID="798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b" Feb 25 08:55:53 crc kubenswrapper[4978]: E0225 08:55:53.409547 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b\": container with ID starting with 798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b not found: ID does not exist" containerID="798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.409590 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b"} err="failed to get container status \"798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b\": rpc error: code = NotFound desc = could not find container \"798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b\": container with ID starting with 798162735ba69e12a2946d135064eb852f19825aec59e9e8b6d728faed2b959b not found: ID does not exist" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.439917 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.440446 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.440460 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.440472 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5714c95c-8a9f-4885-b450-3867dc4ce904-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.440485 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xmqk\" (UniqueName: \"kubernetes.io/projected/5714c95c-8a9f-4885-b450-3867dc4ce904-kube-api-access-9xmqk\") on node \"crc\" DevicePath \"\"" Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.651299 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797c8cd5-csvwh"] Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.660676 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-797c8cd5-csvwh"] Feb 25 08:55:53 crc kubenswrapper[4978]: I0225 08:55:53.796387 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7799fc56d5-cbjnr"] Feb 25 08:55:54 crc kubenswrapper[4978]: I0225 08:55:54.322017 4978 generic.go:334] "Generic (PLEG): container finished" podID="e3d28a89-7825-49d2-bb3d-230edc0a067c" containerID="9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2" exitCode=0 Feb 25 08:55:54 crc kubenswrapper[4978]: I0225 08:55:54.322123 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" event={"ID":"e3d28a89-7825-49d2-bb3d-230edc0a067c","Type":"ContainerDied","Data":"9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2"} Feb 25 08:55:54 crc kubenswrapper[4978]: I0225 08:55:54.322424 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" event={"ID":"e3d28a89-7825-49d2-bb3d-230edc0a067c","Type":"ContainerStarted","Data":"7875de37d6454cf6996f17f479c64186686823e1042a199ff271822d78b125ec"} Feb 25 08:55:55 crc kubenswrapper[4978]: I0225 08:55:55.360074 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5714c95c-8a9f-4885-b450-3867dc4ce904" path="/var/lib/kubelet/pods/5714c95c-8a9f-4885-b450-3867dc4ce904/volumes" Feb 25 08:55:55 crc kubenswrapper[4978]: I0225 08:55:55.361736 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" event={"ID":"e3d28a89-7825-49d2-bb3d-230edc0a067c","Type":"ContainerStarted","Data":"9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247"} Feb 25 08:55:55 crc kubenswrapper[4978]: I0225 08:55:55.361790 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:55:55 crc kubenswrapper[4978]: I0225 08:55:55.377469 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" podStartSLOduration=3.377450499 podStartE2EDuration="3.377450499s" podCreationTimestamp="2026-02-25 08:55:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:55:55.375120447 +0000 UTC m=+7848.814376916" watchObservedRunningTime="2026-02-25 08:55:55.377450499 +0000 UTC m=+7848.816706959" Feb 25 08:55:56 crc kubenswrapper[4978]: I0225 08:55:56.645456 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mmg7b" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerName="registry-server" probeResult="failure" output=< Feb 25 08:55:56 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 08:55:56 crc kubenswrapper[4978]: > Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.141449 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533496-9t8t6"] Feb 25 08:56:00 crc kubenswrapper[4978]: E0225 08:56:00.142482 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5714c95c-8a9f-4885-b450-3867dc4ce904" containerName="dnsmasq-dns" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.142498 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5714c95c-8a9f-4885-b450-3867dc4ce904" containerName="dnsmasq-dns" Feb 25 08:56:00 crc kubenswrapper[4978]: E0225 08:56:00.143265 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5714c95c-8a9f-4885-b450-3867dc4ce904" containerName="init" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.143276 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5714c95c-8a9f-4885-b450-3867dc4ce904" containerName="init" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.143608 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5714c95c-8a9f-4885-b450-3867dc4ce904" containerName="dnsmasq-dns" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.144600 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533496-9t8t6" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.146823 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.147111 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.147231 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.158500 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533496-9t8t6"] Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.217082 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssgg6\" (UniqueName: \"kubernetes.io/projected/72a0184c-d75c-4ecb-a38a-d793e3c7da1d-kube-api-access-ssgg6\") pod \"auto-csr-approver-29533496-9t8t6\" (UID: \"72a0184c-d75c-4ecb-a38a-d793e3c7da1d\") " pod="openshift-infra/auto-csr-approver-29533496-9t8t6" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.318904 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssgg6\" (UniqueName: \"kubernetes.io/projected/72a0184c-d75c-4ecb-a38a-d793e3c7da1d-kube-api-access-ssgg6\") pod \"auto-csr-approver-29533496-9t8t6\" (UID: \"72a0184c-d75c-4ecb-a38a-d793e3c7da1d\") " pod="openshift-infra/auto-csr-approver-29533496-9t8t6" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.338986 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssgg6\" (UniqueName: \"kubernetes.io/projected/72a0184c-d75c-4ecb-a38a-d793e3c7da1d-kube-api-access-ssgg6\") pod \"auto-csr-approver-29533496-9t8t6\" (UID: \"72a0184c-d75c-4ecb-a38a-d793e3c7da1d\") " pod="openshift-infra/auto-csr-approver-29533496-9t8t6" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.464420 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533496-9t8t6" Feb 25 08:56:00 crc kubenswrapper[4978]: I0225 08:56:00.966930 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533496-9t8t6"] Feb 25 08:56:01 crc kubenswrapper[4978]: I0225 08:56:01.421195 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533496-9t8t6" event={"ID":"72a0184c-d75c-4ecb-a38a-d793e3c7da1d","Type":"ContainerStarted","Data":"f091eded8e19e4641e0b38ba8b4c8128ca9c65262624505e9865bcbe96d090a6"} Feb 25 08:56:02 crc kubenswrapper[4978]: I0225 08:56:02.433647 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533496-9t8t6" event={"ID":"72a0184c-d75c-4ecb-a38a-d793e3c7da1d","Type":"ContainerStarted","Data":"57f4b3b77fd4a9550a64e37be2a944b45a7c23612eeae1ce3dbb169cd475bf0e"} Feb 25 08:56:02 crc kubenswrapper[4978]: I0225 08:56:02.461185 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533496-9t8t6" podStartSLOduration=1.490093695 podStartE2EDuration="2.46115332s" podCreationTimestamp="2026-02-25 08:56:00 +0000 UTC" firstStartedPulling="2026-02-25 08:56:00.971112296 +0000 UTC m=+7854.410368765" lastFinishedPulling="2026-02-25 08:56:01.942171931 +0000 UTC m=+7855.381428390" observedRunningTime="2026-02-25 08:56:02.453555064 +0000 UTC m=+7855.892811633" watchObservedRunningTime="2026-02-25 08:56:02.46115332 +0000 UTC m=+7855.900409829" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.314669 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.407597 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dd877b77-rgkct"] Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.408459 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" podUID="8867ff4e-e020-4d2c-a851-40279f46b8fa" containerName="dnsmasq-dns" containerID="cri-o://059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff" gracePeriod=10 Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.476112 4978 generic.go:334] "Generic (PLEG): container finished" podID="72a0184c-d75c-4ecb-a38a-d793e3c7da1d" containerID="57f4b3b77fd4a9550a64e37be2a944b45a7c23612eeae1ce3dbb169cd475bf0e" exitCode=0 Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.476188 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533496-9t8t6" event={"ID":"72a0184c-d75c-4ecb-a38a-d793e3c7da1d","Type":"ContainerDied","Data":"57f4b3b77fd4a9550a64e37be2a944b45a7c23612eeae1ce3dbb169cd475bf0e"} Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.582964 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65cf54b65f-6jwcs"] Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.621334 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.629161 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65cf54b65f-6jwcs"] Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.723641 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-networker\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.723714 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-sb\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.723930 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-cell1\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.724023 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-config\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.724065 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-nb\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.724185 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw9v6\" (UniqueName: \"kubernetes.io/projected/ae3313f2-822b-42cd-826a-daac63f0926c-kube-api-access-bw9v6\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.724240 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-dns-svc\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.826160 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-networker\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.827080 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-sb\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.827144 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-cell1\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.827183 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-config\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.827210 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-nb\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.827265 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw9v6\" (UniqueName: \"kubernetes.io/projected/ae3313f2-822b-42cd-826a-daac63f0926c-kube-api-access-bw9v6\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.827300 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-dns-svc\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.827941 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-networker\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.828334 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-dns-svc\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.832411 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-sb\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.832927 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-nb\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.834975 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-config\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.836581 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-cell1\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.854314 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw9v6\" (UniqueName: \"kubernetes.io/projected/ae3313f2-822b-42cd-826a-daac63f0926c-kube-api-access-bw9v6\") pod \"dnsmasq-dns-65cf54b65f-6jwcs\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.943876 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:03 crc kubenswrapper[4978]: I0225 08:56:03.944255 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.030252 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-cell1\") pod \"8867ff4e-e020-4d2c-a851-40279f46b8fa\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.030304 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-networker\") pod \"8867ff4e-e020-4d2c-a851-40279f46b8fa\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.030430 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvljv\" (UniqueName: \"kubernetes.io/projected/8867ff4e-e020-4d2c-a851-40279f46b8fa-kube-api-access-pvljv\") pod \"8867ff4e-e020-4d2c-a851-40279f46b8fa\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.030522 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-sb\") pod \"8867ff4e-e020-4d2c-a851-40279f46b8fa\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.030575 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-dns-svc\") pod \"8867ff4e-e020-4d2c-a851-40279f46b8fa\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.030594 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-nb\") pod \"8867ff4e-e020-4d2c-a851-40279f46b8fa\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.030634 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-config\") pod \"8867ff4e-e020-4d2c-a851-40279f46b8fa\" (UID: \"8867ff4e-e020-4d2c-a851-40279f46b8fa\") " Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.036772 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8867ff4e-e020-4d2c-a851-40279f46b8fa-kube-api-access-pvljv" (OuterVolumeSpecName: "kube-api-access-pvljv") pod "8867ff4e-e020-4d2c-a851-40279f46b8fa" (UID: "8867ff4e-e020-4d2c-a851-40279f46b8fa"). InnerVolumeSpecName "kube-api-access-pvljv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.094845 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8867ff4e-e020-4d2c-a851-40279f46b8fa" (UID: "8867ff4e-e020-4d2c-a851-40279f46b8fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.103943 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "8867ff4e-e020-4d2c-a851-40279f46b8fa" (UID: "8867ff4e-e020-4d2c-a851-40279f46b8fa"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.124077 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-config" (OuterVolumeSpecName: "config") pod "8867ff4e-e020-4d2c-a851-40279f46b8fa" (UID: "8867ff4e-e020-4d2c-a851-40279f46b8fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.124087 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8867ff4e-e020-4d2c-a851-40279f46b8fa" (UID: "8867ff4e-e020-4d2c-a851-40279f46b8fa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.125664 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-networker" (OuterVolumeSpecName: "openstack-networker") pod "8867ff4e-e020-4d2c-a851-40279f46b8fa" (UID: "8867ff4e-e020-4d2c-a851-40279f46b8fa"). InnerVolumeSpecName "openstack-networker". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.135015 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvljv\" (UniqueName: \"kubernetes.io/projected/8867ff4e-e020-4d2c-a851-40279f46b8fa-kube-api-access-pvljv\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.135051 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.135063 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.135074 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.135082 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.135092 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.141676 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8867ff4e-e020-4d2c-a851-40279f46b8fa" (UID: "8867ff4e-e020-4d2c-a851-40279f46b8fa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.237048 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8867ff4e-e020-4d2c-a851-40279f46b8fa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:04 crc kubenswrapper[4978]: W0225 08:56:04.402297 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae3313f2_822b_42cd_826a_daac63f0926c.slice/crio-5bda5b472b78c8549b087f0ae648ee30d742ad71773ff0f840e5e436533df37e WatchSource:0}: Error finding container 5bda5b472b78c8549b087f0ae648ee30d742ad71773ff0f840e5e436533df37e: Status 404 returned error can't find the container with id 5bda5b472b78c8549b087f0ae648ee30d742ad71773ff0f840e5e436533df37e Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.404186 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65cf54b65f-6jwcs"] Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.487485 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" event={"ID":"ae3313f2-822b-42cd-826a-daac63f0926c","Type":"ContainerStarted","Data":"5bda5b472b78c8549b087f0ae648ee30d742ad71773ff0f840e5e436533df37e"} Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.495774 4978 generic.go:334] "Generic (PLEG): container finished" podID="8867ff4e-e020-4d2c-a851-40279f46b8fa" containerID="059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff" exitCode=0 Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.495897 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.495884 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" event={"ID":"8867ff4e-e020-4d2c-a851-40279f46b8fa","Type":"ContainerDied","Data":"059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff"} Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.496137 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dd877b77-rgkct" event={"ID":"8867ff4e-e020-4d2c-a851-40279f46b8fa","Type":"ContainerDied","Data":"883dcfe1568492dbd63317fd51ff982e67fe8e22cc0aa282f87bd766226af6f2"} Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.496182 4978 scope.go:117] "RemoveContainer" containerID="059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.519661 4978 scope.go:117] "RemoveContainer" containerID="0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.537167 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dd877b77-rgkct"] Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.545696 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7dd877b77-rgkct"] Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.629079 4978 scope.go:117] "RemoveContainer" containerID="059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff" Feb 25 08:56:04 crc kubenswrapper[4978]: E0225 08:56:04.629891 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff\": container with ID starting with 059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff not found: ID does not exist" containerID="059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.629927 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff"} err="failed to get container status \"059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff\": rpc error: code = NotFound desc = could not find container \"059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff\": container with ID starting with 059e7c42a8652545e9f3996576e9d754a16d08ac99d50f6c9e24f9483fd464ff not found: ID does not exist" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.629954 4978 scope.go:117] "RemoveContainer" containerID="0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4" Feb 25 08:56:04 crc kubenswrapper[4978]: E0225 08:56:04.630391 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4\": container with ID starting with 0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4 not found: ID does not exist" containerID="0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.630414 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4"} err="failed to get container status \"0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4\": rpc error: code = NotFound desc = could not find container \"0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4\": container with ID starting with 0fa9257b506986f3559bc168868a133c2d0682b1efb15b4ad88c9558caadd2d4 not found: ID does not exist" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.896926 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533496-9t8t6" Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.949417 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssgg6\" (UniqueName: \"kubernetes.io/projected/72a0184c-d75c-4ecb-a38a-d793e3c7da1d-kube-api-access-ssgg6\") pod \"72a0184c-d75c-4ecb-a38a-d793e3c7da1d\" (UID: \"72a0184c-d75c-4ecb-a38a-d793e3c7da1d\") " Feb 25 08:56:04 crc kubenswrapper[4978]: I0225 08:56:04.955705 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72a0184c-d75c-4ecb-a38a-d793e3c7da1d-kube-api-access-ssgg6" (OuterVolumeSpecName: "kube-api-access-ssgg6") pod "72a0184c-d75c-4ecb-a38a-d793e3c7da1d" (UID: "72a0184c-d75c-4ecb-a38a-d793e3c7da1d"). InnerVolumeSpecName "kube-api-access-ssgg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.052023 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssgg6\" (UniqueName: \"kubernetes.io/projected/72a0184c-d75c-4ecb-a38a-d793e3c7da1d-kube-api-access-ssgg6\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.351443 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8867ff4e-e020-4d2c-a851-40279f46b8fa" path="/var/lib/kubelet/pods/8867ff4e-e020-4d2c-a851-40279f46b8fa/volumes" Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.511658 4978 generic.go:334] "Generic (PLEG): container finished" podID="ae3313f2-822b-42cd-826a-daac63f0926c" containerID="f7a06736fe3496594a681d6fd01f1970c4c0db5e15cc79375fd86f62b61f1859" exitCode=0 Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.511782 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" event={"ID":"ae3313f2-822b-42cd-826a-daac63f0926c","Type":"ContainerDied","Data":"f7a06736fe3496594a681d6fd01f1970c4c0db5e15cc79375fd86f62b61f1859"} Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.520484 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533496-9t8t6" event={"ID":"72a0184c-d75c-4ecb-a38a-d793e3c7da1d","Type":"ContainerDied","Data":"f091eded8e19e4641e0b38ba8b4c8128ca9c65262624505e9865bcbe96d090a6"} Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.520536 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f091eded8e19e4641e0b38ba8b4c8128ca9c65262624505e9865bcbe96d090a6" Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.520564 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533496-9t8t6" Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.546659 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533490-mv8mf"] Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.559826 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533490-mv8mf"] Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.663742 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:56:05 crc kubenswrapper[4978]: I0225 08:56:05.737009 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:56:06 crc kubenswrapper[4978]: I0225 08:56:06.327909 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:56:06 crc kubenswrapper[4978]: E0225 08:56:06.328231 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:56:06 crc kubenswrapper[4978]: I0225 08:56:06.419173 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mmg7b"] Feb 25 08:56:06 crc kubenswrapper[4978]: I0225 08:56:06.550424 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" event={"ID":"ae3313f2-822b-42cd-826a-daac63f0926c","Type":"ContainerStarted","Data":"b6d84f4c7ef5cc6a0a8d1062144e6bef8c995b07806707f434b173641cd2f362"} Feb 25 08:56:06 crc kubenswrapper[4978]: I0225 08:56:06.551485 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:06 crc kubenswrapper[4978]: I0225 08:56:06.580701 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" podStartSLOduration=3.580674646 podStartE2EDuration="3.580674646s" podCreationTimestamp="2026-02-25 08:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 08:56:06.578804158 +0000 UTC m=+7860.018060627" watchObservedRunningTime="2026-02-25 08:56:06.580674646 +0000 UTC m=+7860.019931115" Feb 25 08:56:07 crc kubenswrapper[4978]: I0225 08:56:07.346409 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0" path="/var/lib/kubelet/pods/e4bc8ad1-f5e7-410c-aa35-4b63a5a7fea0/volumes" Feb 25 08:56:07 crc kubenswrapper[4978]: I0225 08:56:07.566720 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mmg7b" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerName="registry-server" containerID="cri-o://cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe" gracePeriod=2 Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.154679 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.224890 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-catalog-content\") pod \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.226394 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n6gw\" (UniqueName: \"kubernetes.io/projected/baa2b0f4-c923-4d74-a112-bf7e5cae275e-kube-api-access-5n6gw\") pod \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.226577 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-utilities\") pod \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\" (UID: \"baa2b0f4-c923-4d74-a112-bf7e5cae275e\") " Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.227815 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-utilities" (OuterVolumeSpecName: "utilities") pod "baa2b0f4-c923-4d74-a112-bf7e5cae275e" (UID: "baa2b0f4-c923-4d74-a112-bf7e5cae275e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.233586 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baa2b0f4-c923-4d74-a112-bf7e5cae275e-kube-api-access-5n6gw" (OuterVolumeSpecName: "kube-api-access-5n6gw") pod "baa2b0f4-c923-4d74-a112-bf7e5cae275e" (UID: "baa2b0f4-c923-4d74-a112-bf7e5cae275e"). InnerVolumeSpecName "kube-api-access-5n6gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.329344 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n6gw\" (UniqueName: \"kubernetes.io/projected/baa2b0f4-c923-4d74-a112-bf7e5cae275e-kube-api-access-5n6gw\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.329405 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.380527 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "baa2b0f4-c923-4d74-a112-bf7e5cae275e" (UID: "baa2b0f4-c923-4d74-a112-bf7e5cae275e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.431529 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baa2b0f4-c923-4d74-a112-bf7e5cae275e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.590035 4978 generic.go:334] "Generic (PLEG): container finished" podID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerID="cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe" exitCode=0 Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.590165 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mmg7b" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.590150 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmg7b" event={"ID":"baa2b0f4-c923-4d74-a112-bf7e5cae275e","Type":"ContainerDied","Data":"cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe"} Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.590263 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mmg7b" event={"ID":"baa2b0f4-c923-4d74-a112-bf7e5cae275e","Type":"ContainerDied","Data":"876380a0b430b9054869e8abaffa57b9ff3c18221b67b3affb7ddc75cfcfb68c"} Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.590303 4978 scope.go:117] "RemoveContainer" containerID="cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.628673 4978 scope.go:117] "RemoveContainer" containerID="a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.642513 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mmg7b"] Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.658976 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mmg7b"] Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.670417 4978 scope.go:117] "RemoveContainer" containerID="9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.728774 4978 scope.go:117] "RemoveContainer" containerID="cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe" Feb 25 08:56:08 crc kubenswrapper[4978]: E0225 08:56:08.730922 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe\": container with ID starting with cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe not found: ID does not exist" containerID="cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.730987 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe"} err="failed to get container status \"cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe\": rpc error: code = NotFound desc = could not find container \"cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe\": container with ID starting with cf38dc905e8231a2342a6825a91634b5156173f0f6a6318459cba0b84caf93fe not found: ID does not exist" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.731029 4978 scope.go:117] "RemoveContainer" containerID="a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3" Feb 25 08:56:08 crc kubenswrapper[4978]: E0225 08:56:08.732690 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3\": container with ID starting with a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3 not found: ID does not exist" containerID="a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.732788 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3"} err="failed to get container status \"a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3\": rpc error: code = NotFound desc = could not find container \"a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3\": container with ID starting with a7dda40ba7e6c84167d62dcc5afb84ac7594c9b6f23ff40480e0b87e7342bbe3 not found: ID does not exist" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.732857 4978 scope.go:117] "RemoveContainer" containerID="9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e" Feb 25 08:56:08 crc kubenswrapper[4978]: E0225 08:56:08.733638 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e\": container with ID starting with 9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e not found: ID does not exist" containerID="9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e" Feb 25 08:56:08 crc kubenswrapper[4978]: I0225 08:56:08.733707 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e"} err="failed to get container status \"9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e\": rpc error: code = NotFound desc = could not find container \"9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e\": container with ID starting with 9e1d18ee72dac11b4878b71f7f5121855a7f6f747d890fccc3a508e98498ba5e not found: ID does not exist" Feb 25 08:56:09 crc kubenswrapper[4978]: I0225 08:56:09.369795 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" path="/var/lib/kubelet/pods/baa2b0f4-c923-4d74-a112-bf7e5cae275e/volumes" Feb 25 08:56:12 crc kubenswrapper[4978]: I0225 08:56:12.047600 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-sxp4q"] Feb 25 08:56:12 crc kubenswrapper[4978]: I0225 08:56:12.064927 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-sxp4q"] Feb 25 08:56:13 crc kubenswrapper[4978]: I0225 08:56:13.341067 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cee2ed39-9b9f-466d-b2b9-89e3b598e4e5" path="/var/lib/kubelet/pods/cee2ed39-9b9f-466d-b2b9-89e3b598e4e5/volumes" Feb 25 08:56:13 crc kubenswrapper[4978]: I0225 08:56:13.945655 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.066960 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7799fc56d5-cbjnr"] Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.067303 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" podUID="e3d28a89-7825-49d2-bb3d-230edc0a067c" containerName="dnsmasq-dns" containerID="cri-o://9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247" gracePeriod=10 Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.576703 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.683992 4978 generic.go:334] "Generic (PLEG): container finished" podID="e3d28a89-7825-49d2-bb3d-230edc0a067c" containerID="9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247" exitCode=0 Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.684056 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" event={"ID":"e3d28a89-7825-49d2-bb3d-230edc0a067c","Type":"ContainerDied","Data":"9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247"} Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.684093 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" event={"ID":"e3d28a89-7825-49d2-bb3d-230edc0a067c","Type":"ContainerDied","Data":"7875de37d6454cf6996f17f479c64186686823e1042a199ff271822d78b125ec"} Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.684117 4978 scope.go:117] "RemoveContainer" containerID="9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.684312 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7799fc56d5-cbjnr" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.686118 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-sb\") pod \"e3d28a89-7825-49d2-bb3d-230edc0a067c\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.686240 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79rnq\" (UniqueName: \"kubernetes.io/projected/e3d28a89-7825-49d2-bb3d-230edc0a067c-kube-api-access-79rnq\") pod \"e3d28a89-7825-49d2-bb3d-230edc0a067c\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.686420 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-cell1\") pod \"e3d28a89-7825-49d2-bb3d-230edc0a067c\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.686599 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-networker\") pod \"e3d28a89-7825-49d2-bb3d-230edc0a067c\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.686642 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-dns-svc\") pod \"e3d28a89-7825-49d2-bb3d-230edc0a067c\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.686659 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-nb\") pod \"e3d28a89-7825-49d2-bb3d-230edc0a067c\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.686704 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-config\") pod \"e3d28a89-7825-49d2-bb3d-230edc0a067c\" (UID: \"e3d28a89-7825-49d2-bb3d-230edc0a067c\") " Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.778508 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3d28a89-7825-49d2-bb3d-230edc0a067c-kube-api-access-79rnq" (OuterVolumeSpecName: "kube-api-access-79rnq") pod "e3d28a89-7825-49d2-bb3d-230edc0a067c" (UID: "e3d28a89-7825-49d2-bb3d-230edc0a067c"). InnerVolumeSpecName "kube-api-access-79rnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.790153 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79rnq\" (UniqueName: \"kubernetes.io/projected/e3d28a89-7825-49d2-bb3d-230edc0a067c-kube-api-access-79rnq\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.818318 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-networker" (OuterVolumeSpecName: "openstack-networker") pod "e3d28a89-7825-49d2-bb3d-230edc0a067c" (UID: "e3d28a89-7825-49d2-bb3d-230edc0a067c"). InnerVolumeSpecName "openstack-networker". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.861674 4978 scope.go:117] "RemoveContainer" containerID="9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.861993 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e3d28a89-7825-49d2-bb3d-230edc0a067c" (UID: "e3d28a89-7825-49d2-bb3d-230edc0a067c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.862179 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "e3d28a89-7825-49d2-bb3d-230edc0a067c" (UID: "e3d28a89-7825-49d2-bb3d-230edc0a067c"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.891892 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.891917 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.891925 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.922080 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-config" (OuterVolumeSpecName: "config") pod "e3d28a89-7825-49d2-bb3d-230edc0a067c" (UID: "e3d28a89-7825-49d2-bb3d-230edc0a067c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.922690 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e3d28a89-7825-49d2-bb3d-230edc0a067c" (UID: "e3d28a89-7825-49d2-bb3d-230edc0a067c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.936725 4978 scope.go:117] "RemoveContainer" containerID="9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247" Feb 25 08:56:14 crc kubenswrapper[4978]: E0225 08:56:14.938493 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247\": container with ID starting with 9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247 not found: ID does not exist" containerID="9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.938544 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247"} err="failed to get container status \"9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247\": rpc error: code = NotFound desc = could not find container \"9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247\": container with ID starting with 9210a2497e81e95d82ab4299c38ff99f02dc443d2237b2ef0ee3da2eb5b36247 not found: ID does not exist" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.938573 4978 scope.go:117] "RemoveContainer" containerID="9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2" Feb 25 08:56:14 crc kubenswrapper[4978]: E0225 08:56:14.939047 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2\": container with ID starting with 9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2 not found: ID does not exist" containerID="9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.939089 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2"} err="failed to get container status \"9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2\": rpc error: code = NotFound desc = could not find container \"9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2\": container with ID starting with 9fe6f35de8fb1bcb0ddc4eca122cebec45b6aa6539bf2c264650bc733f16d2f2 not found: ID does not exist" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.943808 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e3d28a89-7825-49d2-bb3d-230edc0a067c" (UID: "e3d28a89-7825-49d2-bb3d-230edc0a067c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.993778 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.993820 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:14 crc kubenswrapper[4978]: I0225 08:56:14.993836 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3d28a89-7825-49d2-bb3d-230edc0a067c-config\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:15 crc kubenswrapper[4978]: I0225 08:56:15.016891 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7799fc56d5-cbjnr"] Feb 25 08:56:15 crc kubenswrapper[4978]: I0225 08:56:15.025996 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7799fc56d5-cbjnr"] Feb 25 08:56:15 crc kubenswrapper[4978]: I0225 08:56:15.337463 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3d28a89-7825-49d2-bb3d-230edc0a067c" path="/var/lib/kubelet/pods/e3d28a89-7825-49d2-bb3d-230edc0a067c/volumes" Feb 25 08:56:17 crc kubenswrapper[4978]: I0225 08:56:17.339746 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:56:17 crc kubenswrapper[4978]: E0225 08:56:17.340567 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.339202 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph"] Feb 25 08:56:28 crc kubenswrapper[4978]: E0225 08:56:28.340509 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8867ff4e-e020-4d2c-a851-40279f46b8fa" containerName="init" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.340525 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8867ff4e-e020-4d2c-a851-40279f46b8fa" containerName="init" Feb 25 08:56:28 crc kubenswrapper[4978]: E0225 08:56:28.340560 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8867ff4e-e020-4d2c-a851-40279f46b8fa" containerName="dnsmasq-dns" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.340569 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8867ff4e-e020-4d2c-a851-40279f46b8fa" containerName="dnsmasq-dns" Feb 25 08:56:28 crc kubenswrapper[4978]: E0225 08:56:28.340584 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72a0184c-d75c-4ecb-a38a-d793e3c7da1d" containerName="oc" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.340592 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="72a0184c-d75c-4ecb-a38a-d793e3c7da1d" containerName="oc" Feb 25 08:56:28 crc kubenswrapper[4978]: E0225 08:56:28.340611 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d28a89-7825-49d2-bb3d-230edc0a067c" containerName="init" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.340619 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d28a89-7825-49d2-bb3d-230edc0a067c" containerName="init" Feb 25 08:56:28 crc kubenswrapper[4978]: E0225 08:56:28.340633 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerName="extract-utilities" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.340646 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerName="extract-utilities" Feb 25 08:56:28 crc kubenswrapper[4978]: E0225 08:56:28.340668 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerName="extract-content" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.340678 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerName="extract-content" Feb 25 08:56:28 crc kubenswrapper[4978]: E0225 08:56:28.340696 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d28a89-7825-49d2-bb3d-230edc0a067c" containerName="dnsmasq-dns" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.340705 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d28a89-7825-49d2-bb3d-230edc0a067c" containerName="dnsmasq-dns" Feb 25 08:56:28 crc kubenswrapper[4978]: E0225 08:56:28.340748 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerName="registry-server" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.340756 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerName="registry-server" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.341157 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="baa2b0f4-c923-4d74-a112-bf7e5cae275e" containerName="registry-server" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.341193 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8867ff4e-e020-4d2c-a851-40279f46b8fa" containerName="dnsmasq-dns" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.341207 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3d28a89-7825-49d2-bb3d-230edc0a067c" containerName="dnsmasq-dns" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.341255 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="72a0184c-d75c-4ecb-a38a-d793e3c7da1d" containerName="oc" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.347696 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.352259 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.352669 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.352819 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.363512 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.365477 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l"] Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.368169 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.370972 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.371297 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.386240 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l"] Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.407548 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph"] Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.502177 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.502536 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.502632 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktjkn\" (UniqueName: \"kubernetes.io/projected/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-kube-api-access-ktjkn\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.502668 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.502691 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-ssh-key-openstack-networker\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.502738 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdhjw\" (UniqueName: \"kubernetes.io/projected/5deab02b-db5d-47f5-921b-a5b35412877a-kube-api-access-gdhjw\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.502872 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.502896 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.604922 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.605002 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-ssh-key-openstack-networker\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.605084 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdhjw\" (UniqueName: \"kubernetes.io/projected/5deab02b-db5d-47f5-921b-a5b35412877a-kube-api-access-gdhjw\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.605266 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.605315 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.605366 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.605479 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.605529 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktjkn\" (UniqueName: \"kubernetes.io/projected/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-kube-api-access-ktjkn\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.610681 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.611302 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-ssh-key-openstack-networker\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.618706 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.618882 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.619215 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.620789 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.623977 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdhjw\" (UniqueName: \"kubernetes.io/projected/5deab02b-db5d-47f5-921b-a5b35412877a-kube-api-access-gdhjw\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.637638 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktjkn\" (UniqueName: \"kubernetes.io/projected/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-kube-api-access-ktjkn\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cpndph\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.688907 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:28 crc kubenswrapper[4978]: I0225 08:56:28.699870 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:29 crc kubenswrapper[4978]: I0225 08:56:29.609548 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l"] Feb 25 08:56:29 crc kubenswrapper[4978]: I0225 08:56:29.867706 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" event={"ID":"5deab02b-db5d-47f5-921b-a5b35412877a","Type":"ContainerStarted","Data":"028358d14f51e4c94e9fde6576440292a87d3d410fefc6c45e982f4434118222"} Feb 25 08:56:30 crc kubenswrapper[4978]: I0225 08:56:30.308751 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph"] Feb 25 08:56:30 crc kubenswrapper[4978]: I0225 08:56:30.328393 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:56:30 crc kubenswrapper[4978]: E0225 08:56:30.328718 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:56:30 crc kubenswrapper[4978]: I0225 08:56:30.878825 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" event={"ID":"a37f3a9e-cb0d-47f9-8cec-de20cf40a874","Type":"ContainerStarted","Data":"bfccce2af7c4fc6122b758720d04e3354988219d3ff9a5bf48e37ddd493f251d"} Feb 25 08:56:31 crc kubenswrapper[4978]: I0225 08:56:31.149261 4978 scope.go:117] "RemoveContainer" containerID="2763203f1c861e0380f198a1b4d21b94d8355638ad6da04ff4e4a1c1a93ae4a3" Feb 25 08:56:31 crc kubenswrapper[4978]: I0225 08:56:31.174000 4978 scope.go:117] "RemoveContainer" containerID="2db2b82e8d8653b420ee35ddb3dc7d8e6967bc6044563849bb6e8b09eb8d2af0" Feb 25 08:56:31 crc kubenswrapper[4978]: I0225 08:56:31.237482 4978 scope.go:117] "RemoveContainer" containerID="225b0c9dd1125c7894b6a0800aae02d676551673a00a78b8b17ba28d18dca189" Feb 25 08:56:32 crc kubenswrapper[4978]: I0225 08:56:32.721301 4978 scope.go:117] "RemoveContainer" containerID="d04b76696689d976c05d7705c0608c9032fe21d53f4f9a2089fa60ffec5b251a" Feb 25 08:56:40 crc kubenswrapper[4978]: I0225 08:56:40.974561 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" event={"ID":"a37f3a9e-cb0d-47f9-8cec-de20cf40a874","Type":"ContainerStarted","Data":"c1eaee7730197e31cc67e1bacdc3cccc679439aec4e78a5a007d37fa00048a71"} Feb 25 08:56:40 crc kubenswrapper[4978]: I0225 08:56:40.976937 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" event={"ID":"5deab02b-db5d-47f5-921b-a5b35412877a","Type":"ContainerStarted","Data":"bdb2e339142e8cf9c08428206bb7a363b0c54feef6c2c3e446556d28645a1367"} Feb 25 08:56:40 crc kubenswrapper[4978]: I0225 08:56:40.999679 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" podStartSLOduration=3.482500048 podStartE2EDuration="12.999663758s" podCreationTimestamp="2026-02-25 08:56:28 +0000 UTC" firstStartedPulling="2026-02-25 08:56:30.309141736 +0000 UTC m=+7883.748398185" lastFinishedPulling="2026-02-25 08:56:39.826305396 +0000 UTC m=+7893.265561895" observedRunningTime="2026-02-25 08:56:40.996959314 +0000 UTC m=+7894.436215773" watchObservedRunningTime="2026-02-25 08:56:40.999663758 +0000 UTC m=+7894.438920217" Feb 25 08:56:41 crc kubenswrapper[4978]: I0225 08:56:41.032430 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" podStartSLOduration=2.830983943 podStartE2EDuration="13.032408762s" podCreationTimestamp="2026-02-25 08:56:28 +0000 UTC" firstStartedPulling="2026-02-25 08:56:29.615315661 +0000 UTC m=+7883.054572130" lastFinishedPulling="2026-02-25 08:56:39.81674049 +0000 UTC m=+7893.255996949" observedRunningTime="2026-02-25 08:56:41.021800433 +0000 UTC m=+7894.461056892" watchObservedRunningTime="2026-02-25 08:56:41.032408762 +0000 UTC m=+7894.471665241" Feb 25 08:56:41 crc kubenswrapper[4978]: I0225 08:56:41.065226 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-acdf-account-create-update-wfk78"] Feb 25 08:56:41 crc kubenswrapper[4978]: I0225 08:56:41.075167 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-acdf-account-create-update-wfk78"] Feb 25 08:56:41 crc kubenswrapper[4978]: I0225 08:56:41.085419 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-smtfd"] Feb 25 08:56:41 crc kubenswrapper[4978]: I0225 08:56:41.100525 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-smtfd"] Feb 25 08:56:41 crc kubenswrapper[4978]: I0225 08:56:41.327725 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:56:41 crc kubenswrapper[4978]: E0225 08:56:41.328174 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:56:41 crc kubenswrapper[4978]: I0225 08:56:41.338146 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0cb40be-d13e-4d76-a41f-ac97096d1d44" path="/var/lib/kubelet/pods/e0cb40be-d13e-4d76-a41f-ac97096d1d44/volumes" Feb 25 08:56:41 crc kubenswrapper[4978]: I0225 08:56:41.338857 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f" path="/var/lib/kubelet/pods/e7c760da-bbaf-4b92-8b90-5b1b99ed9d7f/volumes" Feb 25 08:56:50 crc kubenswrapper[4978]: E0225 08:56:50.048106 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5deab02b_db5d_47f5_921b_a5b35412877a.slice/crio-conmon-bdb2e339142e8cf9c08428206bb7a363b0c54feef6c2c3e446556d28645a1367.scope\": RecentStats: unable to find data in memory cache]" Feb 25 08:56:50 crc kubenswrapper[4978]: I0225 08:56:50.073199 4978 generic.go:334] "Generic (PLEG): container finished" podID="5deab02b-db5d-47f5-921b-a5b35412877a" containerID="bdb2e339142e8cf9c08428206bb7a363b0c54feef6c2c3e446556d28645a1367" exitCode=0 Feb 25 08:56:50 crc kubenswrapper[4978]: I0225 08:56:50.073253 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" event={"ID":"5deab02b-db5d-47f5-921b-a5b35412877a","Type":"ContainerDied","Data":"bdb2e339142e8cf9c08428206bb7a363b0c54feef6c2c3e446556d28645a1367"} Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.035529 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-dhmxq"] Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.050951 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-dhmxq"] Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.343551 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d009f647-5ba2-4877-b775-47ecd94b3e6f" path="/var/lib/kubelet/pods/d009f647-5ba2-4877-b775-47ecd94b3e6f/volumes" Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.581280 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.720759 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdhjw\" (UniqueName: \"kubernetes.io/projected/5deab02b-db5d-47f5-921b-a5b35412877a-kube-api-access-gdhjw\") pod \"5deab02b-db5d-47f5-921b-a5b35412877a\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.720881 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-pre-adoption-validation-combined-ca-bundle\") pod \"5deab02b-db5d-47f5-921b-a5b35412877a\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.721130 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-inventory\") pod \"5deab02b-db5d-47f5-921b-a5b35412877a\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.721203 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-ssh-key-openstack-networker\") pod \"5deab02b-db5d-47f5-921b-a5b35412877a\" (UID: \"5deab02b-db5d-47f5-921b-a5b35412877a\") " Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.726427 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5deab02b-db5d-47f5-921b-a5b35412877a-kube-api-access-gdhjw" (OuterVolumeSpecName: "kube-api-access-gdhjw") pod "5deab02b-db5d-47f5-921b-a5b35412877a" (UID: "5deab02b-db5d-47f5-921b-a5b35412877a"). InnerVolumeSpecName "kube-api-access-gdhjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.729472 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "5deab02b-db5d-47f5-921b-a5b35412877a" (UID: "5deab02b-db5d-47f5-921b-a5b35412877a"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.755215 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-inventory" (OuterVolumeSpecName: "inventory") pod "5deab02b-db5d-47f5-921b-a5b35412877a" (UID: "5deab02b-db5d-47f5-921b-a5b35412877a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.775449 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "5deab02b-db5d-47f5-921b-a5b35412877a" (UID: "5deab02b-db5d-47f5-921b-a5b35412877a"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.823566 4978 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.823601 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.823612 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/5deab02b-db5d-47f5-921b-a5b35412877a-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:51 crc kubenswrapper[4978]: I0225 08:56:51.823622 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdhjw\" (UniqueName: \"kubernetes.io/projected/5deab02b-db5d-47f5-921b-a5b35412877a-kube-api-access-gdhjw\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:52 crc kubenswrapper[4978]: I0225 08:56:52.097856 4978 generic.go:334] "Generic (PLEG): container finished" podID="a37f3a9e-cb0d-47f9-8cec-de20cf40a874" containerID="c1eaee7730197e31cc67e1bacdc3cccc679439aec4e78a5a007d37fa00048a71" exitCode=0 Feb 25 08:56:52 crc kubenswrapper[4978]: I0225 08:56:52.097988 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" event={"ID":"a37f3a9e-cb0d-47f9-8cec-de20cf40a874","Type":"ContainerDied","Data":"c1eaee7730197e31cc67e1bacdc3cccc679439aec4e78a5a007d37fa00048a71"} Feb 25 08:56:52 crc kubenswrapper[4978]: I0225 08:56:52.100635 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" event={"ID":"5deab02b-db5d-47f5-921b-a5b35412877a","Type":"ContainerDied","Data":"028358d14f51e4c94e9fde6576440292a87d3d410fefc6c45e982f4434118222"} Feb 25 08:56:52 crc kubenswrapper[4978]: I0225 08:56:52.100674 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="028358d14f51e4c94e9fde6576440292a87d3d410fefc6c45e982f4434118222" Feb 25 08:56:52 crc kubenswrapper[4978]: I0225 08:56:52.100723 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l" Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.633527 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.763219 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-inventory\") pod \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.763362 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktjkn\" (UniqueName: \"kubernetes.io/projected/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-kube-api-access-ktjkn\") pod \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.763673 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-ssh-key-openstack-cell1\") pod \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.763775 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-pre-adoption-validation-combined-ca-bundle\") pod \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\" (UID: \"a37f3a9e-cb0d-47f9-8cec-de20cf40a874\") " Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.770742 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "a37f3a9e-cb0d-47f9-8cec-de20cf40a874" (UID: "a37f3a9e-cb0d-47f9-8cec-de20cf40a874"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.772287 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-kube-api-access-ktjkn" (OuterVolumeSpecName: "kube-api-access-ktjkn") pod "a37f3a9e-cb0d-47f9-8cec-de20cf40a874" (UID: "a37f3a9e-cb0d-47f9-8cec-de20cf40a874"). InnerVolumeSpecName "kube-api-access-ktjkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.793991 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-inventory" (OuterVolumeSpecName: "inventory") pod "a37f3a9e-cb0d-47f9-8cec-de20cf40a874" (UID: "a37f3a9e-cb0d-47f9-8cec-de20cf40a874"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.824841 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "a37f3a9e-cb0d-47f9-8cec-de20cf40a874" (UID: "a37f3a9e-cb0d-47f9-8cec-de20cf40a874"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.866908 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.866965 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktjkn\" (UniqueName: \"kubernetes.io/projected/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-kube-api-access-ktjkn\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.866988 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:53 crc kubenswrapper[4978]: I0225 08:56:53.867008 4978 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a37f3a9e-cb0d-47f9-8cec-de20cf40a874-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 08:56:54 crc kubenswrapper[4978]: I0225 08:56:54.133746 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" event={"ID":"a37f3a9e-cb0d-47f9-8cec-de20cf40a874","Type":"ContainerDied","Data":"bfccce2af7c4fc6122b758720d04e3354988219d3ff9a5bf48e37ddd493f251d"} Feb 25 08:56:54 crc kubenswrapper[4978]: I0225 08:56:54.133788 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfccce2af7c4fc6122b758720d04e3354988219d3ff9a5bf48e37ddd493f251d" Feb 25 08:56:54 crc kubenswrapper[4978]: I0225 08:56:54.133801 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cpndph" Feb 25 08:56:55 crc kubenswrapper[4978]: I0225 08:56:55.328132 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:56:55 crc kubenswrapper[4978]: E0225 08:56:55.328563 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.598923 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm"] Feb 25 08:56:56 crc kubenswrapper[4978]: E0225 08:56:56.599867 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5deab02b-db5d-47f5-921b-a5b35412877a" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.599889 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5deab02b-db5d-47f5-921b-a5b35412877a" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Feb 25 08:56:56 crc kubenswrapper[4978]: E0225 08:56:56.599909 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a37f3a9e-cb0d-47f9-8cec-de20cf40a874" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.599918 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a37f3a9e-cb0d-47f9-8cec-de20cf40a874" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.600165 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5deab02b-db5d-47f5-921b-a5b35412877a" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-networ" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.600204 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a37f3a9e-cb0d-47f9-8cec-de20cf40a874" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.601270 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.604260 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.604878 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.608540 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.609824 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.617234 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8"] Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.618982 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.622036 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.622481 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.639196 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm"] Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.639266 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8"] Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.728954 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.729022 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.729055 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.729188 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.729262 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppgzr\" (UniqueName: \"kubernetes.io/projected/6abcdd51-69ae-42d5-9708-7088d9911458-kube-api-access-ppgzr\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.729665 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-ssh-key-openstack-networker\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.729757 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.729840 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2zzj\" (UniqueName: \"kubernetes.io/projected/6153aa3b-10fb-4bba-8103-71d3f8347d0d-kube-api-access-k2zzj\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.831785 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2zzj\" (UniqueName: \"kubernetes.io/projected/6153aa3b-10fb-4bba-8103-71d3f8347d0d-kube-api-access-k2zzj\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.831846 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.831891 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.831915 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.831942 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.831961 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppgzr\" (UniqueName: \"kubernetes.io/projected/6abcdd51-69ae-42d5-9708-7088d9911458-kube-api-access-ppgzr\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.832037 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-ssh-key-openstack-networker\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.832082 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.844648 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.844688 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-ssh-key-openstack-networker\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.844919 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.844726 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.845150 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.845154 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.850197 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2zzj\" (UniqueName: \"kubernetes.io/projected/6153aa3b-10fb-4bba-8103-71d3f8347d0d-kube-api-access-k2zzj\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.851628 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppgzr\" (UniqueName: \"kubernetes.io/projected/6abcdd51-69ae-42d5-9708-7088d9911458-kube-api-access-ppgzr\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.965707 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 08:56:56 crc kubenswrapper[4978]: I0225 08:56:56.981630 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 08:56:57 crc kubenswrapper[4978]: I0225 08:56:57.589893 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm"] Feb 25 08:56:57 crc kubenswrapper[4978]: I0225 08:56:57.734705 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8"] Feb 25 08:56:58 crc kubenswrapper[4978]: I0225 08:56:58.178135 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" event={"ID":"6153aa3b-10fb-4bba-8103-71d3f8347d0d","Type":"ContainerStarted","Data":"d363620a7a3ceb07796ecd2e17132b42a4fe7c7faddd7b8f59b5a66d49c2fb03"} Feb 25 08:56:58 crc kubenswrapper[4978]: I0225 08:56:58.179758 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" event={"ID":"6abcdd51-69ae-42d5-9708-7088d9911458","Type":"ContainerStarted","Data":"1aecf0a41564322e277ec855264a26acb40ad03cfb9d1c7b40e9c5f0f7c5102c"} Feb 25 08:56:59 crc kubenswrapper[4978]: I0225 08:56:59.194702 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" event={"ID":"6abcdd51-69ae-42d5-9708-7088d9911458","Type":"ContainerStarted","Data":"85656ef656077c83751a5c3845c4f90e2f262eaaca42c0427a37198d34b9c5ff"} Feb 25 08:56:59 crc kubenswrapper[4978]: I0225 08:56:59.196897 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" event={"ID":"6153aa3b-10fb-4bba-8103-71d3f8347d0d","Type":"ContainerStarted","Data":"2fc601d65c08af1bb46c5b9e8e89c223bb2165d8d4c611fb499b83c6d256e7a3"} Feb 25 08:56:59 crc kubenswrapper[4978]: I0225 08:56:59.238266 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" podStartSLOduration=2.807728597 podStartE2EDuration="3.238240345s" podCreationTimestamp="2026-02-25 08:56:56 +0000 UTC" firstStartedPulling="2026-02-25 08:56:57.745746736 +0000 UTC m=+7911.185003195" lastFinishedPulling="2026-02-25 08:56:58.176258474 +0000 UTC m=+7911.615514943" observedRunningTime="2026-02-25 08:56:59.218159382 +0000 UTC m=+7912.657415881" watchObservedRunningTime="2026-02-25 08:56:59.238240345 +0000 UTC m=+7912.677496814" Feb 25 08:56:59 crc kubenswrapper[4978]: I0225 08:56:59.256962 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" podStartSLOduration=2.816074876 podStartE2EDuration="3.256942685s" podCreationTimestamp="2026-02-25 08:56:56 +0000 UTC" firstStartedPulling="2026-02-25 08:56:57.597694349 +0000 UTC m=+7911.036950848" lastFinishedPulling="2026-02-25 08:56:58.038562188 +0000 UTC m=+7911.477818657" observedRunningTime="2026-02-25 08:56:59.238795042 +0000 UTC m=+7912.678051541" watchObservedRunningTime="2026-02-25 08:56:59.256942685 +0000 UTC m=+7912.696199144" Feb 25 08:57:09 crc kubenswrapper[4978]: I0225 08:57:09.328830 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:57:09 crc kubenswrapper[4978]: E0225 08:57:09.330142 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:57:21 crc kubenswrapper[4978]: I0225 08:57:21.327736 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:57:21 crc kubenswrapper[4978]: E0225 08:57:21.328633 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:57:32 crc kubenswrapper[4978]: I0225 08:57:32.921615 4978 scope.go:117] "RemoveContainer" containerID="3a0280a607543b4588fc74f488860c3ac84a29d25c45ef37d747296342df7091" Feb 25 08:57:32 crc kubenswrapper[4978]: I0225 08:57:32.948901 4978 scope.go:117] "RemoveContainer" containerID="faa24b9256a8f0a12e21502a1ca50a38220d731f994ae9b0d8a028a6765261de" Feb 25 08:57:33 crc kubenswrapper[4978]: I0225 08:57:33.022664 4978 scope.go:117] "RemoveContainer" containerID="389ff171ff2aab135beef50dd73fbd983bd61c38428fc1c39575fdcc7c58769b" Feb 25 08:57:35 crc kubenswrapper[4978]: I0225 08:57:35.328699 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:57:35 crc kubenswrapper[4978]: E0225 08:57:35.329630 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:57:46 crc kubenswrapper[4978]: I0225 08:57:46.327474 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:57:46 crc kubenswrapper[4978]: E0225 08:57:46.328330 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:57:51 crc kubenswrapper[4978]: I0225 08:57:51.054650 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8d1c-account-create-update-zhr8r"] Feb 25 08:57:51 crc kubenswrapper[4978]: I0225 08:57:51.084636 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8d1c-account-create-update-zhr8r"] Feb 25 08:57:51 crc kubenswrapper[4978]: I0225 08:57:51.347041 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7cf88a7-6d20-4909-a39a-08fbe51bd45a" path="/var/lib/kubelet/pods/a7cf88a7-6d20-4909-a39a-08fbe51bd45a/volumes" Feb 25 08:57:51 crc kubenswrapper[4978]: I0225 08:57:51.742257 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8msh6"] Feb 25 08:57:51 crc kubenswrapper[4978]: I0225 08:57:51.745036 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:51 crc kubenswrapper[4978]: I0225 08:57:51.752053 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8msh6"] Feb 25 08:57:51 crc kubenswrapper[4978]: I0225 08:57:51.909810 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-catalog-content\") pod \"certified-operators-8msh6\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:51 crc kubenswrapper[4978]: I0225 08:57:51.910139 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-utilities\") pod \"certified-operators-8msh6\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:51 crc kubenswrapper[4978]: I0225 08:57:51.910227 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxzkz\" (UniqueName: \"kubernetes.io/projected/4bdecaab-9b43-474e-8dbf-b0adeb06015a-kube-api-access-wxzkz\") pod \"certified-operators-8msh6\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.011626 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-catalog-content\") pod \"certified-operators-8msh6\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.011729 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-utilities\") pod \"certified-operators-8msh6\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.011747 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxzkz\" (UniqueName: \"kubernetes.io/projected/4bdecaab-9b43-474e-8dbf-b0adeb06015a-kube-api-access-wxzkz\") pod \"certified-operators-8msh6\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.012489 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-utilities\") pod \"certified-operators-8msh6\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.012822 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-catalog-content\") pod \"certified-operators-8msh6\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.031993 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-ww92q"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.040198 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-ww92q"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.050355 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-bf46-account-create-update-v8jtg"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.050412 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxzkz\" (UniqueName: \"kubernetes.io/projected/4bdecaab-9b43-474e-8dbf-b0adeb06015a-kube-api-access-wxzkz\") pod \"certified-operators-8msh6\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.061020 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-r7rr7"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.071629 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-d46hl"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.072332 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.080171 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ce19-account-create-update-99rrk"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.089108 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ce19-account-create-update-99rrk"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.098106 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-bf46-account-create-update-v8jtg"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.105707 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-r7rr7"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.113420 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-d46hl"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.532574 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8msh6"] Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.847448 4978 generic.go:334] "Generic (PLEG): container finished" podID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerID="0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213" exitCode=0 Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.847488 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8msh6" event={"ID":"4bdecaab-9b43-474e-8dbf-b0adeb06015a","Type":"ContainerDied","Data":"0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213"} Feb 25 08:57:52 crc kubenswrapper[4978]: I0225 08:57:52.847512 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8msh6" event={"ID":"4bdecaab-9b43-474e-8dbf-b0adeb06015a","Type":"ContainerStarted","Data":"5636d57ddbc46286f33801341c61d2f3d6e83af24989c17e6c1b6117324be8f8"} Feb 25 08:57:53 crc kubenswrapper[4978]: I0225 08:57:53.360636 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f36fcfd-2834-4e66-ad11-16d96d2cbf02" path="/var/lib/kubelet/pods/0f36fcfd-2834-4e66-ad11-16d96d2cbf02/volumes" Feb 25 08:57:53 crc kubenswrapper[4978]: I0225 08:57:53.362070 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="105c1f03-d5a4-43d3-93d7-ce9682adddab" path="/var/lib/kubelet/pods/105c1f03-d5a4-43d3-93d7-ce9682adddab/volumes" Feb 25 08:57:53 crc kubenswrapper[4978]: I0225 08:57:53.362939 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14ebdf50-6a9b-486a-81b7-9d1738e47101" path="/var/lib/kubelet/pods/14ebdf50-6a9b-486a-81b7-9d1738e47101/volumes" Feb 25 08:57:53 crc kubenswrapper[4978]: I0225 08:57:53.363809 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4365133f-7067-4dd7-9ef3-77d8c1312e89" path="/var/lib/kubelet/pods/4365133f-7067-4dd7-9ef3-77d8c1312e89/volumes" Feb 25 08:57:53 crc kubenswrapper[4978]: I0225 08:57:53.365027 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dadb5f9-c6c6-43e0-98a2-917dd5748656" path="/var/lib/kubelet/pods/4dadb5f9-c6c6-43e0-98a2-917dd5748656/volumes" Feb 25 08:57:54 crc kubenswrapper[4978]: I0225 08:57:54.876392 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8msh6" event={"ID":"4bdecaab-9b43-474e-8dbf-b0adeb06015a","Type":"ContainerStarted","Data":"487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082"} Feb 25 08:57:55 crc kubenswrapper[4978]: I0225 08:57:55.896840 4978 generic.go:334] "Generic (PLEG): container finished" podID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerID="487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082" exitCode=0 Feb 25 08:57:55 crc kubenswrapper[4978]: I0225 08:57:55.896940 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8msh6" event={"ID":"4bdecaab-9b43-474e-8dbf-b0adeb06015a","Type":"ContainerDied","Data":"487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082"} Feb 25 08:57:56 crc kubenswrapper[4978]: I0225 08:57:56.911688 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8msh6" event={"ID":"4bdecaab-9b43-474e-8dbf-b0adeb06015a","Type":"ContainerStarted","Data":"7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81"} Feb 25 08:57:56 crc kubenswrapper[4978]: I0225 08:57:56.937293 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8msh6" podStartSLOduration=2.454319177 podStartE2EDuration="5.937273262s" podCreationTimestamp="2026-02-25 08:57:51 +0000 UTC" firstStartedPulling="2026-02-25 08:57:52.849426267 +0000 UTC m=+7966.288682766" lastFinishedPulling="2026-02-25 08:57:56.332380392 +0000 UTC m=+7969.771636851" observedRunningTime="2026-02-25 08:57:56.930767751 +0000 UTC m=+7970.370024220" watchObservedRunningTime="2026-02-25 08:57:56.937273262 +0000 UTC m=+7970.376529721" Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.191129 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533498-rks4h"] Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.194745 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533498-rks4h" Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.199296 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.199333 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.199601 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.205758 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533498-rks4h"] Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.309700 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtngb\" (UniqueName: \"kubernetes.io/projected/7a921f4a-dc79-4d25-befd-cd15da7fa69e-kube-api-access-qtngb\") pod \"auto-csr-approver-29533498-rks4h\" (UID: \"7a921f4a-dc79-4d25-befd-cd15da7fa69e\") " pod="openshift-infra/auto-csr-approver-29533498-rks4h" Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.328584 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:58:00 crc kubenswrapper[4978]: E0225 08:58:00.329191 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.412090 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtngb\" (UniqueName: \"kubernetes.io/projected/7a921f4a-dc79-4d25-befd-cd15da7fa69e-kube-api-access-qtngb\") pod \"auto-csr-approver-29533498-rks4h\" (UID: \"7a921f4a-dc79-4d25-befd-cd15da7fa69e\") " pod="openshift-infra/auto-csr-approver-29533498-rks4h" Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.439682 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtngb\" (UniqueName: \"kubernetes.io/projected/7a921f4a-dc79-4d25-befd-cd15da7fa69e-kube-api-access-qtngb\") pod \"auto-csr-approver-29533498-rks4h\" (UID: \"7a921f4a-dc79-4d25-befd-cd15da7fa69e\") " pod="openshift-infra/auto-csr-approver-29533498-rks4h" Feb 25 08:58:00 crc kubenswrapper[4978]: I0225 08:58:00.528529 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533498-rks4h" Feb 25 08:58:01 crc kubenswrapper[4978]: I0225 08:58:01.070140 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533498-rks4h"] Feb 25 08:58:01 crc kubenswrapper[4978]: I0225 08:58:01.984228 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533498-rks4h" event={"ID":"7a921f4a-dc79-4d25-befd-cd15da7fa69e","Type":"ContainerStarted","Data":"a4825017544b37487f993d70dd403d12b688eb671fcbc5061c75ebc09fd089ab"} Feb 25 08:58:02 crc kubenswrapper[4978]: I0225 08:58:02.072654 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:58:02 crc kubenswrapper[4978]: I0225 08:58:02.072710 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:58:02 crc kubenswrapper[4978]: I0225 08:58:02.142258 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:58:03 crc kubenswrapper[4978]: I0225 08:58:03.001095 4978 generic.go:334] "Generic (PLEG): container finished" podID="7a921f4a-dc79-4d25-befd-cd15da7fa69e" containerID="711d8657ef40f52e0b9ad94dd20d364d4a896473ef84ed5755b65a5e7ddc27b8" exitCode=0 Feb 25 08:58:03 crc kubenswrapper[4978]: I0225 08:58:03.001204 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533498-rks4h" event={"ID":"7a921f4a-dc79-4d25-befd-cd15da7fa69e","Type":"ContainerDied","Data":"711d8657ef40f52e0b9ad94dd20d364d4a896473ef84ed5755b65a5e7ddc27b8"} Feb 25 08:58:03 crc kubenswrapper[4978]: I0225 08:58:03.090044 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:58:03 crc kubenswrapper[4978]: I0225 08:58:03.165016 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8msh6"] Feb 25 08:58:04 crc kubenswrapper[4978]: I0225 08:58:04.394501 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533498-rks4h" Feb 25 08:58:04 crc kubenswrapper[4978]: I0225 08:58:04.514606 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtngb\" (UniqueName: \"kubernetes.io/projected/7a921f4a-dc79-4d25-befd-cd15da7fa69e-kube-api-access-qtngb\") pod \"7a921f4a-dc79-4d25-befd-cd15da7fa69e\" (UID: \"7a921f4a-dc79-4d25-befd-cd15da7fa69e\") " Feb 25 08:58:04 crc kubenswrapper[4978]: I0225 08:58:04.524043 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a921f4a-dc79-4d25-befd-cd15da7fa69e-kube-api-access-qtngb" (OuterVolumeSpecName: "kube-api-access-qtngb") pod "7a921f4a-dc79-4d25-befd-cd15da7fa69e" (UID: "7a921f4a-dc79-4d25-befd-cd15da7fa69e"). InnerVolumeSpecName "kube-api-access-qtngb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:58:04 crc kubenswrapper[4978]: I0225 08:58:04.617301 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtngb\" (UniqueName: \"kubernetes.io/projected/7a921f4a-dc79-4d25-befd-cd15da7fa69e-kube-api-access-qtngb\") on node \"crc\" DevicePath \"\"" Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.024645 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533498-rks4h" event={"ID":"7a921f4a-dc79-4d25-befd-cd15da7fa69e","Type":"ContainerDied","Data":"a4825017544b37487f993d70dd403d12b688eb671fcbc5061c75ebc09fd089ab"} Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.024707 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4825017544b37487f993d70dd403d12b688eb671fcbc5061c75ebc09fd089ab" Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.024750 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8msh6" podUID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerName="registry-server" containerID="cri-o://7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81" gracePeriod=2 Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.024799 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533498-rks4h" Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.467489 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533492-bpxgq"] Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.477792 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533492-bpxgq"] Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.544976 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.643536 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-utilities\") pod \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.643740 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-catalog-content\") pod \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.643799 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxzkz\" (UniqueName: \"kubernetes.io/projected/4bdecaab-9b43-474e-8dbf-b0adeb06015a-kube-api-access-wxzkz\") pod \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\" (UID: \"4bdecaab-9b43-474e-8dbf-b0adeb06015a\") " Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.644596 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-utilities" (OuterVolumeSpecName: "utilities") pod "4bdecaab-9b43-474e-8dbf-b0adeb06015a" (UID: "4bdecaab-9b43-474e-8dbf-b0adeb06015a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.648815 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bdecaab-9b43-474e-8dbf-b0adeb06015a-kube-api-access-wxzkz" (OuterVolumeSpecName: "kube-api-access-wxzkz") pod "4bdecaab-9b43-474e-8dbf-b0adeb06015a" (UID: "4bdecaab-9b43-474e-8dbf-b0adeb06015a"). InnerVolumeSpecName "kube-api-access-wxzkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.705165 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4bdecaab-9b43-474e-8dbf-b0adeb06015a" (UID: "4bdecaab-9b43-474e-8dbf-b0adeb06015a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.747510 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxzkz\" (UniqueName: \"kubernetes.io/projected/4bdecaab-9b43-474e-8dbf-b0adeb06015a-kube-api-access-wxzkz\") on node \"crc\" DevicePath \"\"" Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.747539 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:58:05 crc kubenswrapper[4978]: I0225 08:58:05.747551 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4bdecaab-9b43-474e-8dbf-b0adeb06015a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.044728 4978 generic.go:334] "Generic (PLEG): container finished" podID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerID="7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81" exitCode=0 Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.044768 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8msh6" event={"ID":"4bdecaab-9b43-474e-8dbf-b0adeb06015a","Type":"ContainerDied","Data":"7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81"} Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.044780 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8msh6" Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.044800 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8msh6" event={"ID":"4bdecaab-9b43-474e-8dbf-b0adeb06015a","Type":"ContainerDied","Data":"5636d57ddbc46286f33801341c61d2f3d6e83af24989c17e6c1b6117324be8f8"} Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.044822 4978 scope.go:117] "RemoveContainer" containerID="7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81" Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.083252 4978 scope.go:117] "RemoveContainer" containerID="487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082" Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.109316 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8msh6"] Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.125420 4978 scope.go:117] "RemoveContainer" containerID="0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213" Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.126660 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8msh6"] Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.180055 4978 scope.go:117] "RemoveContainer" containerID="7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81" Feb 25 08:58:06 crc kubenswrapper[4978]: E0225 08:58:06.180871 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81\": container with ID starting with 7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81 not found: ID does not exist" containerID="7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81" Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.180954 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81"} err="failed to get container status \"7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81\": rpc error: code = NotFound desc = could not find container \"7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81\": container with ID starting with 7936a427fc9f22eb229e215e66bbbebbac6c79cc99edc78e538f7088b8a7dc81 not found: ID does not exist" Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.180990 4978 scope.go:117] "RemoveContainer" containerID="487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082" Feb 25 08:58:06 crc kubenswrapper[4978]: E0225 08:58:06.181471 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082\": container with ID starting with 487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082 not found: ID does not exist" containerID="487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082" Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.181540 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082"} err="failed to get container status \"487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082\": rpc error: code = NotFound desc = could not find container \"487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082\": container with ID starting with 487c1641d2a1cffde962a600c28e34b5c976e62b6686ba4efacfd21dc2926082 not found: ID does not exist" Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.181572 4978 scope.go:117] "RemoveContainer" containerID="0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213" Feb 25 08:58:06 crc kubenswrapper[4978]: E0225 08:58:06.182329 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213\": container with ID starting with 0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213 not found: ID does not exist" containerID="0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213" Feb 25 08:58:06 crc kubenswrapper[4978]: I0225 08:58:06.182382 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213"} err="failed to get container status \"0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213\": rpc error: code = NotFound desc = could not find container \"0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213\": container with ID starting with 0b206a0676fab04a50b356307891a6f9e80404f93a9c44236cd908f9f520d213 not found: ID does not exist" Feb 25 08:58:07 crc kubenswrapper[4978]: I0225 08:58:07.346857 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" path="/var/lib/kubelet/pods/4bdecaab-9b43-474e-8dbf-b0adeb06015a/volumes" Feb 25 08:58:07 crc kubenswrapper[4978]: I0225 08:58:07.348238 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="708a02e1-f441-447f-9f0d-91cf374e76d1" path="/var/lib/kubelet/pods/708a02e1-f441-447f-9f0d-91cf374e76d1/volumes" Feb 25 08:58:10 crc kubenswrapper[4978]: I0225 08:58:10.047769 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wg2x6"] Feb 25 08:58:10 crc kubenswrapper[4978]: I0225 08:58:10.063618 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wg2x6"] Feb 25 08:58:11 crc kubenswrapper[4978]: I0225 08:58:11.341832 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c9e435e-8baf-4b51-8bf3-e43fc7dccf37" path="/var/lib/kubelet/pods/2c9e435e-8baf-4b51-8bf3-e43fc7dccf37/volumes" Feb 25 08:58:14 crc kubenswrapper[4978]: I0225 08:58:14.328877 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:58:14 crc kubenswrapper[4978]: E0225 08:58:14.329658 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:58:23 crc kubenswrapper[4978]: I0225 08:58:23.055969 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-mfd2w"] Feb 25 08:58:23 crc kubenswrapper[4978]: I0225 08:58:23.076668 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-mfd2w"] Feb 25 08:58:23 crc kubenswrapper[4978]: I0225 08:58:23.348726 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cf80876-ea3d-4885-90fa-cd977f45aee4" path="/var/lib/kubelet/pods/5cf80876-ea3d-4885-90fa-cd977f45aee4/volumes" Feb 25 08:58:24 crc kubenswrapper[4978]: I0225 08:58:24.045055 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-6lvjg"] Feb 25 08:58:24 crc kubenswrapper[4978]: I0225 08:58:24.063369 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-6lvjg"] Feb 25 08:58:25 crc kubenswrapper[4978]: I0225 08:58:25.327955 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:58:25 crc kubenswrapper[4978]: E0225 08:58:25.328770 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:58:25 crc kubenswrapper[4978]: I0225 08:58:25.346545 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55021e7b-713c-4697-a80b-49e0a23cff45" path="/var/lib/kubelet/pods/55021e7b-713c-4697-a80b-49e0a23cff45/volumes" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.222064 4978 scope.go:117] "RemoveContainer" containerID="3924ccbeb1105d7edb99f7895483c66bc1d515f33759ed49e1e9895226b097e8" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.250297 4978 scope.go:117] "RemoveContainer" containerID="c28cf767dd6d6ef31292b5a97fd23d78ef85205f6de7a28d720aec88b635f002" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.288879 4978 scope.go:117] "RemoveContainer" containerID="8a86ab3bcf2f60da67cdebd1d4fadad5f4755e71ea59fc896f3955ecfe5b556c" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.346430 4978 scope.go:117] "RemoveContainer" containerID="c0d820bd7d896cdafac234504b4db3ba443257d9c44c75b6c1a402bc60277a34" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.402639 4978 scope.go:117] "RemoveContainer" containerID="751f9625f412705ac75b12905ac0ccb115dc26186b4e8a15eae405a4eefba4e8" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.432279 4978 scope.go:117] "RemoveContainer" containerID="fbb0605b18bbb774138802fe465d1ceb037c141f087e0f5a2d3699959eb4dfb4" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.482558 4978 scope.go:117] "RemoveContainer" containerID="702d53de66165222bd7c2d97cd679ada7c060aed8645c986e09b3f658c29a92e" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.511810 4978 scope.go:117] "RemoveContainer" containerID="6e2cc42ac90096e89d9fa3b15d415de4a4496a34981b7bd378b8763c917cc9d1" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.528277 4978 scope.go:117] "RemoveContainer" containerID="6d859e57b682f1cb9b5daffd35493c5a5862453b86aad2e0c87e4c11ffafaa1b" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.554457 4978 scope.go:117] "RemoveContainer" containerID="a9a430476c450bdbcda78400c12c520b9944453552bd31acb86f41886183e9a2" Feb 25 08:58:33 crc kubenswrapper[4978]: I0225 08:58:33.597300 4978 scope.go:117] "RemoveContainer" containerID="f1c0033a1fabc69444f5f5fe3bfecb54fd07596c276d57302c7597782c6b1ed1" Feb 25 08:58:37 crc kubenswrapper[4978]: I0225 08:58:37.342310 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:58:37 crc kubenswrapper[4978]: E0225 08:58:37.343821 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.621082 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5p592"] Feb 25 08:58:46 crc kubenswrapper[4978]: E0225 08:58:46.622064 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerName="registry-server" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.622079 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerName="registry-server" Feb 25 08:58:46 crc kubenswrapper[4978]: E0225 08:58:46.622091 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a921f4a-dc79-4d25-befd-cd15da7fa69e" containerName="oc" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.622099 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a921f4a-dc79-4d25-befd-cd15da7fa69e" containerName="oc" Feb 25 08:58:46 crc kubenswrapper[4978]: E0225 08:58:46.622120 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerName="extract-utilities" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.622129 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerName="extract-utilities" Feb 25 08:58:46 crc kubenswrapper[4978]: E0225 08:58:46.622157 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerName="extract-content" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.622164 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerName="extract-content" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.622414 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bdecaab-9b43-474e-8dbf-b0adeb06015a" containerName="registry-server" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.622435 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a921f4a-dc79-4d25-befd-cd15da7fa69e" containerName="oc" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.626046 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.668637 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5p592"] Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.699309 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnmmj\" (UniqueName: \"kubernetes.io/projected/559fb5ba-d9bf-4e77-8d81-41fd0d809088-kube-api-access-hnmmj\") pod \"redhat-marketplace-5p592\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.699387 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-utilities\") pod \"redhat-marketplace-5p592\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.699410 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-catalog-content\") pod \"redhat-marketplace-5p592\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.801720 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnmmj\" (UniqueName: \"kubernetes.io/projected/559fb5ba-d9bf-4e77-8d81-41fd0d809088-kube-api-access-hnmmj\") pod \"redhat-marketplace-5p592\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.801783 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-utilities\") pod \"redhat-marketplace-5p592\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.801808 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-catalog-content\") pod \"redhat-marketplace-5p592\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.802492 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-catalog-content\") pod \"redhat-marketplace-5p592\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.802566 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-utilities\") pod \"redhat-marketplace-5p592\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.819879 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnmmj\" (UniqueName: \"kubernetes.io/projected/559fb5ba-d9bf-4e77-8d81-41fd0d809088-kube-api-access-hnmmj\") pod \"redhat-marketplace-5p592\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:46 crc kubenswrapper[4978]: I0225 08:58:46.981647 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:47 crc kubenswrapper[4978]: I0225 08:58:47.421535 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5p592"] Feb 25 08:58:47 crc kubenswrapper[4978]: I0225 08:58:47.572062 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p592" event={"ID":"559fb5ba-d9bf-4e77-8d81-41fd0d809088","Type":"ContainerStarted","Data":"a55e0f4513dd1f0c32ec621b1271d5d4db71d92b71590dffa32d39fff8aede7a"} Feb 25 08:58:48 crc kubenswrapper[4978]: I0225 08:58:48.582599 4978 generic.go:334] "Generic (PLEG): container finished" podID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerID="b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2" exitCode=0 Feb 25 08:58:48 crc kubenswrapper[4978]: I0225 08:58:48.582658 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p592" event={"ID":"559fb5ba-d9bf-4e77-8d81-41fd0d809088","Type":"ContainerDied","Data":"b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2"} Feb 25 08:58:49 crc kubenswrapper[4978]: I0225 08:58:49.328558 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:58:49 crc kubenswrapper[4978]: E0225 08:58:49.328941 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:58:49 crc kubenswrapper[4978]: I0225 08:58:49.595410 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p592" event={"ID":"559fb5ba-d9bf-4e77-8d81-41fd0d809088","Type":"ContainerStarted","Data":"6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40"} Feb 25 08:58:50 crc kubenswrapper[4978]: I0225 08:58:50.608719 4978 generic.go:334] "Generic (PLEG): container finished" podID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerID="6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40" exitCode=0 Feb 25 08:58:50 crc kubenswrapper[4978]: I0225 08:58:50.608786 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p592" event={"ID":"559fb5ba-d9bf-4e77-8d81-41fd0d809088","Type":"ContainerDied","Data":"6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40"} Feb 25 08:58:51 crc kubenswrapper[4978]: I0225 08:58:51.620913 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p592" event={"ID":"559fb5ba-d9bf-4e77-8d81-41fd0d809088","Type":"ContainerStarted","Data":"1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117"} Feb 25 08:58:51 crc kubenswrapper[4978]: I0225 08:58:51.667117 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5p592" podStartSLOduration=3.260776002 podStartE2EDuration="5.667088161s" podCreationTimestamp="2026-02-25 08:58:46 +0000 UTC" firstStartedPulling="2026-02-25 08:58:48.585323235 +0000 UTC m=+8022.024579724" lastFinishedPulling="2026-02-25 08:58:50.991635414 +0000 UTC m=+8024.430891883" observedRunningTime="2026-02-25 08:58:51.656002727 +0000 UTC m=+8025.095259186" watchObservedRunningTime="2026-02-25 08:58:51.667088161 +0000 UTC m=+8025.106344670" Feb 25 08:58:56 crc kubenswrapper[4978]: I0225 08:58:56.982265 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:56 crc kubenswrapper[4978]: I0225 08:58:56.982809 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:57 crc kubenswrapper[4978]: I0225 08:58:57.061237 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:57 crc kubenswrapper[4978]: I0225 08:58:57.768143 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:58:57 crc kubenswrapper[4978]: I0225 08:58:57.839426 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5p592"] Feb 25 08:58:59 crc kubenswrapper[4978]: I0225 08:58:59.762845 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5p592" podUID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerName="registry-server" containerID="cri-o://1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117" gracePeriod=2 Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.409645 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.488239 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-catalog-content\") pod \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.488676 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnmmj\" (UniqueName: \"kubernetes.io/projected/559fb5ba-d9bf-4e77-8d81-41fd0d809088-kube-api-access-hnmmj\") pod \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.489026 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-utilities\") pod \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\" (UID: \"559fb5ba-d9bf-4e77-8d81-41fd0d809088\") " Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.490404 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-utilities" (OuterVolumeSpecName: "utilities") pod "559fb5ba-d9bf-4e77-8d81-41fd0d809088" (UID: "559fb5ba-d9bf-4e77-8d81-41fd0d809088"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.502689 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/559fb5ba-d9bf-4e77-8d81-41fd0d809088-kube-api-access-hnmmj" (OuterVolumeSpecName: "kube-api-access-hnmmj") pod "559fb5ba-d9bf-4e77-8d81-41fd0d809088" (UID: "559fb5ba-d9bf-4e77-8d81-41fd0d809088"). InnerVolumeSpecName "kube-api-access-hnmmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.527731 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "559fb5ba-d9bf-4e77-8d81-41fd0d809088" (UID: "559fb5ba-d9bf-4e77-8d81-41fd0d809088"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.592857 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.592900 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnmmj\" (UniqueName: \"kubernetes.io/projected/559fb5ba-d9bf-4e77-8d81-41fd0d809088-kube-api-access-hnmmj\") on node \"crc\" DevicePath \"\"" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.592920 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/559fb5ba-d9bf-4e77-8d81-41fd0d809088-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.784287 4978 generic.go:334] "Generic (PLEG): container finished" podID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerID="1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117" exitCode=0 Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.784420 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5p592" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.784584 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p592" event={"ID":"559fb5ba-d9bf-4e77-8d81-41fd0d809088","Type":"ContainerDied","Data":"1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117"} Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.788506 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5p592" event={"ID":"559fb5ba-d9bf-4e77-8d81-41fd0d809088","Type":"ContainerDied","Data":"a55e0f4513dd1f0c32ec621b1271d5d4db71d92b71590dffa32d39fff8aede7a"} Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.788573 4978 scope.go:117] "RemoveContainer" containerID="1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.828627 4978 scope.go:117] "RemoveContainer" containerID="6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.854028 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5p592"] Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.883938 4978 scope.go:117] "RemoveContainer" containerID="b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.887412 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5p592"] Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.945858 4978 scope.go:117] "RemoveContainer" containerID="1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117" Feb 25 08:59:00 crc kubenswrapper[4978]: E0225 08:59:00.946258 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117\": container with ID starting with 1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117 not found: ID does not exist" containerID="1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.946296 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117"} err="failed to get container status \"1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117\": rpc error: code = NotFound desc = could not find container \"1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117\": container with ID starting with 1b52a7b446d06cf587db63c13a2d9104e64e615113ffd01235b33f3833dfd117 not found: ID does not exist" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.946320 4978 scope.go:117] "RemoveContainer" containerID="6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40" Feb 25 08:59:00 crc kubenswrapper[4978]: E0225 08:59:00.946630 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40\": container with ID starting with 6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40 not found: ID does not exist" containerID="6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.946673 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40"} err="failed to get container status \"6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40\": rpc error: code = NotFound desc = could not find container \"6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40\": container with ID starting with 6b5979e14b056f02d621335dc65c5a5ea7898733629872624bc4352bfe69cf40 not found: ID does not exist" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.946705 4978 scope.go:117] "RemoveContainer" containerID="b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2" Feb 25 08:59:00 crc kubenswrapper[4978]: E0225 08:59:00.946979 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2\": container with ID starting with b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2 not found: ID does not exist" containerID="b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2" Feb 25 08:59:00 crc kubenswrapper[4978]: I0225 08:59:00.947011 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2"} err="failed to get container status \"b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2\": rpc error: code = NotFound desc = could not find container \"b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2\": container with ID starting with b1c7819100394a6c70911d6542faa61dd0c736eb36c8f7a8ca513ea3369583e2 not found: ID does not exist" Feb 25 08:59:01 crc kubenswrapper[4978]: I0225 08:59:01.328359 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:59:01 crc kubenswrapper[4978]: E0225 08:59:01.328885 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:59:01 crc kubenswrapper[4978]: I0225 08:59:01.342570 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" path="/var/lib/kubelet/pods/559fb5ba-d9bf-4e77-8d81-41fd0d809088/volumes" Feb 25 08:59:12 crc kubenswrapper[4978]: I0225 08:59:12.064260 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-d794z"] Feb 25 08:59:12 crc kubenswrapper[4978]: I0225 08:59:12.077387 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-d794z"] Feb 25 08:59:13 crc kubenswrapper[4978]: I0225 08:59:13.342606 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2" path="/var/lib/kubelet/pods/339cbafa-b3c1-4af8-9ac4-ee698a0d3ff2/volumes" Feb 25 08:59:15 crc kubenswrapper[4978]: I0225 08:59:15.327695 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:59:15 crc kubenswrapper[4978]: E0225 08:59:15.328493 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:59:29 crc kubenswrapper[4978]: I0225 08:59:29.328414 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:59:29 crc kubenswrapper[4978]: E0225 08:59:29.329527 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:59:33 crc kubenswrapper[4978]: I0225 08:59:33.920998 4978 scope.go:117] "RemoveContainer" containerID="1185ba4cd40e1397a90efca975c8396c8c484410a8e717821da832987a228fb6" Feb 25 08:59:43 crc kubenswrapper[4978]: I0225 08:59:43.329313 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:59:43 crc kubenswrapper[4978]: E0225 08:59:43.330176 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 08:59:57 crc kubenswrapper[4978]: I0225 08:59:57.340841 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 08:59:58 crc kubenswrapper[4978]: I0225 08:59:58.471842 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"71fe33810cada46fd9054688d7deaba29b642bf93d0651d2c0a3a47b2f670de1"} Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.170578 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2"] Feb 25 09:00:00 crc kubenswrapper[4978]: E0225 09:00:00.171494 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerName="extract-content" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.171508 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerName="extract-content" Feb 25 09:00:00 crc kubenswrapper[4978]: E0225 09:00:00.171523 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerName="extract-utilities" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.171530 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerName="extract-utilities" Feb 25 09:00:00 crc kubenswrapper[4978]: E0225 09:00:00.171561 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerName="registry-server" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.171568 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerName="registry-server" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.171742 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="559fb5ba-d9bf-4e77-8d81-41fd0d809088" containerName="registry-server" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.172452 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.174875 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.179991 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.182196 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533500-gg4p9"] Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.183952 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533500-gg4p9" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.187883 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.187980 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.188075 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.195107 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2"] Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.212130 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533500-gg4p9"] Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.356268 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d7114d5-5975-480f-acbc-66ffd0aff346-config-volume\") pod \"collect-profiles-29533500-4rsv2\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.356326 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpdl8\" (UniqueName: \"kubernetes.io/projected/259ed9e0-f44d-49d9-b800-df10bf4df3da-kube-api-access-wpdl8\") pod \"auto-csr-approver-29533500-gg4p9\" (UID: \"259ed9e0-f44d-49d9-b800-df10bf4df3da\") " pod="openshift-infra/auto-csr-approver-29533500-gg4p9" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.356974 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d7114d5-5975-480f-acbc-66ffd0aff346-secret-volume\") pod \"collect-profiles-29533500-4rsv2\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.357344 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4554\" (UniqueName: \"kubernetes.io/projected/0d7114d5-5975-480f-acbc-66ffd0aff346-kube-api-access-x4554\") pod \"collect-profiles-29533500-4rsv2\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.458949 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d7114d5-5975-480f-acbc-66ffd0aff346-config-volume\") pod \"collect-profiles-29533500-4rsv2\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.459025 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpdl8\" (UniqueName: \"kubernetes.io/projected/259ed9e0-f44d-49d9-b800-df10bf4df3da-kube-api-access-wpdl8\") pod \"auto-csr-approver-29533500-gg4p9\" (UID: \"259ed9e0-f44d-49d9-b800-df10bf4df3da\") " pod="openshift-infra/auto-csr-approver-29533500-gg4p9" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.459126 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d7114d5-5975-480f-acbc-66ffd0aff346-secret-volume\") pod \"collect-profiles-29533500-4rsv2\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.459171 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4554\" (UniqueName: \"kubernetes.io/projected/0d7114d5-5975-480f-acbc-66ffd0aff346-kube-api-access-x4554\") pod \"collect-profiles-29533500-4rsv2\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.460216 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d7114d5-5975-480f-acbc-66ffd0aff346-config-volume\") pod \"collect-profiles-29533500-4rsv2\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.468246 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d7114d5-5975-480f-acbc-66ffd0aff346-secret-volume\") pod \"collect-profiles-29533500-4rsv2\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.476099 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpdl8\" (UniqueName: \"kubernetes.io/projected/259ed9e0-f44d-49d9-b800-df10bf4df3da-kube-api-access-wpdl8\") pod \"auto-csr-approver-29533500-gg4p9\" (UID: \"259ed9e0-f44d-49d9-b800-df10bf4df3da\") " pod="openshift-infra/auto-csr-approver-29533500-gg4p9" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.480241 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4554\" (UniqueName: \"kubernetes.io/projected/0d7114d5-5975-480f-acbc-66ffd0aff346-kube-api-access-x4554\") pod \"collect-profiles-29533500-4rsv2\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.538693 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.546751 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533500-gg4p9" Feb 25 09:00:00 crc kubenswrapper[4978]: I0225 09:00:00.927106 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533500-gg4p9"] Feb 25 09:00:00 crc kubenswrapper[4978]: W0225 09:00:00.928590 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod259ed9e0_f44d_49d9_b800_df10bf4df3da.slice/crio-59dea8c99129621e8586048b644afcc9f339a5a755e33fc2ad7cd4dad25a5962 WatchSource:0}: Error finding container 59dea8c99129621e8586048b644afcc9f339a5a755e33fc2ad7cd4dad25a5962: Status 404 returned error can't find the container with id 59dea8c99129621e8586048b644afcc9f339a5a755e33fc2ad7cd4dad25a5962 Feb 25 09:00:01 crc kubenswrapper[4978]: I0225 09:00:01.182519 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2"] Feb 25 09:00:01 crc kubenswrapper[4978]: I0225 09:00:01.513197 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533500-gg4p9" event={"ID":"259ed9e0-f44d-49d9-b800-df10bf4df3da","Type":"ContainerStarted","Data":"59dea8c99129621e8586048b644afcc9f339a5a755e33fc2ad7cd4dad25a5962"} Feb 25 09:00:01 crc kubenswrapper[4978]: I0225 09:00:01.515106 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" event={"ID":"0d7114d5-5975-480f-acbc-66ffd0aff346","Type":"ContainerStarted","Data":"643169a3c71c7a40fb55504469b404bbde60718bc95d306ab88068f5a5127cd4"} Feb 25 09:00:01 crc kubenswrapper[4978]: I0225 09:00:01.515147 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" event={"ID":"0d7114d5-5975-480f-acbc-66ffd0aff346","Type":"ContainerStarted","Data":"f41fef907671f61c0addc71c3c8d32f637c5ed69cfba51ad97cdb6e390ef81ee"} Feb 25 09:00:01 crc kubenswrapper[4978]: I0225 09:00:01.539862 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" podStartSLOduration=1.539839571 podStartE2EDuration="1.539839571s" podCreationTimestamp="2026-02-25 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:00:01.533893866 +0000 UTC m=+8094.973150355" watchObservedRunningTime="2026-02-25 09:00:01.539839571 +0000 UTC m=+8094.979096030" Feb 25 09:00:02 crc kubenswrapper[4978]: I0225 09:00:02.533576 4978 generic.go:334] "Generic (PLEG): container finished" podID="0d7114d5-5975-480f-acbc-66ffd0aff346" containerID="643169a3c71c7a40fb55504469b404bbde60718bc95d306ab88068f5a5127cd4" exitCode=0 Feb 25 09:00:02 crc kubenswrapper[4978]: I0225 09:00:02.533663 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" event={"ID":"0d7114d5-5975-480f-acbc-66ffd0aff346","Type":"ContainerDied","Data":"643169a3c71c7a40fb55504469b404bbde60718bc95d306ab88068f5a5127cd4"} Feb 25 09:00:03 crc kubenswrapper[4978]: I0225 09:00:03.964624 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.151102 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d7114d5-5975-480f-acbc-66ffd0aff346-config-volume\") pod \"0d7114d5-5975-480f-acbc-66ffd0aff346\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.151285 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4554\" (UniqueName: \"kubernetes.io/projected/0d7114d5-5975-480f-acbc-66ffd0aff346-kube-api-access-x4554\") pod \"0d7114d5-5975-480f-acbc-66ffd0aff346\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.151363 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d7114d5-5975-480f-acbc-66ffd0aff346-secret-volume\") pod \"0d7114d5-5975-480f-acbc-66ffd0aff346\" (UID: \"0d7114d5-5975-480f-acbc-66ffd0aff346\") " Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.151885 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d7114d5-5975-480f-acbc-66ffd0aff346-config-volume" (OuterVolumeSpecName: "config-volume") pod "0d7114d5-5975-480f-acbc-66ffd0aff346" (UID: "0d7114d5-5975-480f-acbc-66ffd0aff346"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.156820 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d7114d5-5975-480f-acbc-66ffd0aff346-kube-api-access-x4554" (OuterVolumeSpecName: "kube-api-access-x4554") pod "0d7114d5-5975-480f-acbc-66ffd0aff346" (UID: "0d7114d5-5975-480f-acbc-66ffd0aff346"). InnerVolumeSpecName "kube-api-access-x4554". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.157240 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d7114d5-5975-480f-acbc-66ffd0aff346-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0d7114d5-5975-480f-acbc-66ffd0aff346" (UID: "0d7114d5-5975-480f-acbc-66ffd0aff346"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.237278 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7"] Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.243067 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533455-v2rl7"] Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.253851 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d7114d5-5975-480f-acbc-66ffd0aff346-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.253880 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4554\" (UniqueName: \"kubernetes.io/projected/0d7114d5-5975-480f-acbc-66ffd0aff346-kube-api-access-x4554\") on node \"crc\" DevicePath \"\"" Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.253890 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d7114d5-5975-480f-acbc-66ffd0aff346-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.562877 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" event={"ID":"0d7114d5-5975-480f-acbc-66ffd0aff346","Type":"ContainerDied","Data":"f41fef907671f61c0addc71c3c8d32f637c5ed69cfba51ad97cdb6e390ef81ee"} Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.563236 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f41fef907671f61c0addc71c3c8d32f637c5ed69cfba51ad97cdb6e390ef81ee" Feb 25 09:00:04 crc kubenswrapper[4978]: I0225 09:00:04.563029 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2" Feb 25 09:00:05 crc kubenswrapper[4978]: I0225 09:00:05.348159 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="302baf2f-e0af-4f5e-88dd-36682e379d58" path="/var/lib/kubelet/pods/302baf2f-e0af-4f5e-88dd-36682e379d58/volumes" Feb 25 09:00:09 crc kubenswrapper[4978]: I0225 09:00:09.623788 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533500-gg4p9" event={"ID":"259ed9e0-f44d-49d9-b800-df10bf4df3da","Type":"ContainerStarted","Data":"4b527beba7c63d5d7de5c8ff43fe3202c98f6f0377a97e19e3cca162a61c297f"} Feb 25 09:00:09 crc kubenswrapper[4978]: I0225 09:00:09.654969 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533500-gg4p9" podStartSLOduration=1.743024605 podStartE2EDuration="9.654939012s" podCreationTimestamp="2026-02-25 09:00:00 +0000 UTC" firstStartedPulling="2026-02-25 09:00:00.930783581 +0000 UTC m=+8094.370040040" lastFinishedPulling="2026-02-25 09:00:08.842697998 +0000 UTC m=+8102.281954447" observedRunningTime="2026-02-25 09:00:09.643669303 +0000 UTC m=+8103.082925822" watchObservedRunningTime="2026-02-25 09:00:09.654939012 +0000 UTC m=+8103.094195511" Feb 25 09:00:10 crc kubenswrapper[4978]: I0225 09:00:10.647126 4978 generic.go:334] "Generic (PLEG): container finished" podID="259ed9e0-f44d-49d9-b800-df10bf4df3da" containerID="4b527beba7c63d5d7de5c8ff43fe3202c98f6f0377a97e19e3cca162a61c297f" exitCode=0 Feb 25 09:00:10 crc kubenswrapper[4978]: I0225 09:00:10.647213 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533500-gg4p9" event={"ID":"259ed9e0-f44d-49d9-b800-df10bf4df3da","Type":"ContainerDied","Data":"4b527beba7c63d5d7de5c8ff43fe3202c98f6f0377a97e19e3cca162a61c297f"} Feb 25 09:00:12 crc kubenswrapper[4978]: I0225 09:00:12.126942 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533500-gg4p9" Feb 25 09:00:12 crc kubenswrapper[4978]: I0225 09:00:12.242391 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpdl8\" (UniqueName: \"kubernetes.io/projected/259ed9e0-f44d-49d9-b800-df10bf4df3da-kube-api-access-wpdl8\") pod \"259ed9e0-f44d-49d9-b800-df10bf4df3da\" (UID: \"259ed9e0-f44d-49d9-b800-df10bf4df3da\") " Feb 25 09:00:12 crc kubenswrapper[4978]: I0225 09:00:12.251063 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/259ed9e0-f44d-49d9-b800-df10bf4df3da-kube-api-access-wpdl8" (OuterVolumeSpecName: "kube-api-access-wpdl8") pod "259ed9e0-f44d-49d9-b800-df10bf4df3da" (UID: "259ed9e0-f44d-49d9-b800-df10bf4df3da"). InnerVolumeSpecName "kube-api-access-wpdl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:00:12 crc kubenswrapper[4978]: I0225 09:00:12.345270 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpdl8\" (UniqueName: \"kubernetes.io/projected/259ed9e0-f44d-49d9-b800-df10bf4df3da-kube-api-access-wpdl8\") on node \"crc\" DevicePath \"\"" Feb 25 09:00:12 crc kubenswrapper[4978]: I0225 09:00:12.669992 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533500-gg4p9" event={"ID":"259ed9e0-f44d-49d9-b800-df10bf4df3da","Type":"ContainerDied","Data":"59dea8c99129621e8586048b644afcc9f339a5a755e33fc2ad7cd4dad25a5962"} Feb 25 09:00:12 crc kubenswrapper[4978]: I0225 09:00:12.670048 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59dea8c99129621e8586048b644afcc9f339a5a755e33fc2ad7cd4dad25a5962" Feb 25 09:00:12 crc kubenswrapper[4978]: I0225 09:00:12.670068 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533500-gg4p9" Feb 25 09:00:12 crc kubenswrapper[4978]: I0225 09:00:12.720999 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533494-fqpmk"] Feb 25 09:00:12 crc kubenswrapper[4978]: I0225 09:00:12.729541 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533494-fqpmk"] Feb 25 09:00:13 crc kubenswrapper[4978]: I0225 09:00:13.346442 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95208faf-d12c-48b1-aaab-98bb4c11a953" path="/var/lib/kubelet/pods/95208faf-d12c-48b1-aaab-98bb4c11a953/volumes" Feb 25 09:00:34 crc kubenswrapper[4978]: I0225 09:00:34.037551 4978 scope.go:117] "RemoveContainer" containerID="eb3b71bb78df7a040c81acbd60e451a59b5384296b041111b8b5140598dfe1a6" Feb 25 09:00:34 crc kubenswrapper[4978]: I0225 09:00:34.077835 4978 scope.go:117] "RemoveContainer" containerID="14d1938a85c40947dd515537c458414c5216a230ed0fb5efbf541226d287f404" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.186514 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29533501-f2btx"] Feb 25 09:01:00 crc kubenswrapper[4978]: E0225 09:01:00.187939 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="259ed9e0-f44d-49d9-b800-df10bf4df3da" containerName="oc" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.187975 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="259ed9e0-f44d-49d9-b800-df10bf4df3da" containerName="oc" Feb 25 09:01:00 crc kubenswrapper[4978]: E0225 09:01:00.188038 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d7114d5-5975-480f-acbc-66ffd0aff346" containerName="collect-profiles" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.188058 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d7114d5-5975-480f-acbc-66ffd0aff346" containerName="collect-profiles" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.188651 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="259ed9e0-f44d-49d9-b800-df10bf4df3da" containerName="oc" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.188746 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d7114d5-5975-480f-acbc-66ffd0aff346" containerName="collect-profiles" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.190596 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.217077 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29533501-f2btx"] Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.257400 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-combined-ca-bundle\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.257436 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-config-data\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.257480 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-fernet-keys\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.257513 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxc2z\" (UniqueName: \"kubernetes.io/projected/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-kube-api-access-zxc2z\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.358198 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-combined-ca-bundle\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.358258 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-config-data\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.358295 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-fernet-keys\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.358391 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxc2z\" (UniqueName: \"kubernetes.io/projected/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-kube-api-access-zxc2z\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.364654 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-fernet-keys\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.365319 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-config-data\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.367736 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-combined-ca-bundle\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.397622 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxc2z\" (UniqueName: \"kubernetes.io/projected/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-kube-api-access-zxc2z\") pod \"keystone-cron-29533501-f2btx\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:00 crc kubenswrapper[4978]: I0225 09:01:00.515496 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:01 crc kubenswrapper[4978]: I0225 09:01:01.036710 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29533501-f2btx"] Feb 25 09:01:01 crc kubenswrapper[4978]: W0225 09:01:01.039074 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d5bf4f2_3ff6_405c_85a5_4423fbd1d88e.slice/crio-3d51f6cb3e4c683cd5c385b5e470b3947fed842379aee42e66b11523c30b09ad WatchSource:0}: Error finding container 3d51f6cb3e4c683cd5c385b5e470b3947fed842379aee42e66b11523c30b09ad: Status 404 returned error can't find the container with id 3d51f6cb3e4c683cd5c385b5e470b3947fed842379aee42e66b11523c30b09ad Feb 25 09:01:01 crc kubenswrapper[4978]: I0225 09:01:01.264650 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533501-f2btx" event={"ID":"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e","Type":"ContainerStarted","Data":"a08fa9a6ff15cddce66b66e1744db8982d51cb800e3db6bba248c17d4a7b7a0d"} Feb 25 09:01:01 crc kubenswrapper[4978]: I0225 09:01:01.264718 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533501-f2btx" event={"ID":"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e","Type":"ContainerStarted","Data":"3d51f6cb3e4c683cd5c385b5e470b3947fed842379aee42e66b11523c30b09ad"} Feb 25 09:01:02 crc kubenswrapper[4978]: I0225 09:01:02.296409 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29533501-f2btx" podStartSLOduration=2.296354359 podStartE2EDuration="2.296354359s" podCreationTimestamp="2026-02-25 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:01:02.290130267 +0000 UTC m=+8155.729386756" watchObservedRunningTime="2026-02-25 09:01:02.296354359 +0000 UTC m=+8155.735610828" Feb 25 09:01:05 crc kubenswrapper[4978]: I0225 09:01:05.310607 4978 generic.go:334] "Generic (PLEG): container finished" podID="5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e" containerID="a08fa9a6ff15cddce66b66e1744db8982d51cb800e3db6bba248c17d4a7b7a0d" exitCode=0 Feb 25 09:01:05 crc kubenswrapper[4978]: I0225 09:01:05.310705 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533501-f2btx" event={"ID":"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e","Type":"ContainerDied","Data":"a08fa9a6ff15cddce66b66e1744db8982d51cb800e3db6bba248c17d4a7b7a0d"} Feb 25 09:01:06 crc kubenswrapper[4978]: I0225 09:01:06.832016 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:06 crc kubenswrapper[4978]: I0225 09:01:06.924195 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-fernet-keys\") pod \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " Feb 25 09:01:06 crc kubenswrapper[4978]: I0225 09:01:06.924249 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-config-data\") pod \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " Feb 25 09:01:06 crc kubenswrapper[4978]: I0225 09:01:06.924524 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-combined-ca-bundle\") pod \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " Feb 25 09:01:06 crc kubenswrapper[4978]: I0225 09:01:06.924560 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxc2z\" (UniqueName: \"kubernetes.io/projected/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-kube-api-access-zxc2z\") pod \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\" (UID: \"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e\") " Feb 25 09:01:06 crc kubenswrapper[4978]: I0225 09:01:06.931718 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-kube-api-access-zxc2z" (OuterVolumeSpecName: "kube-api-access-zxc2z") pod "5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e" (UID: "5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e"). InnerVolumeSpecName "kube-api-access-zxc2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:01:06 crc kubenswrapper[4978]: I0225 09:01:06.938614 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e" (UID: "5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:01:06 crc kubenswrapper[4978]: I0225 09:01:06.956597 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e" (UID: "5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:01:06 crc kubenswrapper[4978]: I0225 09:01:06.983487 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-config-data" (OuterVolumeSpecName: "config-data") pod "5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e" (UID: "5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:01:07 crc kubenswrapper[4978]: I0225 09:01:07.027230 4978 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:07 crc kubenswrapper[4978]: I0225 09:01:07.027268 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:07 crc kubenswrapper[4978]: I0225 09:01:07.027280 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:07 crc kubenswrapper[4978]: I0225 09:01:07.027291 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxc2z\" (UniqueName: \"kubernetes.io/projected/5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e-kube-api-access-zxc2z\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:07 crc kubenswrapper[4978]: I0225 09:01:07.345060 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533501-f2btx" event={"ID":"5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e","Type":"ContainerDied","Data":"3d51f6cb3e4c683cd5c385b5e470b3947fed842379aee42e66b11523c30b09ad"} Feb 25 09:01:07 crc kubenswrapper[4978]: I0225 09:01:07.345138 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d51f6cb3e4c683cd5c385b5e470b3947fed842379aee42e66b11523c30b09ad" Feb 25 09:01:07 crc kubenswrapper[4978]: I0225 09:01:07.345289 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533501-f2btx" Feb 25 09:01:26 crc kubenswrapper[4978]: I0225 09:01:26.559474 4978 generic.go:334] "Generic (PLEG): container finished" podID="6abcdd51-69ae-42d5-9708-7088d9911458" containerID="85656ef656077c83751a5c3845c4f90e2f262eaaca42c0427a37198d34b9c5ff" exitCode=0 Feb 25 09:01:26 crc kubenswrapper[4978]: I0225 09:01:26.559743 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" event={"ID":"6abcdd51-69ae-42d5-9708-7088d9911458","Type":"ContainerDied","Data":"85656ef656077c83751a5c3845c4f90e2f262eaaca42c0427a37198d34b9c5ff"} Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.069792 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.241898 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-inventory\") pod \"6abcdd51-69ae-42d5-9708-7088d9911458\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.241982 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppgzr\" (UniqueName: \"kubernetes.io/projected/6abcdd51-69ae-42d5-9708-7088d9911458-kube-api-access-ppgzr\") pod \"6abcdd51-69ae-42d5-9708-7088d9911458\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.242044 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-tripleo-cleanup-combined-ca-bundle\") pod \"6abcdd51-69ae-42d5-9708-7088d9911458\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.242167 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-ssh-key-openstack-networker\") pod \"6abcdd51-69ae-42d5-9708-7088d9911458\" (UID: \"6abcdd51-69ae-42d5-9708-7088d9911458\") " Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.249633 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "6abcdd51-69ae-42d5-9708-7088d9911458" (UID: "6abcdd51-69ae-42d5-9708-7088d9911458"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.249908 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6abcdd51-69ae-42d5-9708-7088d9911458-kube-api-access-ppgzr" (OuterVolumeSpecName: "kube-api-access-ppgzr") pod "6abcdd51-69ae-42d5-9708-7088d9911458" (UID: "6abcdd51-69ae-42d5-9708-7088d9911458"). InnerVolumeSpecName "kube-api-access-ppgzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.273274 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "6abcdd51-69ae-42d5-9708-7088d9911458" (UID: "6abcdd51-69ae-42d5-9708-7088d9911458"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.276261 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-inventory" (OuterVolumeSpecName: "inventory") pod "6abcdd51-69ae-42d5-9708-7088d9911458" (UID: "6abcdd51-69ae-42d5-9708-7088d9911458"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.345108 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.345163 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppgzr\" (UniqueName: \"kubernetes.io/projected/6abcdd51-69ae-42d5-9708-7088d9911458-kube-api-access-ppgzr\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.345186 4978 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.345205 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/6abcdd51-69ae-42d5-9708-7088d9911458-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.581699 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" event={"ID":"6abcdd51-69ae-42d5-9708-7088d9911458","Type":"ContainerDied","Data":"1aecf0a41564322e277ec855264a26acb40ad03cfb9d1c7b40e9c5f0f7c5102c"} Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.581742 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1aecf0a41564322e277ec855264a26acb40ad03cfb9d1c7b40e9c5f0f7c5102c" Feb 25 09:01:28 crc kubenswrapper[4978]: I0225 09:01:28.581819 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8" Feb 25 09:01:32 crc kubenswrapper[4978]: I0225 09:01:32.968292 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s785w"] Feb 25 09:01:32 crc kubenswrapper[4978]: E0225 09:01:32.969707 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6abcdd51-69ae-42d5-9708-7088d9911458" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Feb 25 09:01:32 crc kubenswrapper[4978]: I0225 09:01:32.969737 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6abcdd51-69ae-42d5-9708-7088d9911458" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Feb 25 09:01:32 crc kubenswrapper[4978]: E0225 09:01:32.969764 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e" containerName="keystone-cron" Feb 25 09:01:32 crc kubenswrapper[4978]: I0225 09:01:32.969781 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e" containerName="keystone-cron" Feb 25 09:01:32 crc kubenswrapper[4978]: I0225 09:01:32.970127 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6abcdd51-69ae-42d5-9708-7088d9911458" containerName="tripleo-cleanup-tripleo-cleanup-openstack-networker" Feb 25 09:01:32 crc kubenswrapper[4978]: I0225 09:01:32.970170 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e" containerName="keystone-cron" Feb 25 09:01:32 crc kubenswrapper[4978]: I0225 09:01:32.973058 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:32 crc kubenswrapper[4978]: I0225 09:01:32.990957 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s785w"] Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.075058 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-catalog-content\") pod \"community-operators-s785w\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.075867 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lskq9\" (UniqueName: \"kubernetes.io/projected/e24e92a5-c015-4531-beaf-4829ae60cb1d-kube-api-access-lskq9\") pod \"community-operators-s785w\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.075982 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-utilities\") pod \"community-operators-s785w\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.177893 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-catalog-content\") pod \"community-operators-s785w\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.177978 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lskq9\" (UniqueName: \"kubernetes.io/projected/e24e92a5-c015-4531-beaf-4829ae60cb1d-kube-api-access-lskq9\") pod \"community-operators-s785w\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.178016 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-utilities\") pod \"community-operators-s785w\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.178677 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-utilities\") pod \"community-operators-s785w\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.178768 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-catalog-content\") pod \"community-operators-s785w\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.202337 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lskq9\" (UniqueName: \"kubernetes.io/projected/e24e92a5-c015-4531-beaf-4829ae60cb1d-kube-api-access-lskq9\") pod \"community-operators-s785w\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.306535 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:33 crc kubenswrapper[4978]: I0225 09:01:33.839926 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s785w"] Feb 25 09:01:34 crc kubenswrapper[4978]: I0225 09:01:34.659709 4978 generic.go:334] "Generic (PLEG): container finished" podID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerID="12b6b8ef135eb5a078d696152786ba835f6511018a32c8418df6284f34c7aff3" exitCode=0 Feb 25 09:01:34 crc kubenswrapper[4978]: I0225 09:01:34.659802 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s785w" event={"ID":"e24e92a5-c015-4531-beaf-4829ae60cb1d","Type":"ContainerDied","Data":"12b6b8ef135eb5a078d696152786ba835f6511018a32c8418df6284f34c7aff3"} Feb 25 09:01:34 crc kubenswrapper[4978]: I0225 09:01:34.660312 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s785w" event={"ID":"e24e92a5-c015-4531-beaf-4829ae60cb1d","Type":"ContainerStarted","Data":"6955cf63af50d763a7247bd001526310ca3a823f37f6692f7d8a2fcff7e652bc"} Feb 25 09:01:34 crc kubenswrapper[4978]: I0225 09:01:34.663348 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:01:36 crc kubenswrapper[4978]: I0225 09:01:36.686318 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s785w" event={"ID":"e24e92a5-c015-4531-beaf-4829ae60cb1d","Type":"ContainerStarted","Data":"0bf0492e49d183958582013bd2e1ac8b4e6705126edfc98fad0909b63aa4c491"} Feb 25 09:01:37 crc kubenswrapper[4978]: I0225 09:01:37.699260 4978 generic.go:334] "Generic (PLEG): container finished" podID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerID="0bf0492e49d183958582013bd2e1ac8b4e6705126edfc98fad0909b63aa4c491" exitCode=0 Feb 25 09:01:37 crc kubenswrapper[4978]: I0225 09:01:37.699318 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s785w" event={"ID":"e24e92a5-c015-4531-beaf-4829ae60cb1d","Type":"ContainerDied","Data":"0bf0492e49d183958582013bd2e1ac8b4e6705126edfc98fad0909b63aa4c491"} Feb 25 09:01:38 crc kubenswrapper[4978]: I0225 09:01:38.709479 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s785w" event={"ID":"e24e92a5-c015-4531-beaf-4829ae60cb1d","Type":"ContainerStarted","Data":"1c3227fd76b1a3f448d220a82b48541757b91a286314c549037e2d5a1348a93b"} Feb 25 09:01:38 crc kubenswrapper[4978]: I0225 09:01:38.746414 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s785w" podStartSLOduration=3.131187957 podStartE2EDuration="6.74639126s" podCreationTimestamp="2026-02-25 09:01:32 +0000 UTC" firstStartedPulling="2026-02-25 09:01:34.66309936 +0000 UTC m=+8188.102355819" lastFinishedPulling="2026-02-25 09:01:38.278302653 +0000 UTC m=+8191.717559122" observedRunningTime="2026-02-25 09:01:38.741970172 +0000 UTC m=+8192.181226631" watchObservedRunningTime="2026-02-25 09:01:38.74639126 +0000 UTC m=+8192.185647719" Feb 25 09:01:43 crc kubenswrapper[4978]: I0225 09:01:43.307722 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:43 crc kubenswrapper[4978]: I0225 09:01:43.308915 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:43 crc kubenswrapper[4978]: I0225 09:01:43.385276 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:43 crc kubenswrapper[4978]: I0225 09:01:43.878493 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.343147 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s785w"] Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.344099 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s785w" podUID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerName="registry-server" containerID="cri-o://1c3227fd76b1a3f448d220a82b48541757b91a286314c549037e2d5a1348a93b" gracePeriod=2 Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.832449 4978 generic.go:334] "Generic (PLEG): container finished" podID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerID="1c3227fd76b1a3f448d220a82b48541757b91a286314c549037e2d5a1348a93b" exitCode=0 Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.832525 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s785w" event={"ID":"e24e92a5-c015-4531-beaf-4829ae60cb1d","Type":"ContainerDied","Data":"1c3227fd76b1a3f448d220a82b48541757b91a286314c549037e2d5a1348a93b"} Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.832861 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s785w" event={"ID":"e24e92a5-c015-4531-beaf-4829ae60cb1d","Type":"ContainerDied","Data":"6955cf63af50d763a7247bd001526310ca3a823f37f6692f7d8a2fcff7e652bc"} Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.832886 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6955cf63af50d763a7247bd001526310ca3a823f37f6692f7d8a2fcff7e652bc" Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.853679 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.934809 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-utilities\") pod \"e24e92a5-c015-4531-beaf-4829ae60cb1d\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.934974 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lskq9\" (UniqueName: \"kubernetes.io/projected/e24e92a5-c015-4531-beaf-4829ae60cb1d-kube-api-access-lskq9\") pod \"e24e92a5-c015-4531-beaf-4829ae60cb1d\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.937958 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-catalog-content\") pod \"e24e92a5-c015-4531-beaf-4829ae60cb1d\" (UID: \"e24e92a5-c015-4531-beaf-4829ae60cb1d\") " Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.941460 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-utilities" (OuterVolumeSpecName: "utilities") pod "e24e92a5-c015-4531-beaf-4829ae60cb1d" (UID: "e24e92a5-c015-4531-beaf-4829ae60cb1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:01:47 crc kubenswrapper[4978]: I0225 09:01:47.944662 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e24e92a5-c015-4531-beaf-4829ae60cb1d-kube-api-access-lskq9" (OuterVolumeSpecName: "kube-api-access-lskq9") pod "e24e92a5-c015-4531-beaf-4829ae60cb1d" (UID: "e24e92a5-c015-4531-beaf-4829ae60cb1d"). InnerVolumeSpecName "kube-api-access-lskq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:01:48 crc kubenswrapper[4978]: I0225 09:01:48.028255 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e24e92a5-c015-4531-beaf-4829ae60cb1d" (UID: "e24e92a5-c015-4531-beaf-4829ae60cb1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:01:48 crc kubenswrapper[4978]: I0225 09:01:48.041023 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:48 crc kubenswrapper[4978]: I0225 09:01:48.041062 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e24e92a5-c015-4531-beaf-4829ae60cb1d-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:48 crc kubenswrapper[4978]: I0225 09:01:48.041075 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lskq9\" (UniqueName: \"kubernetes.io/projected/e24e92a5-c015-4531-beaf-4829ae60cb1d-kube-api-access-lskq9\") on node \"crc\" DevicePath \"\"" Feb 25 09:01:48 crc kubenswrapper[4978]: I0225 09:01:48.845630 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s785w" Feb 25 09:01:48 crc kubenswrapper[4978]: I0225 09:01:48.904686 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s785w"] Feb 25 09:01:48 crc kubenswrapper[4978]: I0225 09:01:48.913965 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s785w"] Feb 25 09:01:49 crc kubenswrapper[4978]: I0225 09:01:49.338863 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e24e92a5-c015-4531-beaf-4829ae60cb1d" path="/var/lib/kubelet/pods/e24e92a5-c015-4531-beaf-4829ae60cb1d/volumes" Feb 25 09:01:52 crc kubenswrapper[4978]: I0225 09:01:52.046790 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-96h4w"] Feb 25 09:01:52 crc kubenswrapper[4978]: I0225 09:01:52.058783 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-4f6e-account-create-update-ldpx7"] Feb 25 09:01:52 crc kubenswrapper[4978]: I0225 09:01:52.072038 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-96h4w"] Feb 25 09:01:52 crc kubenswrapper[4978]: I0225 09:01:52.082422 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-4f6e-account-create-update-ldpx7"] Feb 25 09:01:53 crc kubenswrapper[4978]: I0225 09:01:53.347569 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a0d295e-5470-4e5d-aeac-04bfd4acb5a7" path="/var/lib/kubelet/pods/9a0d295e-5470-4e5d-aeac-04bfd4acb5a7/volumes" Feb 25 09:01:53 crc kubenswrapper[4978]: I0225 09:01:53.349529 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eff4497f-701f-4a54-afd0-b34020a27d89" path="/var/lib/kubelet/pods/eff4497f-701f-4a54-afd0-b34020a27d89/volumes" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.177743 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533502-hknv5"] Feb 25 09:02:00 crc kubenswrapper[4978]: E0225 09:02:00.179226 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerName="registry-server" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.179255 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerName="registry-server" Feb 25 09:02:00 crc kubenswrapper[4978]: E0225 09:02:00.179292 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerName="extract-utilities" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.179306 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerName="extract-utilities" Feb 25 09:02:00 crc kubenswrapper[4978]: E0225 09:02:00.179343 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerName="extract-content" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.179358 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerName="extract-content" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.179752 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e24e92a5-c015-4531-beaf-4829ae60cb1d" containerName="registry-server" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.180929 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533502-hknv5" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.183301 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.183657 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.184545 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.189421 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533502-hknv5"] Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.318566 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw22t\" (UniqueName: \"kubernetes.io/projected/eccaaabf-9269-4f30-bc2e-475f3ffb3ef2-kube-api-access-bw22t\") pod \"auto-csr-approver-29533502-hknv5\" (UID: \"eccaaabf-9269-4f30-bc2e-475f3ffb3ef2\") " pod="openshift-infra/auto-csr-approver-29533502-hknv5" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.421099 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw22t\" (UniqueName: \"kubernetes.io/projected/eccaaabf-9269-4f30-bc2e-475f3ffb3ef2-kube-api-access-bw22t\") pod \"auto-csr-approver-29533502-hknv5\" (UID: \"eccaaabf-9269-4f30-bc2e-475f3ffb3ef2\") " pod="openshift-infra/auto-csr-approver-29533502-hknv5" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.453484 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw22t\" (UniqueName: \"kubernetes.io/projected/eccaaabf-9269-4f30-bc2e-475f3ffb3ef2-kube-api-access-bw22t\") pod \"auto-csr-approver-29533502-hknv5\" (UID: \"eccaaabf-9269-4f30-bc2e-475f3ffb3ef2\") " pod="openshift-infra/auto-csr-approver-29533502-hknv5" Feb 25 09:02:00 crc kubenswrapper[4978]: I0225 09:02:00.509754 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533502-hknv5" Feb 25 09:02:01 crc kubenswrapper[4978]: I0225 09:02:01.070032 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533502-hknv5"] Feb 25 09:02:02 crc kubenswrapper[4978]: I0225 09:02:02.023045 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533502-hknv5" event={"ID":"eccaaabf-9269-4f30-bc2e-475f3ffb3ef2","Type":"ContainerStarted","Data":"515a3e3197eee161f9cdebd821ba54b7a38d0d5cfd57d6c05188ce68a64e4721"} Feb 25 09:02:03 crc kubenswrapper[4978]: I0225 09:02:03.035104 4978 generic.go:334] "Generic (PLEG): container finished" podID="eccaaabf-9269-4f30-bc2e-475f3ffb3ef2" containerID="32864be2c22e82a39acdd8fd94f91b2d8b73fb8ef2c7b433857059ca569d0c18" exitCode=0 Feb 25 09:02:03 crc kubenswrapper[4978]: I0225 09:02:03.035203 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533502-hknv5" event={"ID":"eccaaabf-9269-4f30-bc2e-475f3ffb3ef2","Type":"ContainerDied","Data":"32864be2c22e82a39acdd8fd94f91b2d8b73fb8ef2c7b433857059ca569d0c18"} Feb 25 09:02:04 crc kubenswrapper[4978]: I0225 09:02:04.428016 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533502-hknv5" Feb 25 09:02:04 crc kubenswrapper[4978]: I0225 09:02:04.455620 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw22t\" (UniqueName: \"kubernetes.io/projected/eccaaabf-9269-4f30-bc2e-475f3ffb3ef2-kube-api-access-bw22t\") pod \"eccaaabf-9269-4f30-bc2e-475f3ffb3ef2\" (UID: \"eccaaabf-9269-4f30-bc2e-475f3ffb3ef2\") " Feb 25 09:02:04 crc kubenswrapper[4978]: I0225 09:02:04.465805 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eccaaabf-9269-4f30-bc2e-475f3ffb3ef2-kube-api-access-bw22t" (OuterVolumeSpecName: "kube-api-access-bw22t") pod "eccaaabf-9269-4f30-bc2e-475f3ffb3ef2" (UID: "eccaaabf-9269-4f30-bc2e-475f3ffb3ef2"). InnerVolumeSpecName "kube-api-access-bw22t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:02:04 crc kubenswrapper[4978]: I0225 09:02:04.558523 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw22t\" (UniqueName: \"kubernetes.io/projected/eccaaabf-9269-4f30-bc2e-475f3ffb3ef2-kube-api-access-bw22t\") on node \"crc\" DevicePath \"\"" Feb 25 09:02:05 crc kubenswrapper[4978]: I0225 09:02:05.063208 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533502-hknv5" event={"ID":"eccaaabf-9269-4f30-bc2e-475f3ffb3ef2","Type":"ContainerDied","Data":"515a3e3197eee161f9cdebd821ba54b7a38d0d5cfd57d6c05188ce68a64e4721"} Feb 25 09:02:05 crc kubenswrapper[4978]: I0225 09:02:05.063671 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="515a3e3197eee161f9cdebd821ba54b7a38d0d5cfd57d6c05188ce68a64e4721" Feb 25 09:02:05 crc kubenswrapper[4978]: I0225 09:02:05.063291 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533502-hknv5" Feb 25 09:02:05 crc kubenswrapper[4978]: I0225 09:02:05.510441 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533496-9t8t6"] Feb 25 09:02:05 crc kubenswrapper[4978]: I0225 09:02:05.514711 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533496-9t8t6"] Feb 25 09:02:06 crc kubenswrapper[4978]: I0225 09:02:06.051597 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-nndrq"] Feb 25 09:02:06 crc kubenswrapper[4978]: I0225 09:02:06.066652 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-nndrq"] Feb 25 09:02:07 crc kubenswrapper[4978]: I0225 09:02:07.370618 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fce5bd5-727e-4ee2-8e67-1c1c8d731f47" path="/var/lib/kubelet/pods/0fce5bd5-727e-4ee2-8e67-1c1c8d731f47/volumes" Feb 25 09:02:07 crc kubenswrapper[4978]: I0225 09:02:07.374139 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72a0184c-d75c-4ecb-a38a-d793e3c7da1d" path="/var/lib/kubelet/pods/72a0184c-d75c-4ecb-a38a-d793e3c7da1d/volumes" Feb 25 09:02:16 crc kubenswrapper[4978]: I0225 09:02:16.540576 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:02:16 crc kubenswrapper[4978]: I0225 09:02:16.541442 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:02:34 crc kubenswrapper[4978]: I0225 09:02:34.280525 4978 scope.go:117] "RemoveContainer" containerID="7bde35d6ec1dfe86c589a8031091208db9f0cb65836b801550c1e40ddc7b6a88" Feb 25 09:02:34 crc kubenswrapper[4978]: I0225 09:02:34.322719 4978 scope.go:117] "RemoveContainer" containerID="9cdd00f31fccb854d98fb2b38f6893d5de7bb91821633da5d292a7dab0404c3b" Feb 25 09:02:34 crc kubenswrapper[4978]: I0225 09:02:34.397980 4978 scope.go:117] "RemoveContainer" containerID="ad73f201cbe54de15f3284657aad0c76020e7e4c4bf5c97e4612bbc37d4e60c2" Feb 25 09:02:34 crc kubenswrapper[4978]: I0225 09:02:34.461394 4978 scope.go:117] "RemoveContainer" containerID="57f4b3b77fd4a9550a64e37be2a944b45a7c23612eeae1ce3dbb169cd475bf0e" Feb 25 09:02:46 crc kubenswrapper[4978]: I0225 09:02:46.540297 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:02:46 crc kubenswrapper[4978]: I0225 09:02:46.541023 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:03:16 crc kubenswrapper[4978]: I0225 09:03:16.540820 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:03:16 crc kubenswrapper[4978]: I0225 09:03:16.541334 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:03:16 crc kubenswrapper[4978]: I0225 09:03:16.541411 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 09:03:16 crc kubenswrapper[4978]: I0225 09:03:16.542444 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"71fe33810cada46fd9054688d7deaba29b642bf93d0651d2c0a3a47b2f670de1"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 09:03:16 crc kubenswrapper[4978]: I0225 09:03:16.542517 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://71fe33810cada46fd9054688d7deaba29b642bf93d0651d2c0a3a47b2f670de1" gracePeriod=600 Feb 25 09:03:16 crc kubenswrapper[4978]: I0225 09:03:16.975648 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="71fe33810cada46fd9054688d7deaba29b642bf93d0651d2c0a3a47b2f670de1" exitCode=0 Feb 25 09:03:16 crc kubenswrapper[4978]: I0225 09:03:16.975808 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"71fe33810cada46fd9054688d7deaba29b642bf93d0651d2c0a3a47b2f670de1"} Feb 25 09:03:16 crc kubenswrapper[4978]: I0225 09:03:16.975957 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe"} Feb 25 09:03:16 crc kubenswrapper[4978]: I0225 09:03:16.975986 4978 scope.go:117] "RemoveContainer" containerID="de94e1faf764974f1b9db4d32d9357a7826564a05a04f71348be960f64341595" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.183199 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533504-gtft2"] Feb 25 09:04:00 crc kubenswrapper[4978]: E0225 09:04:00.184237 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eccaaabf-9269-4f30-bc2e-475f3ffb3ef2" containerName="oc" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.184252 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="eccaaabf-9269-4f30-bc2e-475f3ffb3ef2" containerName="oc" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.184549 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="eccaaabf-9269-4f30-bc2e-475f3ffb3ef2" containerName="oc" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.185492 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533504-gtft2" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.189621 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.189952 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.190062 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.196771 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533504-gtft2"] Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.275355 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z6h5\" (UniqueName: \"kubernetes.io/projected/57f4189e-c7ed-418b-93e7-76caed0ff2b1-kube-api-access-4z6h5\") pod \"auto-csr-approver-29533504-gtft2\" (UID: \"57f4189e-c7ed-418b-93e7-76caed0ff2b1\") " pod="openshift-infra/auto-csr-approver-29533504-gtft2" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.378613 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z6h5\" (UniqueName: \"kubernetes.io/projected/57f4189e-c7ed-418b-93e7-76caed0ff2b1-kube-api-access-4z6h5\") pod \"auto-csr-approver-29533504-gtft2\" (UID: \"57f4189e-c7ed-418b-93e7-76caed0ff2b1\") " pod="openshift-infra/auto-csr-approver-29533504-gtft2" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.405696 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z6h5\" (UniqueName: \"kubernetes.io/projected/57f4189e-c7ed-418b-93e7-76caed0ff2b1-kube-api-access-4z6h5\") pod \"auto-csr-approver-29533504-gtft2\" (UID: \"57f4189e-c7ed-418b-93e7-76caed0ff2b1\") " pod="openshift-infra/auto-csr-approver-29533504-gtft2" Feb 25 09:04:00 crc kubenswrapper[4978]: I0225 09:04:00.516154 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533504-gtft2" Feb 25 09:04:01 crc kubenswrapper[4978]: I0225 09:04:01.018738 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533504-gtft2"] Feb 25 09:04:01 crc kubenswrapper[4978]: I0225 09:04:01.502086 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533504-gtft2" event={"ID":"57f4189e-c7ed-418b-93e7-76caed0ff2b1","Type":"ContainerStarted","Data":"fac9a092445c0652754e1b37467f3292c82d04846517590063268ed975836e24"} Feb 25 09:04:02 crc kubenswrapper[4978]: I0225 09:04:02.524868 4978 generic.go:334] "Generic (PLEG): container finished" podID="57f4189e-c7ed-418b-93e7-76caed0ff2b1" containerID="0aea4f8aef7964c8da272a2a1195dfd6a729e4fc63301d1665c3e6642c209fe4" exitCode=0 Feb 25 09:04:02 crc kubenswrapper[4978]: I0225 09:04:02.525215 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533504-gtft2" event={"ID":"57f4189e-c7ed-418b-93e7-76caed0ff2b1","Type":"ContainerDied","Data":"0aea4f8aef7964c8da272a2a1195dfd6a729e4fc63301d1665c3e6642c209fe4"} Feb 25 09:04:04 crc kubenswrapper[4978]: I0225 09:04:04.005392 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533504-gtft2" Feb 25 09:04:04 crc kubenswrapper[4978]: I0225 09:04:04.164658 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z6h5\" (UniqueName: \"kubernetes.io/projected/57f4189e-c7ed-418b-93e7-76caed0ff2b1-kube-api-access-4z6h5\") pod \"57f4189e-c7ed-418b-93e7-76caed0ff2b1\" (UID: \"57f4189e-c7ed-418b-93e7-76caed0ff2b1\") " Feb 25 09:04:04 crc kubenswrapper[4978]: I0225 09:04:04.172327 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57f4189e-c7ed-418b-93e7-76caed0ff2b1-kube-api-access-4z6h5" (OuterVolumeSpecName: "kube-api-access-4z6h5") pod "57f4189e-c7ed-418b-93e7-76caed0ff2b1" (UID: "57f4189e-c7ed-418b-93e7-76caed0ff2b1"). InnerVolumeSpecName "kube-api-access-4z6h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:04:04 crc kubenswrapper[4978]: I0225 09:04:04.267246 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z6h5\" (UniqueName: \"kubernetes.io/projected/57f4189e-c7ed-418b-93e7-76caed0ff2b1-kube-api-access-4z6h5\") on node \"crc\" DevicePath \"\"" Feb 25 09:04:04 crc kubenswrapper[4978]: I0225 09:04:04.551301 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533504-gtft2" event={"ID":"57f4189e-c7ed-418b-93e7-76caed0ff2b1","Type":"ContainerDied","Data":"fac9a092445c0652754e1b37467f3292c82d04846517590063268ed975836e24"} Feb 25 09:04:04 crc kubenswrapper[4978]: I0225 09:04:04.551525 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fac9a092445c0652754e1b37467f3292c82d04846517590063268ed975836e24" Feb 25 09:04:04 crc kubenswrapper[4978]: I0225 09:04:04.551336 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533504-gtft2" Feb 25 09:04:05 crc kubenswrapper[4978]: I0225 09:04:05.113172 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533498-rks4h"] Feb 25 09:04:05 crc kubenswrapper[4978]: I0225 09:04:05.128023 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533498-rks4h"] Feb 25 09:04:05 crc kubenswrapper[4978]: I0225 09:04:05.348748 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a921f4a-dc79-4d25-befd-cd15da7fa69e" path="/var/lib/kubelet/pods/7a921f4a-dc79-4d25-befd-cd15da7fa69e/volumes" Feb 25 09:04:31 crc kubenswrapper[4978]: I0225 09:04:31.070107 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-9ssj7"] Feb 25 09:04:31 crc kubenswrapper[4978]: I0225 09:04:31.088207 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-072e-account-create-update-g24l2"] Feb 25 09:04:31 crc kubenswrapper[4978]: I0225 09:04:31.107359 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-072e-account-create-update-g24l2"] Feb 25 09:04:31 crc kubenswrapper[4978]: I0225 09:04:31.119460 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-9ssj7"] Feb 25 09:04:31 crc kubenswrapper[4978]: I0225 09:04:31.344115 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a86f5c5d-c083-48de-802a-14e448d14f86" path="/var/lib/kubelet/pods/a86f5c5d-c083-48de-802a-14e448d14f86/volumes" Feb 25 09:04:31 crc kubenswrapper[4978]: I0225 09:04:31.345411 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9" path="/var/lib/kubelet/pods/b3bd6215-9ff7-4e0a-b137-eb7f358f5ac9/volumes" Feb 25 09:04:34 crc kubenswrapper[4978]: I0225 09:04:34.646699 4978 scope.go:117] "RemoveContainer" containerID="f778e1e7fd9c03303a600b3f9c819c27f3da07697e4eef6b9ee347e1e708d3af" Feb 25 09:04:34 crc kubenswrapper[4978]: I0225 09:04:34.687278 4978 scope.go:117] "RemoveContainer" containerID="9161200378d1f79de6f0c779c891966368613ba5aef3d369d83ba1e49285975c" Feb 25 09:04:34 crc kubenswrapper[4978]: I0225 09:04:34.766847 4978 scope.go:117] "RemoveContainer" containerID="711d8657ef40f52e0b9ad94dd20d364d4a896473ef84ed5755b65a5e7ddc27b8" Feb 25 09:04:47 crc kubenswrapper[4978]: I0225 09:04:47.048959 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-2jmrd"] Feb 25 09:04:47 crc kubenswrapper[4978]: I0225 09:04:47.070841 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-2jmrd"] Feb 25 09:04:47 crc kubenswrapper[4978]: I0225 09:04:47.346998 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67524ebc-c6e6-4d7e-a175-e939f659a952" path="/var/lib/kubelet/pods/67524ebc-c6e6-4d7e-a175-e939f659a952/volumes" Feb 25 09:05:16 crc kubenswrapper[4978]: I0225 09:05:16.540869 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:05:16 crc kubenswrapper[4978]: I0225 09:05:16.541743 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:05:34 crc kubenswrapper[4978]: I0225 09:05:34.939109 4978 scope.go:117] "RemoveContainer" containerID="66755eae1e385188325158cac2f0a047e582d3529a37909f220e5032796601ce" Feb 25 09:05:35 crc kubenswrapper[4978]: I0225 09:05:35.668854 4978 generic.go:334] "Generic (PLEG): container finished" podID="6153aa3b-10fb-4bba-8103-71d3f8347d0d" containerID="2fc601d65c08af1bb46c5b9e8e89c223bb2165d8d4c611fb499b83c6d256e7a3" exitCode=0 Feb 25 09:05:35 crc kubenswrapper[4978]: I0225 09:05:35.668970 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" event={"ID":"6153aa3b-10fb-4bba-8103-71d3f8347d0d","Type":"ContainerDied","Data":"2fc601d65c08af1bb46c5b9e8e89c223bb2165d8d4c611fb499b83c6d256e7a3"} Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.200091 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.386933 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2zzj\" (UniqueName: \"kubernetes.io/projected/6153aa3b-10fb-4bba-8103-71d3f8347d0d-kube-api-access-k2zzj\") pod \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.387069 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-tripleo-cleanup-combined-ca-bundle\") pod \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.387146 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-ssh-key-openstack-cell1\") pod \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.387305 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-inventory\") pod \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\" (UID: \"6153aa3b-10fb-4bba-8103-71d3f8347d0d\") " Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.396039 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "6153aa3b-10fb-4bba-8103-71d3f8347d0d" (UID: "6153aa3b-10fb-4bba-8103-71d3f8347d0d"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.398768 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6153aa3b-10fb-4bba-8103-71d3f8347d0d-kube-api-access-k2zzj" (OuterVolumeSpecName: "kube-api-access-k2zzj") pod "6153aa3b-10fb-4bba-8103-71d3f8347d0d" (UID: "6153aa3b-10fb-4bba-8103-71d3f8347d0d"). InnerVolumeSpecName "kube-api-access-k2zzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.437187 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-inventory" (OuterVolumeSpecName: "inventory") pod "6153aa3b-10fb-4bba-8103-71d3f8347d0d" (UID: "6153aa3b-10fb-4bba-8103-71d3f8347d0d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.440244 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "6153aa3b-10fb-4bba-8103-71d3f8347d0d" (UID: "6153aa3b-10fb-4bba-8103-71d3f8347d0d"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.490575 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2zzj\" (UniqueName: \"kubernetes.io/projected/6153aa3b-10fb-4bba-8103-71d3f8347d0d-kube-api-access-k2zzj\") on node \"crc\" DevicePath \"\"" Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.490628 4978 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.490656 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.490687 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6153aa3b-10fb-4bba-8103-71d3f8347d0d-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.698172 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" event={"ID":"6153aa3b-10fb-4bba-8103-71d3f8347d0d","Type":"ContainerDied","Data":"d363620a7a3ceb07796ecd2e17132b42a4fe7c7faddd7b8f59b5a66d49c2fb03"} Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.698231 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d363620a7a3ceb07796ecd2e17132b42a4fe7c7faddd7b8f59b5a66d49c2fb03" Feb 25 09:05:37 crc kubenswrapper[4978]: I0225 09:05:37.698283 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm" Feb 25 09:05:46 crc kubenswrapper[4978]: I0225 09:05:46.540646 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:05:46 crc kubenswrapper[4978]: I0225 09:05:46.542956 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.146924 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533506-kjrl4"] Feb 25 09:06:00 crc kubenswrapper[4978]: E0225 09:06:00.154498 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6153aa3b-10fb-4bba-8103-71d3f8347d0d" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.154641 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6153aa3b-10fb-4bba-8103-71d3f8347d0d" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Feb 25 09:06:00 crc kubenswrapper[4978]: E0225 09:06:00.154744 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57f4189e-c7ed-418b-93e7-76caed0ff2b1" containerName="oc" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.154827 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="57f4189e-c7ed-418b-93e7-76caed0ff2b1" containerName="oc" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.155162 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6153aa3b-10fb-4bba-8103-71d3f8347d0d" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.155298 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="57f4189e-c7ed-418b-93e7-76caed0ff2b1" containerName="oc" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.156277 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533506-kjrl4" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.162635 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.163029 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.163119 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.174467 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533506-kjrl4"] Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.253704 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48jgh\" (UniqueName: \"kubernetes.io/projected/02bb15c4-7c69-4b27-bd36-bc13eb17cdce-kube-api-access-48jgh\") pod \"auto-csr-approver-29533506-kjrl4\" (UID: \"02bb15c4-7c69-4b27-bd36-bc13eb17cdce\") " pod="openshift-infra/auto-csr-approver-29533506-kjrl4" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.355054 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48jgh\" (UniqueName: \"kubernetes.io/projected/02bb15c4-7c69-4b27-bd36-bc13eb17cdce-kube-api-access-48jgh\") pod \"auto-csr-approver-29533506-kjrl4\" (UID: \"02bb15c4-7c69-4b27-bd36-bc13eb17cdce\") " pod="openshift-infra/auto-csr-approver-29533506-kjrl4" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.375608 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48jgh\" (UniqueName: \"kubernetes.io/projected/02bb15c4-7c69-4b27-bd36-bc13eb17cdce-kube-api-access-48jgh\") pod \"auto-csr-approver-29533506-kjrl4\" (UID: \"02bb15c4-7c69-4b27-bd36-bc13eb17cdce\") " pod="openshift-infra/auto-csr-approver-29533506-kjrl4" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.497535 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533506-kjrl4" Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.809440 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533506-kjrl4"] Feb 25 09:06:00 crc kubenswrapper[4978]: W0225 09:06:00.809884 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02bb15c4_7c69_4b27_bd36_bc13eb17cdce.slice/crio-3d997102d367053fda2aa1509609a07f7d4b612b72ea824e2e20ce3f5b88751e WatchSource:0}: Error finding container 3d997102d367053fda2aa1509609a07f7d4b612b72ea824e2e20ce3f5b88751e: Status 404 returned error can't find the container with id 3d997102d367053fda2aa1509609a07f7d4b612b72ea824e2e20ce3f5b88751e Feb 25 09:06:00 crc kubenswrapper[4978]: I0225 09:06:00.957483 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533506-kjrl4" event={"ID":"02bb15c4-7c69-4b27-bd36-bc13eb17cdce","Type":"ContainerStarted","Data":"3d997102d367053fda2aa1509609a07f7d4b612b72ea824e2e20ce3f5b88751e"} Feb 25 09:06:02 crc kubenswrapper[4978]: I0225 09:06:02.980745 4978 generic.go:334] "Generic (PLEG): container finished" podID="02bb15c4-7c69-4b27-bd36-bc13eb17cdce" containerID="307a04b86c55772b23c949535a67b84987c7219eada766a80ac3faadd1ca6ef5" exitCode=0 Feb 25 09:06:02 crc kubenswrapper[4978]: I0225 09:06:02.980817 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533506-kjrl4" event={"ID":"02bb15c4-7c69-4b27-bd36-bc13eb17cdce","Type":"ContainerDied","Data":"307a04b86c55772b23c949535a67b84987c7219eada766a80ac3faadd1ca6ef5"} Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.443778 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533506-kjrl4" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.458147 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48jgh\" (UniqueName: \"kubernetes.io/projected/02bb15c4-7c69-4b27-bd36-bc13eb17cdce-kube-api-access-48jgh\") pod \"02bb15c4-7c69-4b27-bd36-bc13eb17cdce\" (UID: \"02bb15c4-7c69-4b27-bd36-bc13eb17cdce\") " Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.472617 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02bb15c4-7c69-4b27-bd36-bc13eb17cdce-kube-api-access-48jgh" (OuterVolumeSpecName: "kube-api-access-48jgh") pod "02bb15c4-7c69-4b27-bd36-bc13eb17cdce" (UID: "02bb15c4-7c69-4b27-bd36-bc13eb17cdce"). InnerVolumeSpecName "kube-api-access-48jgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.561570 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48jgh\" (UniqueName: \"kubernetes.io/projected/02bb15c4-7c69-4b27-bd36-bc13eb17cdce-kube-api-access-48jgh\") on node \"crc\" DevicePath \"\"" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.938718 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-jjn5l"] Feb 25 09:06:04 crc kubenswrapper[4978]: E0225 09:06:04.939546 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02bb15c4-7c69-4b27-bd36-bc13eb17cdce" containerName="oc" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.939584 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="02bb15c4-7c69-4b27-bd36-bc13eb17cdce" containerName="oc" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.939991 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="02bb15c4-7c69-4b27-bd36-bc13eb17cdce" containerName="oc" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.941345 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.945706 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.945882 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.946287 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.946634 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.957583 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-jjn5l"] Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.967894 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-vhgfz"] Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.969524 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.992904 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-inventory\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.992975 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6zkw\" (UniqueName: \"kubernetes.io/projected/060597fe-f0f4-45f6-bea3-5faa8b657c41-kube-api-access-k6zkw\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.993130 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.993229 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.998040 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 09:06:04 crc kubenswrapper[4978]: I0225 09:06:04.998299 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.007746 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533506-kjrl4" event={"ID":"02bb15c4-7c69-4b27-bd36-bc13eb17cdce","Type":"ContainerDied","Data":"3d997102d367053fda2aa1509609a07f7d4b612b72ea824e2e20ce3f5b88751e"} Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.007786 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d997102d367053fda2aa1509609a07f7d4b612b72ea824e2e20ce3f5b88751e" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.007843 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533506-kjrl4" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.024880 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-vhgfz"] Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.094817 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-ssh-key-openstack-networker\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.094891 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.094955 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.095006 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggnw8\" (UniqueName: \"kubernetes.io/projected/c811c196-e2b2-4230-b91d-89c22efbd036-kube-api-access-ggnw8\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.095034 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-inventory\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.095083 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6zkw\" (UniqueName: \"kubernetes.io/projected/060597fe-f0f4-45f6-bea3-5faa8b657c41-kube-api-access-k6zkw\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.095187 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-inventory\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.095216 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.102665 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.104347 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.109952 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-inventory\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.120870 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6zkw\" (UniqueName: \"kubernetes.io/projected/060597fe-f0f4-45f6-bea3-5faa8b657c41-kube-api-access-k6zkw\") pod \"bootstrap-openstack-openstack-cell1-jjn5l\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.196945 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-ssh-key-openstack-networker\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.197023 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.197081 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggnw8\" (UniqueName: \"kubernetes.io/projected/c811c196-e2b2-4230-b91d-89c22efbd036-kube-api-access-ggnw8\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.197175 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-inventory\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.203034 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.203154 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-ssh-key-openstack-networker\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.203471 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-inventory\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.214231 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggnw8\" (UniqueName: \"kubernetes.io/projected/c811c196-e2b2-4230-b91d-89c22efbd036-kube-api-access-ggnw8\") pod \"bootstrap-openstack-openstack-networker-vhgfz\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.318021 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.330944 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.532979 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533500-gg4p9"] Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.541610 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533500-gg4p9"] Feb 25 09:06:05 crc kubenswrapper[4978]: I0225 09:06:05.978274 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-jjn5l"] Feb 25 09:06:06 crc kubenswrapper[4978]: I0225 09:06:06.021116 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" event={"ID":"060597fe-f0f4-45f6-bea3-5faa8b657c41","Type":"ContainerStarted","Data":"8f7d5d4f3e47ae093389a43230a2b3a59011fd465c8c2b0c3706791647335daf"} Feb 25 09:06:06 crc kubenswrapper[4978]: W0225 09:06:06.573883 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc811c196_e2b2_4230_b91d_89c22efbd036.slice/crio-f9f4af1842ef3f0f2cefa933ee017be84a2d21536910d75a698bb246a0e9b429 WatchSource:0}: Error finding container f9f4af1842ef3f0f2cefa933ee017be84a2d21536910d75a698bb246a0e9b429: Status 404 returned error can't find the container with id f9f4af1842ef3f0f2cefa933ee017be84a2d21536910d75a698bb246a0e9b429 Feb 25 09:06:06 crc kubenswrapper[4978]: I0225 09:06:06.580453 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-networker-vhgfz"] Feb 25 09:06:07 crc kubenswrapper[4978]: I0225 09:06:07.034198 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" event={"ID":"060597fe-f0f4-45f6-bea3-5faa8b657c41","Type":"ContainerStarted","Data":"9c3261c4a30cfc30943e7771b5e245dda40f467ee76a25d93bb5385a4bfee96f"} Feb 25 09:06:07 crc kubenswrapper[4978]: I0225 09:06:07.036341 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" event={"ID":"c811c196-e2b2-4230-b91d-89c22efbd036","Type":"ContainerStarted","Data":"f9f4af1842ef3f0f2cefa933ee017be84a2d21536910d75a698bb246a0e9b429"} Feb 25 09:06:07 crc kubenswrapper[4978]: I0225 09:06:07.057117 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" podStartSLOduration=2.621996404 podStartE2EDuration="3.05709849s" podCreationTimestamp="2026-02-25 09:06:04 +0000 UTC" firstStartedPulling="2026-02-25 09:06:05.994329325 +0000 UTC m=+8459.433585784" lastFinishedPulling="2026-02-25 09:06:06.429431371 +0000 UTC m=+8459.868687870" observedRunningTime="2026-02-25 09:06:07.051449875 +0000 UTC m=+8460.490706364" watchObservedRunningTime="2026-02-25 09:06:07.05709849 +0000 UTC m=+8460.496354969" Feb 25 09:06:07 crc kubenswrapper[4978]: I0225 09:06:07.347714 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="259ed9e0-f44d-49d9-b800-df10bf4df3da" path="/var/lib/kubelet/pods/259ed9e0-f44d-49d9-b800-df10bf4df3da/volumes" Feb 25 09:06:08 crc kubenswrapper[4978]: I0225 09:06:08.049398 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" event={"ID":"c811c196-e2b2-4230-b91d-89c22efbd036","Type":"ContainerStarted","Data":"67cfc1604a63b1d5d17bb42165abcd99edf423e2df9648339949403fe1460806"} Feb 25 09:06:08 crc kubenswrapper[4978]: I0225 09:06:08.078281 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" podStartSLOduration=3.717297576 podStartE2EDuration="4.078253114s" podCreationTimestamp="2026-02-25 09:06:04 +0000 UTC" firstStartedPulling="2026-02-25 09:06:06.57598696 +0000 UTC m=+8460.015243439" lastFinishedPulling="2026-02-25 09:06:06.936942518 +0000 UTC m=+8460.376198977" observedRunningTime="2026-02-25 09:06:08.078044098 +0000 UTC m=+8461.517300567" watchObservedRunningTime="2026-02-25 09:06:08.078253114 +0000 UTC m=+8461.517509613" Feb 25 09:06:16 crc kubenswrapper[4978]: I0225 09:06:16.540522 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:06:16 crc kubenswrapper[4978]: I0225 09:06:16.541138 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:06:16 crc kubenswrapper[4978]: I0225 09:06:16.541208 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 09:06:16 crc kubenswrapper[4978]: I0225 09:06:16.543483 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 09:06:16 crc kubenswrapper[4978]: I0225 09:06:16.543656 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" gracePeriod=600 Feb 25 09:06:16 crc kubenswrapper[4978]: E0225 09:06:16.683118 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:06:17 crc kubenswrapper[4978]: I0225 09:06:17.191331 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" exitCode=0 Feb 25 09:06:17 crc kubenswrapper[4978]: I0225 09:06:17.191478 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe"} Feb 25 09:06:17 crc kubenswrapper[4978]: I0225 09:06:17.191752 4978 scope.go:117] "RemoveContainer" containerID="71fe33810cada46fd9054688d7deaba29b642bf93d0651d2c0a3a47b2f670de1" Feb 25 09:06:17 crc kubenswrapper[4978]: I0225 09:06:17.193096 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:06:17 crc kubenswrapper[4978]: E0225 09:06:17.196422 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:06:29 crc kubenswrapper[4978]: I0225 09:06:29.327726 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:06:29 crc kubenswrapper[4978]: E0225 09:06:29.328766 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:06:35 crc kubenswrapper[4978]: I0225 09:06:35.051432 4978 scope.go:117] "RemoveContainer" containerID="4b527beba7c63d5d7de5c8ff43fe3202c98f6f0377a97e19e3cca162a61c297f" Feb 25 09:06:40 crc kubenswrapper[4978]: I0225 09:06:40.329078 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:06:40 crc kubenswrapper[4978]: E0225 09:06:40.330154 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.234189 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w5cjf"] Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.239011 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.266600 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w5cjf"] Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.310022 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v26fs\" (UniqueName: \"kubernetes.io/projected/df43f3d1-d6d9-478f-96be-83201bccdd24-kube-api-access-v26fs\") pod \"redhat-operators-w5cjf\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.310184 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-catalog-content\") pod \"redhat-operators-w5cjf\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.310219 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-utilities\") pod \"redhat-operators-w5cjf\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.412377 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-utilities\") pod \"redhat-operators-w5cjf\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.412508 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v26fs\" (UniqueName: \"kubernetes.io/projected/df43f3d1-d6d9-478f-96be-83201bccdd24-kube-api-access-v26fs\") pod \"redhat-operators-w5cjf\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.412688 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-catalog-content\") pod \"redhat-operators-w5cjf\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.412971 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-utilities\") pod \"redhat-operators-w5cjf\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.413076 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-catalog-content\") pod \"redhat-operators-w5cjf\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.435023 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v26fs\" (UniqueName: \"kubernetes.io/projected/df43f3d1-d6d9-478f-96be-83201bccdd24-kube-api-access-v26fs\") pod \"redhat-operators-w5cjf\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:44 crc kubenswrapper[4978]: I0225 09:06:44.564875 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:45 crc kubenswrapper[4978]: I0225 09:06:45.049773 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w5cjf"] Feb 25 09:06:45 crc kubenswrapper[4978]: W0225 09:06:45.056307 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf43f3d1_d6d9_478f_96be_83201bccdd24.slice/crio-220f4b0a68aa2d19c9088aeb415f5ef19b51497c39820ce639f02c28456b9436 WatchSource:0}: Error finding container 220f4b0a68aa2d19c9088aeb415f5ef19b51497c39820ce639f02c28456b9436: Status 404 returned error can't find the container with id 220f4b0a68aa2d19c9088aeb415f5ef19b51497c39820ce639f02c28456b9436 Feb 25 09:06:45 crc kubenswrapper[4978]: I0225 09:06:45.517909 4978 generic.go:334] "Generic (PLEG): container finished" podID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerID="d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3" exitCode=0 Feb 25 09:06:45 crc kubenswrapper[4978]: I0225 09:06:45.518022 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5cjf" event={"ID":"df43f3d1-d6d9-478f-96be-83201bccdd24","Type":"ContainerDied","Data":"d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3"} Feb 25 09:06:45 crc kubenswrapper[4978]: I0225 09:06:45.518225 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5cjf" event={"ID":"df43f3d1-d6d9-478f-96be-83201bccdd24","Type":"ContainerStarted","Data":"220f4b0a68aa2d19c9088aeb415f5ef19b51497c39820ce639f02c28456b9436"} Feb 25 09:06:45 crc kubenswrapper[4978]: I0225 09:06:45.520795 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:06:46 crc kubenswrapper[4978]: I0225 09:06:46.535668 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5cjf" event={"ID":"df43f3d1-d6d9-478f-96be-83201bccdd24","Type":"ContainerStarted","Data":"6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63"} Feb 25 09:06:51 crc kubenswrapper[4978]: I0225 09:06:51.328431 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:06:51 crc kubenswrapper[4978]: E0225 09:06:51.330094 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:06:51 crc kubenswrapper[4978]: I0225 09:06:51.593126 4978 generic.go:334] "Generic (PLEG): container finished" podID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerID="6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63" exitCode=0 Feb 25 09:06:51 crc kubenswrapper[4978]: I0225 09:06:51.593189 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5cjf" event={"ID":"df43f3d1-d6d9-478f-96be-83201bccdd24","Type":"ContainerDied","Data":"6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63"} Feb 25 09:06:52 crc kubenswrapper[4978]: I0225 09:06:52.609791 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5cjf" event={"ID":"df43f3d1-d6d9-478f-96be-83201bccdd24","Type":"ContainerStarted","Data":"acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8"} Feb 25 09:06:52 crc kubenswrapper[4978]: I0225 09:06:52.653200 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w5cjf" podStartSLOduration=2.016402784 podStartE2EDuration="8.653166442s" podCreationTimestamp="2026-02-25 09:06:44 +0000 UTC" firstStartedPulling="2026-02-25 09:06:45.520561338 +0000 UTC m=+8498.959817797" lastFinishedPulling="2026-02-25 09:06:52.157324956 +0000 UTC m=+8505.596581455" observedRunningTime="2026-02-25 09:06:52.637816046 +0000 UTC m=+8506.077072515" watchObservedRunningTime="2026-02-25 09:06:52.653166442 +0000 UTC m=+8506.092422931" Feb 25 09:06:54 crc kubenswrapper[4978]: I0225 09:06:54.565401 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:54 crc kubenswrapper[4978]: I0225 09:06:54.565454 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:06:55 crc kubenswrapper[4978]: I0225 09:06:55.618504 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w5cjf" podUID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerName="registry-server" probeResult="failure" output=< Feb 25 09:06:55 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 09:06:55 crc kubenswrapper[4978]: > Feb 25 09:07:04 crc kubenswrapper[4978]: I0225 09:07:04.650990 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:07:04 crc kubenswrapper[4978]: I0225 09:07:04.750306 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:07:04 crc kubenswrapper[4978]: I0225 09:07:04.912771 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w5cjf"] Feb 25 09:07:05 crc kubenswrapper[4978]: I0225 09:07:05.758705 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w5cjf" podUID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerName="registry-server" containerID="cri-o://acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8" gracePeriod=2 Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.244223 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.327916 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:07:06 crc kubenswrapper[4978]: E0225 09:07:06.328139 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.347654 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v26fs\" (UniqueName: \"kubernetes.io/projected/df43f3d1-d6d9-478f-96be-83201bccdd24-kube-api-access-v26fs\") pod \"df43f3d1-d6d9-478f-96be-83201bccdd24\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.347990 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-utilities\") pod \"df43f3d1-d6d9-478f-96be-83201bccdd24\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.348115 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-catalog-content\") pod \"df43f3d1-d6d9-478f-96be-83201bccdd24\" (UID: \"df43f3d1-d6d9-478f-96be-83201bccdd24\") " Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.348867 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-utilities" (OuterVolumeSpecName: "utilities") pod "df43f3d1-d6d9-478f-96be-83201bccdd24" (UID: "df43f3d1-d6d9-478f-96be-83201bccdd24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.349441 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.362701 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df43f3d1-d6d9-478f-96be-83201bccdd24-kube-api-access-v26fs" (OuterVolumeSpecName: "kube-api-access-v26fs") pod "df43f3d1-d6d9-478f-96be-83201bccdd24" (UID: "df43f3d1-d6d9-478f-96be-83201bccdd24"). InnerVolumeSpecName "kube-api-access-v26fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.453576 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v26fs\" (UniqueName: \"kubernetes.io/projected/df43f3d1-d6d9-478f-96be-83201bccdd24-kube-api-access-v26fs\") on node \"crc\" DevicePath \"\"" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.502010 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df43f3d1-d6d9-478f-96be-83201bccdd24" (UID: "df43f3d1-d6d9-478f-96be-83201bccdd24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.556399 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df43f3d1-d6d9-478f-96be-83201bccdd24-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.774465 4978 generic.go:334] "Generic (PLEG): container finished" podID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerID="acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8" exitCode=0 Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.774516 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5cjf" event={"ID":"df43f3d1-d6d9-478f-96be-83201bccdd24","Type":"ContainerDied","Data":"acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8"} Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.775530 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5cjf" event={"ID":"df43f3d1-d6d9-478f-96be-83201bccdd24","Type":"ContainerDied","Data":"220f4b0a68aa2d19c9088aeb415f5ef19b51497c39820ce639f02c28456b9436"} Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.774597 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5cjf" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.775569 4978 scope.go:117] "RemoveContainer" containerID="acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.803981 4978 scope.go:117] "RemoveContainer" containerID="6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.829224 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w5cjf"] Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.842397 4978 scope.go:117] "RemoveContainer" containerID="d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.889192 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w5cjf"] Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.905207 4978 scope.go:117] "RemoveContainer" containerID="acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8" Feb 25 09:07:06 crc kubenswrapper[4978]: E0225 09:07:06.905765 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8\": container with ID starting with acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8 not found: ID does not exist" containerID="acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.905812 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8"} err="failed to get container status \"acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8\": rpc error: code = NotFound desc = could not find container \"acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8\": container with ID starting with acd9374fa38c2e71210d39176911f229c816f28e5b4f425a4464ad20b26e93a8 not found: ID does not exist" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.905837 4978 scope.go:117] "RemoveContainer" containerID="6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63" Feb 25 09:07:06 crc kubenswrapper[4978]: E0225 09:07:06.906259 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63\": container with ID starting with 6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63 not found: ID does not exist" containerID="6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.906290 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63"} err="failed to get container status \"6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63\": rpc error: code = NotFound desc = could not find container \"6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63\": container with ID starting with 6b2fbb8057cb90bb28a344b070ff02dcabb2dd5af130064fbad60ce28f536d63 not found: ID does not exist" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.906310 4978 scope.go:117] "RemoveContainer" containerID="d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3" Feb 25 09:07:06 crc kubenswrapper[4978]: E0225 09:07:06.906691 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3\": container with ID starting with d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3 not found: ID does not exist" containerID="d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3" Feb 25 09:07:06 crc kubenswrapper[4978]: I0225 09:07:06.906707 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3"} err="failed to get container status \"d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3\": rpc error: code = NotFound desc = could not find container \"d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3\": container with ID starting with d4edb369852f48cee5ad1d2298f403e9a033b92ef8c5e9b3d1b7a5087c1f56e3 not found: ID does not exist" Feb 25 09:07:07 crc kubenswrapper[4978]: I0225 09:07:07.349935 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df43f3d1-d6d9-478f-96be-83201bccdd24" path="/var/lib/kubelet/pods/df43f3d1-d6d9-478f-96be-83201bccdd24/volumes" Feb 25 09:07:19 crc kubenswrapper[4978]: I0225 09:07:19.327679 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:07:19 crc kubenswrapper[4978]: E0225 09:07:19.329479 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:07:33 crc kubenswrapper[4978]: I0225 09:07:33.327133 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:07:33 crc kubenswrapper[4978]: E0225 09:07:33.328079 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:07:35 crc kubenswrapper[4978]: I0225 09:07:35.182169 4978 scope.go:117] "RemoveContainer" containerID="12b6b8ef135eb5a078d696152786ba835f6511018a32c8418df6284f34c7aff3" Feb 25 09:07:45 crc kubenswrapper[4978]: I0225 09:07:45.327791 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:07:45 crc kubenswrapper[4978]: E0225 09:07:45.328437 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:07:58 crc kubenswrapper[4978]: I0225 09:07:58.328327 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:07:58 crc kubenswrapper[4978]: E0225 09:07:58.330388 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.140902 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533508-28cvw"] Feb 25 09:08:00 crc kubenswrapper[4978]: E0225 09:08:00.141836 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerName="extract-content" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.141850 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerName="extract-content" Feb 25 09:08:00 crc kubenswrapper[4978]: E0225 09:08:00.141872 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerName="registry-server" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.141878 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerName="registry-server" Feb 25 09:08:00 crc kubenswrapper[4978]: E0225 09:08:00.141894 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerName="extract-utilities" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.141900 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerName="extract-utilities" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.142065 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="df43f3d1-d6d9-478f-96be-83201bccdd24" containerName="registry-server" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.142758 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533508-28cvw" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.145906 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.146060 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.146332 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.153394 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533508-28cvw"] Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.230691 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfhtb\" (UniqueName: \"kubernetes.io/projected/4d7e4fd9-5585-44d8-9498-dd11cec75fa8-kube-api-access-kfhtb\") pod \"auto-csr-approver-29533508-28cvw\" (UID: \"4d7e4fd9-5585-44d8-9498-dd11cec75fa8\") " pod="openshift-infra/auto-csr-approver-29533508-28cvw" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.332340 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfhtb\" (UniqueName: \"kubernetes.io/projected/4d7e4fd9-5585-44d8-9498-dd11cec75fa8-kube-api-access-kfhtb\") pod \"auto-csr-approver-29533508-28cvw\" (UID: \"4d7e4fd9-5585-44d8-9498-dd11cec75fa8\") " pod="openshift-infra/auto-csr-approver-29533508-28cvw" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.369149 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfhtb\" (UniqueName: \"kubernetes.io/projected/4d7e4fd9-5585-44d8-9498-dd11cec75fa8-kube-api-access-kfhtb\") pod \"auto-csr-approver-29533508-28cvw\" (UID: \"4d7e4fd9-5585-44d8-9498-dd11cec75fa8\") " pod="openshift-infra/auto-csr-approver-29533508-28cvw" Feb 25 09:08:00 crc kubenswrapper[4978]: I0225 09:08:00.464028 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533508-28cvw" Feb 25 09:08:01 crc kubenswrapper[4978]: W0225 09:08:01.018150 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d7e4fd9_5585_44d8_9498_dd11cec75fa8.slice/crio-f3e510e6a96434c800e0ba1052afd7d2e9b45cadb2c6b26d36c408857669d0f5 WatchSource:0}: Error finding container f3e510e6a96434c800e0ba1052afd7d2e9b45cadb2c6b26d36c408857669d0f5: Status 404 returned error can't find the container with id f3e510e6a96434c800e0ba1052afd7d2e9b45cadb2c6b26d36c408857669d0f5 Feb 25 09:08:01 crc kubenswrapper[4978]: I0225 09:08:01.025526 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533508-28cvw"] Feb 25 09:08:01 crc kubenswrapper[4978]: I0225 09:08:01.434272 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533508-28cvw" event={"ID":"4d7e4fd9-5585-44d8-9498-dd11cec75fa8","Type":"ContainerStarted","Data":"f3e510e6a96434c800e0ba1052afd7d2e9b45cadb2c6b26d36c408857669d0f5"} Feb 25 09:08:02 crc kubenswrapper[4978]: I0225 09:08:02.445236 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533508-28cvw" event={"ID":"4d7e4fd9-5585-44d8-9498-dd11cec75fa8","Type":"ContainerStarted","Data":"dbd06cbbc2d528d2b04a966950f772bc7c45d2eac157596ed7d04f560338dd56"} Feb 25 09:08:02 crc kubenswrapper[4978]: I0225 09:08:02.465225 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533508-28cvw" podStartSLOduration=1.628385768 podStartE2EDuration="2.465196143s" podCreationTimestamp="2026-02-25 09:08:00 +0000 UTC" firstStartedPulling="2026-02-25 09:08:01.020690777 +0000 UTC m=+8574.459947236" lastFinishedPulling="2026-02-25 09:08:01.857501112 +0000 UTC m=+8575.296757611" observedRunningTime="2026-02-25 09:08:02.45703167 +0000 UTC m=+8575.896288149" watchObservedRunningTime="2026-02-25 09:08:02.465196143 +0000 UTC m=+8575.904452622" Feb 25 09:08:03 crc kubenswrapper[4978]: I0225 09:08:03.458236 4978 generic.go:334] "Generic (PLEG): container finished" podID="4d7e4fd9-5585-44d8-9498-dd11cec75fa8" containerID="dbd06cbbc2d528d2b04a966950f772bc7c45d2eac157596ed7d04f560338dd56" exitCode=0 Feb 25 09:08:03 crc kubenswrapper[4978]: I0225 09:08:03.458279 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533508-28cvw" event={"ID":"4d7e4fd9-5585-44d8-9498-dd11cec75fa8","Type":"ContainerDied","Data":"dbd06cbbc2d528d2b04a966950f772bc7c45d2eac157596ed7d04f560338dd56"} Feb 25 09:08:04 crc kubenswrapper[4978]: I0225 09:08:04.847224 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533508-28cvw" Feb 25 09:08:05 crc kubenswrapper[4978]: I0225 09:08:05.037876 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfhtb\" (UniqueName: \"kubernetes.io/projected/4d7e4fd9-5585-44d8-9498-dd11cec75fa8-kube-api-access-kfhtb\") pod \"4d7e4fd9-5585-44d8-9498-dd11cec75fa8\" (UID: \"4d7e4fd9-5585-44d8-9498-dd11cec75fa8\") " Feb 25 09:08:05 crc kubenswrapper[4978]: I0225 09:08:05.045979 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d7e4fd9-5585-44d8-9498-dd11cec75fa8-kube-api-access-kfhtb" (OuterVolumeSpecName: "kube-api-access-kfhtb") pod "4d7e4fd9-5585-44d8-9498-dd11cec75fa8" (UID: "4d7e4fd9-5585-44d8-9498-dd11cec75fa8"). InnerVolumeSpecName "kube-api-access-kfhtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:08:05 crc kubenswrapper[4978]: I0225 09:08:05.140350 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfhtb\" (UniqueName: \"kubernetes.io/projected/4d7e4fd9-5585-44d8-9498-dd11cec75fa8-kube-api-access-kfhtb\") on node \"crc\" DevicePath \"\"" Feb 25 09:08:05 crc kubenswrapper[4978]: I0225 09:08:05.477108 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533508-28cvw" event={"ID":"4d7e4fd9-5585-44d8-9498-dd11cec75fa8","Type":"ContainerDied","Data":"f3e510e6a96434c800e0ba1052afd7d2e9b45cadb2c6b26d36c408857669d0f5"} Feb 25 09:08:05 crc kubenswrapper[4978]: I0225 09:08:05.477420 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3e510e6a96434c800e0ba1052afd7d2e9b45cadb2c6b26d36c408857669d0f5" Feb 25 09:08:05 crc kubenswrapper[4978]: I0225 09:08:05.477177 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533508-28cvw" Feb 25 09:08:05 crc kubenswrapper[4978]: I0225 09:08:05.533679 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533502-hknv5"] Feb 25 09:08:05 crc kubenswrapper[4978]: I0225 09:08:05.542027 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533502-hknv5"] Feb 25 09:08:07 crc kubenswrapper[4978]: I0225 09:08:07.342736 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eccaaabf-9269-4f30-bc2e-475f3ffb3ef2" path="/var/lib/kubelet/pods/eccaaabf-9269-4f30-bc2e-475f3ffb3ef2/volumes" Feb 25 09:08:10 crc kubenswrapper[4978]: I0225 09:08:10.328833 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:08:10 crc kubenswrapper[4978]: E0225 09:08:10.329519 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:08:23 crc kubenswrapper[4978]: I0225 09:08:23.329117 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:08:23 crc kubenswrapper[4978]: E0225 09:08:23.330303 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:08:35 crc kubenswrapper[4978]: I0225 09:08:35.242075 4978 scope.go:117] "RemoveContainer" containerID="1c3227fd76b1a3f448d220a82b48541757b91a286314c549037e2d5a1348a93b" Feb 25 09:08:35 crc kubenswrapper[4978]: I0225 09:08:35.282463 4978 scope.go:117] "RemoveContainer" containerID="0bf0492e49d183958582013bd2e1ac8b4e6705126edfc98fad0909b63aa4c491" Feb 25 09:08:35 crc kubenswrapper[4978]: I0225 09:08:35.321460 4978 scope.go:117] "RemoveContainer" containerID="32864be2c22e82a39acdd8fd94f91b2d8b73fb8ef2c7b433857059ca569d0c18" Feb 25 09:08:36 crc kubenswrapper[4978]: I0225 09:08:36.328203 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:08:36 crc kubenswrapper[4978]: E0225 09:08:36.328768 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.052138 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5r42w"] Feb 25 09:08:41 crc kubenswrapper[4978]: E0225 09:08:41.053160 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7e4fd9-5585-44d8-9498-dd11cec75fa8" containerName="oc" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.053176 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7e4fd9-5585-44d8-9498-dd11cec75fa8" containerName="oc" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.053424 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7e4fd9-5585-44d8-9498-dd11cec75fa8" containerName="oc" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.054833 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.066028 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5r42w"] Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.140208 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsxx9\" (UniqueName: \"kubernetes.io/projected/1e092c44-d603-4b08-b2ed-008d6fc75c03-kube-api-access-hsxx9\") pod \"certified-operators-5r42w\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.140392 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-catalog-content\") pod \"certified-operators-5r42w\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.140871 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-utilities\") pod \"certified-operators-5r42w\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.243150 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-catalog-content\") pod \"certified-operators-5r42w\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.243219 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-utilities\") pod \"certified-operators-5r42w\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.243302 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsxx9\" (UniqueName: \"kubernetes.io/projected/1e092c44-d603-4b08-b2ed-008d6fc75c03-kube-api-access-hsxx9\") pod \"certified-operators-5r42w\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.244024 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-catalog-content\") pod \"certified-operators-5r42w\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.244263 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-utilities\") pod \"certified-operators-5r42w\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.264123 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsxx9\" (UniqueName: \"kubernetes.io/projected/1e092c44-d603-4b08-b2ed-008d6fc75c03-kube-api-access-hsxx9\") pod \"certified-operators-5r42w\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.379218 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.876652 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5r42w"] Feb 25 09:08:41 crc kubenswrapper[4978]: I0225 09:08:41.893401 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r42w" event={"ID":"1e092c44-d603-4b08-b2ed-008d6fc75c03","Type":"ContainerStarted","Data":"148eb8b0b6bac1b33f767b2b333ab97f9d8b68f6189ce1981162c5c03919ca0c"} Feb 25 09:08:42 crc kubenswrapper[4978]: I0225 09:08:42.905541 4978 generic.go:334] "Generic (PLEG): container finished" podID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerID="95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386" exitCode=0 Feb 25 09:08:42 crc kubenswrapper[4978]: I0225 09:08:42.905634 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r42w" event={"ID":"1e092c44-d603-4b08-b2ed-008d6fc75c03","Type":"ContainerDied","Data":"95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386"} Feb 25 09:08:43 crc kubenswrapper[4978]: I0225 09:08:43.921607 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r42w" event={"ID":"1e092c44-d603-4b08-b2ed-008d6fc75c03","Type":"ContainerStarted","Data":"7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a"} Feb 25 09:08:44 crc kubenswrapper[4978]: I0225 09:08:44.937167 4978 generic.go:334] "Generic (PLEG): container finished" podID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerID="7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a" exitCode=0 Feb 25 09:08:44 crc kubenswrapper[4978]: I0225 09:08:44.937233 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r42w" event={"ID":"1e092c44-d603-4b08-b2ed-008d6fc75c03","Type":"ContainerDied","Data":"7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a"} Feb 25 09:08:45 crc kubenswrapper[4978]: I0225 09:08:45.986587 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r42w" event={"ID":"1e092c44-d603-4b08-b2ed-008d6fc75c03","Type":"ContainerStarted","Data":"c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6"} Feb 25 09:08:46 crc kubenswrapper[4978]: I0225 09:08:46.010943 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5r42w" podStartSLOduration=2.563101238 podStartE2EDuration="5.010925547s" podCreationTimestamp="2026-02-25 09:08:41 +0000 UTC" firstStartedPulling="2026-02-25 09:08:42.907769762 +0000 UTC m=+8616.347026231" lastFinishedPulling="2026-02-25 09:08:45.355594071 +0000 UTC m=+8618.794850540" observedRunningTime="2026-02-25 09:08:46.008290425 +0000 UTC m=+8619.447546884" watchObservedRunningTime="2026-02-25 09:08:46.010925547 +0000 UTC m=+8619.450182006" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.039940 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5786c"] Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.050194 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5786c"] Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.050339 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.222065 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-catalog-content\") pod \"redhat-marketplace-5786c\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.222437 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhprc\" (UniqueName: \"kubernetes.io/projected/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-kube-api-access-vhprc\") pod \"redhat-marketplace-5786c\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.222765 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-utilities\") pod \"redhat-marketplace-5786c\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.325647 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-catalog-content\") pod \"redhat-marketplace-5786c\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.325826 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhprc\" (UniqueName: \"kubernetes.io/projected/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-kube-api-access-vhprc\") pod \"redhat-marketplace-5786c\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.325926 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-utilities\") pod \"redhat-marketplace-5786c\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.326223 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-catalog-content\") pod \"redhat-marketplace-5786c\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.326406 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-utilities\") pod \"redhat-marketplace-5786c\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.327322 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:08:48 crc kubenswrapper[4978]: E0225 09:08:48.327617 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.347880 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhprc\" (UniqueName: \"kubernetes.io/projected/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-kube-api-access-vhprc\") pod \"redhat-marketplace-5786c\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.389019 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:48 crc kubenswrapper[4978]: I0225 09:08:48.890684 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5786c"] Feb 25 09:08:49 crc kubenswrapper[4978]: I0225 09:08:49.013227 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5786c" event={"ID":"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79","Type":"ContainerStarted","Data":"7fc044b0071432e1e8c3504147b4d8700322c32b1869a73dc0b8b57a7b8bd5e0"} Feb 25 09:08:50 crc kubenswrapper[4978]: I0225 09:08:50.055771 4978 generic.go:334] "Generic (PLEG): container finished" podID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerID="30549a3b33f6ec3dda56bbc9285e86c7d29f4a96d6f67e5b1a2257630e608d6e" exitCode=0 Feb 25 09:08:50 crc kubenswrapper[4978]: I0225 09:08:50.055837 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5786c" event={"ID":"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79","Type":"ContainerDied","Data":"30549a3b33f6ec3dda56bbc9285e86c7d29f4a96d6f67e5b1a2257630e608d6e"} Feb 25 09:08:51 crc kubenswrapper[4978]: I0225 09:08:51.072340 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5786c" event={"ID":"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79","Type":"ContainerStarted","Data":"f37eaa18f90b4dd2d338d71b81ca4b16e44541947a22845767a9f6b1ece4b1a4"} Feb 25 09:08:51 crc kubenswrapper[4978]: I0225 09:08:51.379801 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:51 crc kubenswrapper[4978]: I0225 09:08:51.379855 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:51 crc kubenswrapper[4978]: I0225 09:08:51.460430 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:52 crc kubenswrapper[4978]: I0225 09:08:52.087167 4978 generic.go:334] "Generic (PLEG): container finished" podID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerID="f37eaa18f90b4dd2d338d71b81ca4b16e44541947a22845767a9f6b1ece4b1a4" exitCode=0 Feb 25 09:08:52 crc kubenswrapper[4978]: I0225 09:08:52.087393 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5786c" event={"ID":"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79","Type":"ContainerDied","Data":"f37eaa18f90b4dd2d338d71b81ca4b16e44541947a22845767a9f6b1ece4b1a4"} Feb 25 09:08:52 crc kubenswrapper[4978]: I0225 09:08:52.151466 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:53 crc kubenswrapper[4978]: I0225 09:08:53.098553 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5786c" event={"ID":"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79","Type":"ContainerStarted","Data":"975cdd422cdb7a0dbff5cbd8d5a90ea8450a4015b33bd7cf34b78e556318dc99"} Feb 25 09:08:53 crc kubenswrapper[4978]: I0225 09:08:53.126843 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5786c" podStartSLOduration=2.682073161 podStartE2EDuration="5.126818373s" podCreationTimestamp="2026-02-25 09:08:48 +0000 UTC" firstStartedPulling="2026-02-25 09:08:50.059407497 +0000 UTC m=+8623.498663986" lastFinishedPulling="2026-02-25 09:08:52.504152699 +0000 UTC m=+8625.943409198" observedRunningTime="2026-02-25 09:08:53.119266759 +0000 UTC m=+8626.558523258" watchObservedRunningTime="2026-02-25 09:08:53.126818373 +0000 UTC m=+8626.566074842" Feb 25 09:08:53 crc kubenswrapper[4978]: I0225 09:08:53.834279 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5r42w"] Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.110983 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5r42w" podUID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerName="registry-server" containerID="cri-o://c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6" gracePeriod=2 Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.709289 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.804934 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsxx9\" (UniqueName: \"kubernetes.io/projected/1e092c44-d603-4b08-b2ed-008d6fc75c03-kube-api-access-hsxx9\") pod \"1e092c44-d603-4b08-b2ed-008d6fc75c03\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.805041 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-utilities\") pod \"1e092c44-d603-4b08-b2ed-008d6fc75c03\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.805309 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-catalog-content\") pod \"1e092c44-d603-4b08-b2ed-008d6fc75c03\" (UID: \"1e092c44-d603-4b08-b2ed-008d6fc75c03\") " Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.806160 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-utilities" (OuterVolumeSpecName: "utilities") pod "1e092c44-d603-4b08-b2ed-008d6fc75c03" (UID: "1e092c44-d603-4b08-b2ed-008d6fc75c03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.819783 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e092c44-d603-4b08-b2ed-008d6fc75c03-kube-api-access-hsxx9" (OuterVolumeSpecName: "kube-api-access-hsxx9") pod "1e092c44-d603-4b08-b2ed-008d6fc75c03" (UID: "1e092c44-d603-4b08-b2ed-008d6fc75c03"). InnerVolumeSpecName "kube-api-access-hsxx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.873094 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e092c44-d603-4b08-b2ed-008d6fc75c03" (UID: "1e092c44-d603-4b08-b2ed-008d6fc75c03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.907079 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.907112 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsxx9\" (UniqueName: \"kubernetes.io/projected/1e092c44-d603-4b08-b2ed-008d6fc75c03-kube-api-access-hsxx9\") on node \"crc\" DevicePath \"\"" Feb 25 09:08:54 crc kubenswrapper[4978]: I0225 09:08:54.907127 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e092c44-d603-4b08-b2ed-008d6fc75c03-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.122746 4978 generic.go:334] "Generic (PLEG): container finished" podID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerID="c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6" exitCode=0 Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.122792 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r42w" event={"ID":"1e092c44-d603-4b08-b2ed-008d6fc75c03","Type":"ContainerDied","Data":"c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6"} Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.122838 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r42w" event={"ID":"1e092c44-d603-4b08-b2ed-008d6fc75c03","Type":"ContainerDied","Data":"148eb8b0b6bac1b33f767b2b333ab97f9d8b68f6189ce1981162c5c03919ca0c"} Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.122861 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r42w" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.122874 4978 scope.go:117] "RemoveContainer" containerID="c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.162230 4978 scope.go:117] "RemoveContainer" containerID="7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.179758 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5r42w"] Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.191067 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5r42w"] Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.211549 4978 scope.go:117] "RemoveContainer" containerID="95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.242138 4978 scope.go:117] "RemoveContainer" containerID="c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6" Feb 25 09:08:55 crc kubenswrapper[4978]: E0225 09:08:55.242637 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6\": container with ID starting with c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6 not found: ID does not exist" containerID="c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.242705 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6"} err="failed to get container status \"c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6\": rpc error: code = NotFound desc = could not find container \"c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6\": container with ID starting with c5f60d818e7850faf451bdca00b9a7188b8e8a58a0b8cbeb0d8ab68b6aefacd6 not found: ID does not exist" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.242746 4978 scope.go:117] "RemoveContainer" containerID="7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a" Feb 25 09:08:55 crc kubenswrapper[4978]: E0225 09:08:55.243085 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a\": container with ID starting with 7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a not found: ID does not exist" containerID="7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.243139 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a"} err="failed to get container status \"7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a\": rpc error: code = NotFound desc = could not find container \"7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a\": container with ID starting with 7f8375551bda8aaf18ec3b7c13b2317352721baaa343351ef80c80c50479d84a not found: ID does not exist" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.243176 4978 scope.go:117] "RemoveContainer" containerID="95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386" Feb 25 09:08:55 crc kubenswrapper[4978]: E0225 09:08:55.243478 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386\": container with ID starting with 95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386 not found: ID does not exist" containerID="95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.243510 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386"} err="failed to get container status \"95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386\": rpc error: code = NotFound desc = could not find container \"95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386\": container with ID starting with 95a7f7155d1294dd0886a5fea0fcc797ba211de5dbb0cf2293f9ac5325b3f386 not found: ID does not exist" Feb 25 09:08:55 crc kubenswrapper[4978]: I0225 09:08:55.339532 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e092c44-d603-4b08-b2ed-008d6fc75c03" path="/var/lib/kubelet/pods/1e092c44-d603-4b08-b2ed-008d6fc75c03/volumes" Feb 25 09:08:58 crc kubenswrapper[4978]: I0225 09:08:58.389867 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:58 crc kubenswrapper[4978]: I0225 09:08:58.391554 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:58 crc kubenswrapper[4978]: I0225 09:08:58.460736 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:59 crc kubenswrapper[4978]: I0225 09:08:59.244828 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:08:59 crc kubenswrapper[4978]: I0225 09:08:59.319818 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5786c"] Feb 25 09:09:01 crc kubenswrapper[4978]: I0225 09:09:01.201531 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5786c" podUID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerName="registry-server" containerID="cri-o://975cdd422cdb7a0dbff5cbd8d5a90ea8450a4015b33bd7cf34b78e556318dc99" gracePeriod=2 Feb 25 09:09:01 crc kubenswrapper[4978]: I0225 09:09:01.327953 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:09:01 crc kubenswrapper[4978]: E0225 09:09:01.328542 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.220994 4978 generic.go:334] "Generic (PLEG): container finished" podID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerID="975cdd422cdb7a0dbff5cbd8d5a90ea8450a4015b33bd7cf34b78e556318dc99" exitCode=0 Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.221041 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5786c" event={"ID":"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79","Type":"ContainerDied","Data":"975cdd422cdb7a0dbff5cbd8d5a90ea8450a4015b33bd7cf34b78e556318dc99"} Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.221114 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5786c" event={"ID":"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79","Type":"ContainerDied","Data":"7fc044b0071432e1e8c3504147b4d8700322c32b1869a73dc0b8b57a7b8bd5e0"} Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.221126 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fc044b0071432e1e8c3504147b4d8700322c32b1869a73dc0b8b57a7b8bd5e0" Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.269398 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.370989 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-catalog-content\") pod \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.371082 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhprc\" (UniqueName: \"kubernetes.io/projected/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-kube-api-access-vhprc\") pod \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.371285 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-utilities\") pod \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\" (UID: \"55cac31d-fabb-49b2-a89b-8cd3eeaf7e79\") " Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.372096 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-utilities" (OuterVolumeSpecName: "utilities") pod "55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" (UID: "55cac31d-fabb-49b2-a89b-8cd3eeaf7e79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.376447 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-kube-api-access-vhprc" (OuterVolumeSpecName: "kube-api-access-vhprc") pod "55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" (UID: "55cac31d-fabb-49b2-a89b-8cd3eeaf7e79"). InnerVolumeSpecName "kube-api-access-vhprc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.413437 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" (UID: "55cac31d-fabb-49b2-a89b-8cd3eeaf7e79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.474353 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.474409 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhprc\" (UniqueName: \"kubernetes.io/projected/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-kube-api-access-vhprc\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:02 crc kubenswrapper[4978]: I0225 09:09:02.474425 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:03 crc kubenswrapper[4978]: I0225 09:09:03.230114 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5786c" Feb 25 09:09:03 crc kubenswrapper[4978]: I0225 09:09:03.266905 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5786c"] Feb 25 09:09:03 crc kubenswrapper[4978]: I0225 09:09:03.276505 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5786c"] Feb 25 09:09:03 crc kubenswrapper[4978]: I0225 09:09:03.338091 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" path="/var/lib/kubelet/pods/55cac31d-fabb-49b2-a89b-8cd3eeaf7e79/volumes" Feb 25 09:09:05 crc kubenswrapper[4978]: I0225 09:09:05.251089 4978 generic.go:334] "Generic (PLEG): container finished" podID="060597fe-f0f4-45f6-bea3-5faa8b657c41" containerID="9c3261c4a30cfc30943e7771b5e245dda40f467ee76a25d93bb5385a4bfee96f" exitCode=0 Feb 25 09:09:05 crc kubenswrapper[4978]: I0225 09:09:05.251137 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" event={"ID":"060597fe-f0f4-45f6-bea3-5faa8b657c41","Type":"ContainerDied","Data":"9c3261c4a30cfc30943e7771b5e245dda40f467ee76a25d93bb5385a4bfee96f"} Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.766790 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.863579 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-inventory\") pod \"060597fe-f0f4-45f6-bea3-5faa8b657c41\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.863871 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-ssh-key-openstack-cell1\") pod \"060597fe-f0f4-45f6-bea3-5faa8b657c41\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.864094 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-bootstrap-combined-ca-bundle\") pod \"060597fe-f0f4-45f6-bea3-5faa8b657c41\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.864134 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6zkw\" (UniqueName: \"kubernetes.io/projected/060597fe-f0f4-45f6-bea3-5faa8b657c41-kube-api-access-k6zkw\") pod \"060597fe-f0f4-45f6-bea3-5faa8b657c41\" (UID: \"060597fe-f0f4-45f6-bea3-5faa8b657c41\") " Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.869321 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/060597fe-f0f4-45f6-bea3-5faa8b657c41-kube-api-access-k6zkw" (OuterVolumeSpecName: "kube-api-access-k6zkw") pod "060597fe-f0f4-45f6-bea3-5faa8b657c41" (UID: "060597fe-f0f4-45f6-bea3-5faa8b657c41"). InnerVolumeSpecName "kube-api-access-k6zkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.869491 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "060597fe-f0f4-45f6-bea3-5faa8b657c41" (UID: "060597fe-f0f4-45f6-bea3-5faa8b657c41"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.895059 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-inventory" (OuterVolumeSpecName: "inventory") pod "060597fe-f0f4-45f6-bea3-5faa8b657c41" (UID: "060597fe-f0f4-45f6-bea3-5faa8b657c41"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.899013 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "060597fe-f0f4-45f6-bea3-5faa8b657c41" (UID: "060597fe-f0f4-45f6-bea3-5faa8b657c41"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.966344 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.966400 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.966415 4978 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/060597fe-f0f4-45f6-bea3-5faa8b657c41-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:06 crc kubenswrapper[4978]: I0225 09:09:06.966430 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6zkw\" (UniqueName: \"kubernetes.io/projected/060597fe-f0f4-45f6-bea3-5faa8b657c41-kube-api-access-k6zkw\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.274004 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" event={"ID":"060597fe-f0f4-45f6-bea3-5faa8b657c41","Type":"ContainerDied","Data":"8f7d5d4f3e47ae093389a43230a2b3a59011fd465c8c2b0c3706791647335daf"} Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.274291 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f7d5d4f3e47ae093389a43230a2b3a59011fd465c8c2b0c3706791647335daf" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.274135 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-jjn5l" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.455382 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-hzw4k"] Feb 25 09:09:07 crc kubenswrapper[4978]: E0225 09:09:07.455883 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerName="extract-utilities" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.455905 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerName="extract-utilities" Feb 25 09:09:07 crc kubenswrapper[4978]: E0225 09:09:07.455948 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerName="registry-server" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.455956 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerName="registry-server" Feb 25 09:09:07 crc kubenswrapper[4978]: E0225 09:09:07.455973 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="060597fe-f0f4-45f6-bea3-5faa8b657c41" containerName="bootstrap-openstack-openstack-cell1" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.456011 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="060597fe-f0f4-45f6-bea3-5faa8b657c41" containerName="bootstrap-openstack-openstack-cell1" Feb 25 09:09:07 crc kubenswrapper[4978]: E0225 09:09:07.456030 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerName="extract-content" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.456038 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerName="extract-content" Feb 25 09:09:07 crc kubenswrapper[4978]: E0225 09:09:07.456049 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerName="extract-utilities" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.456057 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerName="extract-utilities" Feb 25 09:09:07 crc kubenswrapper[4978]: E0225 09:09:07.456071 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerName="extract-content" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.456079 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerName="extract-content" Feb 25 09:09:07 crc kubenswrapper[4978]: E0225 09:09:07.456092 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerName="registry-server" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.456099 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerName="registry-server" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.456413 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e092c44-d603-4b08-b2ed-008d6fc75c03" containerName="registry-server" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.456444 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="060597fe-f0f4-45f6-bea3-5faa8b657c41" containerName="bootstrap-openstack-openstack-cell1" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.456457 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="55cac31d-fabb-49b2-a89b-8cd3eeaf7e79" containerName="registry-server" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.458638 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.461091 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.461390 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.492769 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-hzw4k"] Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.578826 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-inventory\") pod \"download-cache-openstack-openstack-cell1-hzw4k\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.578948 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq9h8\" (UniqueName: \"kubernetes.io/projected/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-kube-api-access-sq9h8\") pod \"download-cache-openstack-openstack-cell1-hzw4k\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.579044 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-hzw4k\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.681018 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-inventory\") pod \"download-cache-openstack-openstack-cell1-hzw4k\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.681130 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq9h8\" (UniqueName: \"kubernetes.io/projected/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-kube-api-access-sq9h8\") pod \"download-cache-openstack-openstack-cell1-hzw4k\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.681204 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-hzw4k\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.686150 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-inventory\") pod \"download-cache-openstack-openstack-cell1-hzw4k\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.695444 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-hzw4k\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.701125 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq9h8\" (UniqueName: \"kubernetes.io/projected/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-kube-api-access-sq9h8\") pod \"download-cache-openstack-openstack-cell1-hzw4k\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:07 crc kubenswrapper[4978]: I0225 09:09:07.785032 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:09:08 crc kubenswrapper[4978]: I0225 09:09:08.409327 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-hzw4k"] Feb 25 09:09:09 crc kubenswrapper[4978]: I0225 09:09:09.302741 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" event={"ID":"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e","Type":"ContainerStarted","Data":"c4b438892b682bb0bf2a5c7723d2b1c6c7334352348c4d00227efd0389c729e5"} Feb 25 09:09:09 crc kubenswrapper[4978]: I0225 09:09:09.303137 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" event={"ID":"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e","Type":"ContainerStarted","Data":"67ed1f6c475edb6fdb9b9bfcfab7e26d2937c910e4c3a3ec33b04376609ff2c5"} Feb 25 09:09:09 crc kubenswrapper[4978]: I0225 09:09:09.323955 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" podStartSLOduration=1.776966012 podStartE2EDuration="2.323934702s" podCreationTimestamp="2026-02-25 09:09:07 +0000 UTC" firstStartedPulling="2026-02-25 09:09:08.412176955 +0000 UTC m=+8641.851433434" lastFinishedPulling="2026-02-25 09:09:08.959145635 +0000 UTC m=+8642.398402124" observedRunningTime="2026-02-25 09:09:09.321905669 +0000 UTC m=+8642.761162148" watchObservedRunningTime="2026-02-25 09:09:09.323934702 +0000 UTC m=+8642.763191171" Feb 25 09:09:13 crc kubenswrapper[4978]: I0225 09:09:13.330982 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:09:13 crc kubenswrapper[4978]: E0225 09:09:13.332152 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:09:14 crc kubenswrapper[4978]: I0225 09:09:14.351670 4978 generic.go:334] "Generic (PLEG): container finished" podID="c811c196-e2b2-4230-b91d-89c22efbd036" containerID="67cfc1604a63b1d5d17bb42165abcd99edf423e2df9648339949403fe1460806" exitCode=0 Feb 25 09:09:14 crc kubenswrapper[4978]: I0225 09:09:14.351748 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" event={"ID":"c811c196-e2b2-4230-b91d-89c22efbd036","Type":"ContainerDied","Data":"67cfc1604a63b1d5d17bb42165abcd99edf423e2df9648339949403fe1460806"} Feb 25 09:09:15 crc kubenswrapper[4978]: I0225 09:09:15.982488 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.067714 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-ssh-key-openstack-networker\") pod \"c811c196-e2b2-4230-b91d-89c22efbd036\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.067815 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-bootstrap-combined-ca-bundle\") pod \"c811c196-e2b2-4230-b91d-89c22efbd036\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.067900 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-inventory\") pod \"c811c196-e2b2-4230-b91d-89c22efbd036\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.068063 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggnw8\" (UniqueName: \"kubernetes.io/projected/c811c196-e2b2-4230-b91d-89c22efbd036-kube-api-access-ggnw8\") pod \"c811c196-e2b2-4230-b91d-89c22efbd036\" (UID: \"c811c196-e2b2-4230-b91d-89c22efbd036\") " Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.073113 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c811c196-e2b2-4230-b91d-89c22efbd036" (UID: "c811c196-e2b2-4230-b91d-89c22efbd036"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.088013 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c811c196-e2b2-4230-b91d-89c22efbd036-kube-api-access-ggnw8" (OuterVolumeSpecName: "kube-api-access-ggnw8") pod "c811c196-e2b2-4230-b91d-89c22efbd036" (UID: "c811c196-e2b2-4230-b91d-89c22efbd036"). InnerVolumeSpecName "kube-api-access-ggnw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.096306 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-inventory" (OuterVolumeSpecName: "inventory") pod "c811c196-e2b2-4230-b91d-89c22efbd036" (UID: "c811c196-e2b2-4230-b91d-89c22efbd036"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.114436 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "c811c196-e2b2-4230-b91d-89c22efbd036" (UID: "c811c196-e2b2-4230-b91d-89c22efbd036"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.171320 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.171581 4978 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.171705 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c811c196-e2b2-4230-b91d-89c22efbd036-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.171787 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggnw8\" (UniqueName: \"kubernetes.io/projected/c811c196-e2b2-4230-b91d-89c22efbd036-kube-api-access-ggnw8\") on node \"crc\" DevicePath \"\"" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.383610 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" event={"ID":"c811c196-e2b2-4230-b91d-89c22efbd036","Type":"ContainerDied","Data":"f9f4af1842ef3f0f2cefa933ee017be84a2d21536910d75a698bb246a0e9b429"} Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.383981 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9f4af1842ef3f0f2cefa933ee017be84a2d21536910d75a698bb246a0e9b429" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.383762 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-networker-vhgfz" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.481167 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-networker-77j4w"] Feb 25 09:09:16 crc kubenswrapper[4978]: E0225 09:09:16.481950 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c811c196-e2b2-4230-b91d-89c22efbd036" containerName="bootstrap-openstack-openstack-networker" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.482066 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c811c196-e2b2-4230-b91d-89c22efbd036" containerName="bootstrap-openstack-openstack-networker" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.482430 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c811c196-e2b2-4230-b91d-89c22efbd036" containerName="bootstrap-openstack-openstack-networker" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.483483 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.486536 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.486937 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.510344 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-networker-77j4w"] Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.578236 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-inventory\") pod \"download-cache-openstack-openstack-networker-77j4w\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.578291 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tf47\" (UniqueName: \"kubernetes.io/projected/c8c5a88a-2673-4723-8967-9f209d416174-kube-api-access-7tf47\") pod \"download-cache-openstack-openstack-networker-77j4w\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.578863 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-ssh-key-openstack-networker\") pod \"download-cache-openstack-openstack-networker-77j4w\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.681541 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-ssh-key-openstack-networker\") pod \"download-cache-openstack-openstack-networker-77j4w\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.681674 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-inventory\") pod \"download-cache-openstack-openstack-networker-77j4w\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.681706 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tf47\" (UniqueName: \"kubernetes.io/projected/c8c5a88a-2673-4723-8967-9f209d416174-kube-api-access-7tf47\") pod \"download-cache-openstack-openstack-networker-77j4w\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.687573 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-ssh-key-openstack-networker\") pod \"download-cache-openstack-openstack-networker-77j4w\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.694221 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-inventory\") pod \"download-cache-openstack-openstack-networker-77j4w\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.706564 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tf47\" (UniqueName: \"kubernetes.io/projected/c8c5a88a-2673-4723-8967-9f209d416174-kube-api-access-7tf47\") pod \"download-cache-openstack-openstack-networker-77j4w\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:16 crc kubenswrapper[4978]: I0225 09:09:16.810558 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:09:17 crc kubenswrapper[4978]: I0225 09:09:17.446632 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-networker-77j4w"] Feb 25 09:09:18 crc kubenswrapper[4978]: I0225 09:09:18.410041 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-77j4w" event={"ID":"c8c5a88a-2673-4723-8967-9f209d416174","Type":"ContainerStarted","Data":"38bf3e9a87fedee0ecccd42a2a9c4b2820339a1ec1ba3d3dbbb7448418811d98"} Feb 25 09:09:18 crc kubenswrapper[4978]: I0225 09:09:18.410512 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-77j4w" event={"ID":"c8c5a88a-2673-4723-8967-9f209d416174","Type":"ContainerStarted","Data":"9b49a72e31eb5ffb0560b63e761fab991486de23d68f81c0078431a19be1dfde"} Feb 25 09:09:18 crc kubenswrapper[4978]: I0225 09:09:18.443816 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-networker-77j4w" podStartSLOduration=1.797113483 podStartE2EDuration="2.44378903s" podCreationTimestamp="2026-02-25 09:09:16 +0000 UTC" firstStartedPulling="2026-02-25 09:09:17.448511707 +0000 UTC m=+8650.887768166" lastFinishedPulling="2026-02-25 09:09:18.095187214 +0000 UTC m=+8651.534443713" observedRunningTime="2026-02-25 09:09:18.424905476 +0000 UTC m=+8651.864161925" watchObservedRunningTime="2026-02-25 09:09:18.44378903 +0000 UTC m=+8651.883045499" Feb 25 09:09:25 crc kubenswrapper[4978]: I0225 09:09:25.328149 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:09:25 crc kubenswrapper[4978]: E0225 09:09:25.329080 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:09:39 crc kubenswrapper[4978]: I0225 09:09:39.328278 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:09:39 crc kubenswrapper[4978]: E0225 09:09:39.329110 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:09:51 crc kubenswrapper[4978]: I0225 09:09:51.329123 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:09:51 crc kubenswrapper[4978]: E0225 09:09:51.331644 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:10:00 crc kubenswrapper[4978]: I0225 09:10:00.169570 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533510-9jj6q"] Feb 25 09:10:00 crc kubenswrapper[4978]: I0225 09:10:00.172203 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533510-9jj6q" Feb 25 09:10:00 crc kubenswrapper[4978]: I0225 09:10:00.178289 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:10:00 crc kubenswrapper[4978]: I0225 09:10:00.178664 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:10:00 crc kubenswrapper[4978]: I0225 09:10:00.178838 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:10:00 crc kubenswrapper[4978]: I0225 09:10:00.198556 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533510-9jj6q"] Feb 25 09:10:00 crc kubenswrapper[4978]: I0225 09:10:00.273123 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-255qb\" (UniqueName: \"kubernetes.io/projected/37ae8605-e2a4-41dd-b5bc-5275301ed3be-kube-api-access-255qb\") pod \"auto-csr-approver-29533510-9jj6q\" (UID: \"37ae8605-e2a4-41dd-b5bc-5275301ed3be\") " pod="openshift-infra/auto-csr-approver-29533510-9jj6q" Feb 25 09:10:00 crc kubenswrapper[4978]: I0225 09:10:00.375083 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-255qb\" (UniqueName: \"kubernetes.io/projected/37ae8605-e2a4-41dd-b5bc-5275301ed3be-kube-api-access-255qb\") pod \"auto-csr-approver-29533510-9jj6q\" (UID: \"37ae8605-e2a4-41dd-b5bc-5275301ed3be\") " pod="openshift-infra/auto-csr-approver-29533510-9jj6q" Feb 25 09:10:00 crc kubenswrapper[4978]: I0225 09:10:00.394034 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-255qb\" (UniqueName: \"kubernetes.io/projected/37ae8605-e2a4-41dd-b5bc-5275301ed3be-kube-api-access-255qb\") pod \"auto-csr-approver-29533510-9jj6q\" (UID: \"37ae8605-e2a4-41dd-b5bc-5275301ed3be\") " pod="openshift-infra/auto-csr-approver-29533510-9jj6q" Feb 25 09:10:00 crc kubenswrapper[4978]: I0225 09:10:00.527691 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533510-9jj6q" Feb 25 09:10:01 crc kubenswrapper[4978]: I0225 09:10:01.685855 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533510-9jj6q"] Feb 25 09:10:01 crc kubenswrapper[4978]: I0225 09:10:01.948921 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533510-9jj6q" event={"ID":"37ae8605-e2a4-41dd-b5bc-5275301ed3be","Type":"ContainerStarted","Data":"8d8f50ee1004b00d06796c5c0f31dfceeeb5883c6dd6e280269cbe1dfbd14e8d"} Feb 25 09:10:04 crc kubenswrapper[4978]: I0225 09:10:04.328557 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:10:04 crc kubenswrapper[4978]: E0225 09:10:04.329125 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:10:04 crc kubenswrapper[4978]: I0225 09:10:04.996030 4978 generic.go:334] "Generic (PLEG): container finished" podID="37ae8605-e2a4-41dd-b5bc-5275301ed3be" containerID="aee1214b39366db4378d9a34da91be3f704cd7816af2f78422e056aab38d3a27" exitCode=0 Feb 25 09:10:04 crc kubenswrapper[4978]: I0225 09:10:04.996119 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533510-9jj6q" event={"ID":"37ae8605-e2a4-41dd-b5bc-5275301ed3be","Type":"ContainerDied","Data":"aee1214b39366db4378d9a34da91be3f704cd7816af2f78422e056aab38d3a27"} Feb 25 09:10:06 crc kubenswrapper[4978]: I0225 09:10:06.400433 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533510-9jj6q" Feb 25 09:10:06 crc kubenswrapper[4978]: I0225 09:10:06.520289 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-255qb\" (UniqueName: \"kubernetes.io/projected/37ae8605-e2a4-41dd-b5bc-5275301ed3be-kube-api-access-255qb\") pod \"37ae8605-e2a4-41dd-b5bc-5275301ed3be\" (UID: \"37ae8605-e2a4-41dd-b5bc-5275301ed3be\") " Feb 25 09:10:06 crc kubenswrapper[4978]: I0225 09:10:06.529714 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37ae8605-e2a4-41dd-b5bc-5275301ed3be-kube-api-access-255qb" (OuterVolumeSpecName: "kube-api-access-255qb") pod "37ae8605-e2a4-41dd-b5bc-5275301ed3be" (UID: "37ae8605-e2a4-41dd-b5bc-5275301ed3be"). InnerVolumeSpecName "kube-api-access-255qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:10:06 crc kubenswrapper[4978]: I0225 09:10:06.623041 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-255qb\" (UniqueName: \"kubernetes.io/projected/37ae8605-e2a4-41dd-b5bc-5275301ed3be-kube-api-access-255qb\") on node \"crc\" DevicePath \"\"" Feb 25 09:10:07 crc kubenswrapper[4978]: I0225 09:10:07.015986 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533510-9jj6q" event={"ID":"37ae8605-e2a4-41dd-b5bc-5275301ed3be","Type":"ContainerDied","Data":"8d8f50ee1004b00d06796c5c0f31dfceeeb5883c6dd6e280269cbe1dfbd14e8d"} Feb 25 09:10:07 crc kubenswrapper[4978]: I0225 09:10:07.016297 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d8f50ee1004b00d06796c5c0f31dfceeeb5883c6dd6e280269cbe1dfbd14e8d" Feb 25 09:10:07 crc kubenswrapper[4978]: I0225 09:10:07.016034 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533510-9jj6q" Feb 25 09:10:07 crc kubenswrapper[4978]: I0225 09:10:07.472720 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533504-gtft2"] Feb 25 09:10:07 crc kubenswrapper[4978]: I0225 09:10:07.481663 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533504-gtft2"] Feb 25 09:10:09 crc kubenswrapper[4978]: I0225 09:10:09.346918 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57f4189e-c7ed-418b-93e7-76caed0ff2b1" path="/var/lib/kubelet/pods/57f4189e-c7ed-418b-93e7-76caed0ff2b1/volumes" Feb 25 09:10:16 crc kubenswrapper[4978]: I0225 09:10:16.328224 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:10:16 crc kubenswrapper[4978]: E0225 09:10:16.329474 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:10:25 crc kubenswrapper[4978]: I0225 09:10:25.243128 4978 generic.go:334] "Generic (PLEG): container finished" podID="c8c5a88a-2673-4723-8967-9f209d416174" containerID="38bf3e9a87fedee0ecccd42a2a9c4b2820339a1ec1ba3d3dbbb7448418811d98" exitCode=0 Feb 25 09:10:25 crc kubenswrapper[4978]: I0225 09:10:25.243200 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-77j4w" event={"ID":"c8c5a88a-2673-4723-8967-9f209d416174","Type":"ContainerDied","Data":"38bf3e9a87fedee0ecccd42a2a9c4b2820339a1ec1ba3d3dbbb7448418811d98"} Feb 25 09:10:26 crc kubenswrapper[4978]: I0225 09:10:26.783351 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:10:26 crc kubenswrapper[4978]: I0225 09:10:26.832747 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tf47\" (UniqueName: \"kubernetes.io/projected/c8c5a88a-2673-4723-8967-9f209d416174-kube-api-access-7tf47\") pod \"c8c5a88a-2673-4723-8967-9f209d416174\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " Feb 25 09:10:26 crc kubenswrapper[4978]: I0225 09:10:26.833215 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-ssh-key-openstack-networker\") pod \"c8c5a88a-2673-4723-8967-9f209d416174\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " Feb 25 09:10:26 crc kubenswrapper[4978]: I0225 09:10:26.833320 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-inventory\") pod \"c8c5a88a-2673-4723-8967-9f209d416174\" (UID: \"c8c5a88a-2673-4723-8967-9f209d416174\") " Feb 25 09:10:26 crc kubenswrapper[4978]: I0225 09:10:26.855633 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c5a88a-2673-4723-8967-9f209d416174-kube-api-access-7tf47" (OuterVolumeSpecName: "kube-api-access-7tf47") pod "c8c5a88a-2673-4723-8967-9f209d416174" (UID: "c8c5a88a-2673-4723-8967-9f209d416174"). InnerVolumeSpecName "kube-api-access-7tf47". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:10:26 crc kubenswrapper[4978]: I0225 09:10:26.903136 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "c8c5a88a-2673-4723-8967-9f209d416174" (UID: "c8c5a88a-2673-4723-8967-9f209d416174"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:10:26 crc kubenswrapper[4978]: I0225 09:10:26.905820 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-inventory" (OuterVolumeSpecName: "inventory") pod "c8c5a88a-2673-4723-8967-9f209d416174" (UID: "c8c5a88a-2673-4723-8967-9f209d416174"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:10:26 crc kubenswrapper[4978]: I0225 09:10:26.948254 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tf47\" (UniqueName: \"kubernetes.io/projected/c8c5a88a-2673-4723-8967-9f209d416174-kube-api-access-7tf47\") on node \"crc\" DevicePath \"\"" Feb 25 09:10:26 crc kubenswrapper[4978]: I0225 09:10:26.948444 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:10:26 crc kubenswrapper[4978]: I0225 09:10:26.948532 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c8c5a88a-2673-4723-8967-9f209d416174-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.265283 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-networker-77j4w" event={"ID":"c8c5a88a-2673-4723-8967-9f209d416174","Type":"ContainerDied","Data":"9b49a72e31eb5ffb0560b63e761fab991486de23d68f81c0078431a19be1dfde"} Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.265344 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b49a72e31eb5ffb0560b63e761fab991486de23d68f81c0078431a19be1dfde" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.265417 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-networker-77j4w" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.381566 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-networker-4nb7l"] Feb 25 09:10:27 crc kubenswrapper[4978]: E0225 09:10:27.382032 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ae8605-e2a4-41dd-b5bc-5275301ed3be" containerName="oc" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.382051 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ae8605-e2a4-41dd-b5bc-5275301ed3be" containerName="oc" Feb 25 09:10:27 crc kubenswrapper[4978]: E0225 09:10:27.382087 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c5a88a-2673-4723-8967-9f209d416174" containerName="download-cache-openstack-openstack-networker" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.382095 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c5a88a-2673-4723-8967-9f209d416174" containerName="download-cache-openstack-openstack-networker" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.382291 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c5a88a-2673-4723-8967-9f209d416174" containerName="download-cache-openstack-openstack-networker" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.382307 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ae8605-e2a4-41dd-b5bc-5275301ed3be" containerName="oc" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.383025 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.385258 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.386153 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.397189 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-networker-4nb7l"] Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.466750 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-ssh-key-openstack-networker\") pod \"configure-network-openstack-openstack-networker-4nb7l\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.466952 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2k7w\" (UniqueName: \"kubernetes.io/projected/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-kube-api-access-t2k7w\") pod \"configure-network-openstack-openstack-networker-4nb7l\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.467247 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-inventory\") pod \"configure-network-openstack-openstack-networker-4nb7l\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.569027 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2k7w\" (UniqueName: \"kubernetes.io/projected/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-kube-api-access-t2k7w\") pod \"configure-network-openstack-openstack-networker-4nb7l\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.569195 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-inventory\") pod \"configure-network-openstack-openstack-networker-4nb7l\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.569333 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-ssh-key-openstack-networker\") pod \"configure-network-openstack-openstack-networker-4nb7l\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.574353 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-ssh-key-openstack-networker\") pod \"configure-network-openstack-openstack-networker-4nb7l\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.575567 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-inventory\") pod \"configure-network-openstack-openstack-networker-4nb7l\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.589314 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2k7w\" (UniqueName: \"kubernetes.io/projected/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-kube-api-access-t2k7w\") pod \"configure-network-openstack-openstack-networker-4nb7l\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:27 crc kubenswrapper[4978]: I0225 09:10:27.702362 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:10:28 crc kubenswrapper[4978]: W0225 09:10:28.329192 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04b8c344_5cce_4f9c_9e0b_181950cb4e8b.slice/crio-aa65ae1e271d4e48d16cb635eb0c9adb6358394e9b48a71c203fa2450c00a89c WatchSource:0}: Error finding container aa65ae1e271d4e48d16cb635eb0c9adb6358394e9b48a71c203fa2450c00a89c: Status 404 returned error can't find the container with id aa65ae1e271d4e48d16cb635eb0c9adb6358394e9b48a71c203fa2450c00a89c Feb 25 09:10:28 crc kubenswrapper[4978]: I0225 09:10:28.333199 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-networker-4nb7l"] Feb 25 09:10:29 crc kubenswrapper[4978]: I0225 09:10:29.289800 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-4nb7l" event={"ID":"04b8c344-5cce-4f9c-9e0b-181950cb4e8b","Type":"ContainerStarted","Data":"5214b72b5662f942f1ceeb77e18351532e5348d1eb69d7d0f18241527e86c4e7"} Feb 25 09:10:29 crc kubenswrapper[4978]: I0225 09:10:29.290080 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-4nb7l" event={"ID":"04b8c344-5cce-4f9c-9e0b-181950cb4e8b","Type":"ContainerStarted","Data":"aa65ae1e271d4e48d16cb635eb0c9adb6358394e9b48a71c203fa2450c00a89c"} Feb 25 09:10:29 crc kubenswrapper[4978]: I0225 09:10:29.320074 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-networker-4nb7l" podStartSLOduration=1.8202556410000001 podStartE2EDuration="2.32005371s" podCreationTimestamp="2026-02-25 09:10:27 +0000 UTC" firstStartedPulling="2026-02-25 09:10:28.33523761 +0000 UTC m=+8721.774494109" lastFinishedPulling="2026-02-25 09:10:28.835035709 +0000 UTC m=+8722.274292178" observedRunningTime="2026-02-25 09:10:29.317762169 +0000 UTC m=+8722.757018658" watchObservedRunningTime="2026-02-25 09:10:29.32005371 +0000 UTC m=+8722.759310169" Feb 25 09:10:29 crc kubenswrapper[4978]: I0225 09:10:29.327805 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:10:29 crc kubenswrapper[4978]: E0225 09:10:29.328154 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:10:35 crc kubenswrapper[4978]: I0225 09:10:35.544501 4978 scope.go:117] "RemoveContainer" containerID="0aea4f8aef7964c8da272a2a1195dfd6a729e4fc63301d1665c3e6642c209fe4" Feb 25 09:10:41 crc kubenswrapper[4978]: I0225 09:10:41.327650 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:10:41 crc kubenswrapper[4978]: E0225 09:10:41.328452 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:10:53 crc kubenswrapper[4978]: I0225 09:10:53.327839 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:10:53 crc kubenswrapper[4978]: E0225 09:10:53.328741 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:11:04 crc kubenswrapper[4978]: I0225 09:11:04.328580 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:11:04 crc kubenswrapper[4978]: E0225 09:11:04.329787 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:11:07 crc kubenswrapper[4978]: I0225 09:11:07.693142 4978 generic.go:334] "Generic (PLEG): container finished" podID="99ce1a38-0659-4c3d-be3f-5311d2a4cd8e" containerID="c4b438892b682bb0bf2a5c7723d2b1c6c7334352348c4d00227efd0389c729e5" exitCode=0 Feb 25 09:11:07 crc kubenswrapper[4978]: I0225 09:11:07.693253 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" event={"ID":"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e","Type":"ContainerDied","Data":"c4b438892b682bb0bf2a5c7723d2b1c6c7334352348c4d00227efd0389c729e5"} Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.297655 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.455589 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq9h8\" (UniqueName: \"kubernetes.io/projected/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-kube-api-access-sq9h8\") pod \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.455894 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-ssh-key-openstack-cell1\") pod \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.456120 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-inventory\") pod \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\" (UID: \"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e\") " Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.462935 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-kube-api-access-sq9h8" (OuterVolumeSpecName: "kube-api-access-sq9h8") pod "99ce1a38-0659-4c3d-be3f-5311d2a4cd8e" (UID: "99ce1a38-0659-4c3d-be3f-5311d2a4cd8e"). InnerVolumeSpecName "kube-api-access-sq9h8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.491646 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-inventory" (OuterVolumeSpecName: "inventory") pod "99ce1a38-0659-4c3d-be3f-5311d2a4cd8e" (UID: "99ce1a38-0659-4c3d-be3f-5311d2a4cd8e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.492236 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "99ce1a38-0659-4c3d-be3f-5311d2a4cd8e" (UID: "99ce1a38-0659-4c3d-be3f-5311d2a4cd8e"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.559006 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq9h8\" (UniqueName: \"kubernetes.io/projected/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-kube-api-access-sq9h8\") on node \"crc\" DevicePath \"\"" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.559058 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.559078 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99ce1a38-0659-4c3d-be3f-5311d2a4cd8e-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.715970 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" event={"ID":"99ce1a38-0659-4c3d-be3f-5311d2a4cd8e","Type":"ContainerDied","Data":"67ed1f6c475edb6fdb9b9bfcfab7e26d2937c910e4c3a3ec33b04376609ff2c5"} Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.716040 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67ed1f6c475edb6fdb9b9bfcfab7e26d2937c910e4c3a3ec33b04376609ff2c5" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.716433 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-hzw4k" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.817731 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-gc692"] Feb 25 09:11:09 crc kubenswrapper[4978]: E0225 09:11:09.818195 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ce1a38-0659-4c3d-be3f-5311d2a4cd8e" containerName="download-cache-openstack-openstack-cell1" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.818220 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ce1a38-0659-4c3d-be3f-5311d2a4cd8e" containerName="download-cache-openstack-openstack-cell1" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.818519 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="99ce1a38-0659-4c3d-be3f-5311d2a4cd8e" containerName="download-cache-openstack-openstack-cell1" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.819276 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.823026 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.823509 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.831872 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-gc692"] Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.967376 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-inventory\") pod \"configure-network-openstack-openstack-cell1-gc692\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.967496 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-gc692\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:09 crc kubenswrapper[4978]: I0225 09:11:09.967549 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbhmh\" (UniqueName: \"kubernetes.io/projected/03aef306-3186-4b8a-9a1e-523ebca2e8fc-kube-api-access-kbhmh\") pod \"configure-network-openstack-openstack-cell1-gc692\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:10 crc kubenswrapper[4978]: I0225 09:11:10.070165 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-gc692\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:10 crc kubenswrapper[4978]: I0225 09:11:10.070316 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbhmh\" (UniqueName: \"kubernetes.io/projected/03aef306-3186-4b8a-9a1e-523ebca2e8fc-kube-api-access-kbhmh\") pod \"configure-network-openstack-openstack-cell1-gc692\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:10 crc kubenswrapper[4978]: I0225 09:11:10.070514 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-inventory\") pod \"configure-network-openstack-openstack-cell1-gc692\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:10 crc kubenswrapper[4978]: I0225 09:11:10.076216 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-gc692\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:10 crc kubenswrapper[4978]: I0225 09:11:10.080158 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-inventory\") pod \"configure-network-openstack-openstack-cell1-gc692\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:10 crc kubenswrapper[4978]: I0225 09:11:10.095098 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbhmh\" (UniqueName: \"kubernetes.io/projected/03aef306-3186-4b8a-9a1e-523ebca2e8fc-kube-api-access-kbhmh\") pod \"configure-network-openstack-openstack-cell1-gc692\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:10 crc kubenswrapper[4978]: I0225 09:11:10.154313 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:11:10 crc kubenswrapper[4978]: W0225 09:11:10.750778 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03aef306_3186_4b8a_9a1e_523ebca2e8fc.slice/crio-47e32296ca32e4b858fea65f506e1cbfc05187414bb15d71e10f21eb610ad623 WatchSource:0}: Error finding container 47e32296ca32e4b858fea65f506e1cbfc05187414bb15d71e10f21eb610ad623: Status 404 returned error can't find the container with id 47e32296ca32e4b858fea65f506e1cbfc05187414bb15d71e10f21eb610ad623 Feb 25 09:11:10 crc kubenswrapper[4978]: I0225 09:11:10.752061 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-gc692"] Feb 25 09:11:11 crc kubenswrapper[4978]: I0225 09:11:11.735319 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-gc692" event={"ID":"03aef306-3186-4b8a-9a1e-523ebca2e8fc","Type":"ContainerStarted","Data":"e06689688035bd4659e9538531ebe3c19a074cb02ea7a0c5c13bb56669d22a69"} Feb 25 09:11:11 crc kubenswrapper[4978]: I0225 09:11:11.736011 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-gc692" event={"ID":"03aef306-3186-4b8a-9a1e-523ebca2e8fc","Type":"ContainerStarted","Data":"47e32296ca32e4b858fea65f506e1cbfc05187414bb15d71e10f21eb610ad623"} Feb 25 09:11:11 crc kubenswrapper[4978]: I0225 09:11:11.756812 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-gc692" podStartSLOduration=2.348857856 podStartE2EDuration="2.756795289s" podCreationTimestamp="2026-02-25 09:11:09 +0000 UTC" firstStartedPulling="2026-02-25 09:11:10.754279171 +0000 UTC m=+8764.193535630" lastFinishedPulling="2026-02-25 09:11:11.162216574 +0000 UTC m=+8764.601473063" observedRunningTime="2026-02-25 09:11:11.749622296 +0000 UTC m=+8765.188878755" watchObservedRunningTime="2026-02-25 09:11:11.756795289 +0000 UTC m=+8765.196051748" Feb 25 09:11:16 crc kubenswrapper[4978]: I0225 09:11:16.328204 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:11:16 crc kubenswrapper[4978]: E0225 09:11:16.329452 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:11:29 crc kubenswrapper[4978]: I0225 09:11:29.331328 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:11:29 crc kubenswrapper[4978]: I0225 09:11:29.966168 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"6e09bc70e0ad7207309daaa14091dbcb18dfd0d3b48bf2f83a1437dfb4535285"} Feb 25 09:11:34 crc kubenswrapper[4978]: I0225 09:11:34.012342 4978 generic.go:334] "Generic (PLEG): container finished" podID="04b8c344-5cce-4f9c-9e0b-181950cb4e8b" containerID="5214b72b5662f942f1ceeb77e18351532e5348d1eb69d7d0f18241527e86c4e7" exitCode=0 Feb 25 09:11:34 crc kubenswrapper[4978]: I0225 09:11:34.012476 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-4nb7l" event={"ID":"04b8c344-5cce-4f9c-9e0b-181950cb4e8b","Type":"ContainerDied","Data":"5214b72b5662f942f1ceeb77e18351532e5348d1eb69d7d0f18241527e86c4e7"} Feb 25 09:11:35 crc kubenswrapper[4978]: I0225 09:11:35.537121 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:11:35 crc kubenswrapper[4978]: I0225 09:11:35.611228 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-inventory\") pod \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " Feb 25 09:11:35 crc kubenswrapper[4978]: I0225 09:11:35.611641 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-ssh-key-openstack-networker\") pod \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " Feb 25 09:11:35 crc kubenswrapper[4978]: I0225 09:11:35.611876 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2k7w\" (UniqueName: \"kubernetes.io/projected/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-kube-api-access-t2k7w\") pod \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\" (UID: \"04b8c344-5cce-4f9c-9e0b-181950cb4e8b\") " Feb 25 09:11:35 crc kubenswrapper[4978]: I0225 09:11:35.632091 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-kube-api-access-t2k7w" (OuterVolumeSpecName: "kube-api-access-t2k7w") pod "04b8c344-5cce-4f9c-9e0b-181950cb4e8b" (UID: "04b8c344-5cce-4f9c-9e0b-181950cb4e8b"). InnerVolumeSpecName "kube-api-access-t2k7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:11:35 crc kubenswrapper[4978]: I0225 09:11:35.647692 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "04b8c344-5cce-4f9c-9e0b-181950cb4e8b" (UID: "04b8c344-5cce-4f9c-9e0b-181950cb4e8b"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:11:35 crc kubenswrapper[4978]: I0225 09:11:35.654688 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-inventory" (OuterVolumeSpecName: "inventory") pod "04b8c344-5cce-4f9c-9e0b-181950cb4e8b" (UID: "04b8c344-5cce-4f9c-9e0b-181950cb4e8b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:11:35 crc kubenswrapper[4978]: I0225 09:11:35.714907 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2k7w\" (UniqueName: \"kubernetes.io/projected/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-kube-api-access-t2k7w\") on node \"crc\" DevicePath \"\"" Feb 25 09:11:35 crc kubenswrapper[4978]: I0225 09:11:35.714949 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:11:35 crc kubenswrapper[4978]: I0225 09:11:35.714962 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/04b8c344-5cce-4f9c-9e0b-181950cb4e8b-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.048587 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-networker-4nb7l" event={"ID":"04b8c344-5cce-4f9c-9e0b-181950cb4e8b","Type":"ContainerDied","Data":"aa65ae1e271d4e48d16cb635eb0c9adb6358394e9b48a71c203fa2450c00a89c"} Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.048895 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa65ae1e271d4e48d16cb635eb0c9adb6358394e9b48a71c203fa2450c00a89c" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.048649 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-networker-4nb7l" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.131746 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-networker-jl46h"] Feb 25 09:11:36 crc kubenswrapper[4978]: E0225 09:11:36.132273 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b8c344-5cce-4f9c-9e0b-181950cb4e8b" containerName="configure-network-openstack-openstack-networker" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.132289 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b8c344-5cce-4f9c-9e0b-181950cb4e8b" containerName="configure-network-openstack-openstack-networker" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.132577 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b8c344-5cce-4f9c-9e0b-181950cb4e8b" containerName="configure-network-openstack-openstack-networker" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.135338 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.139001 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.139557 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.151663 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-networker-jl46h"] Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.227302 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsps2\" (UniqueName: \"kubernetes.io/projected/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-kube-api-access-tsps2\") pod \"validate-network-openstack-openstack-networker-jl46h\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.227498 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-ssh-key-openstack-networker\") pod \"validate-network-openstack-openstack-networker-jl46h\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.227548 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-inventory\") pod \"validate-network-openstack-openstack-networker-jl46h\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.330388 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-ssh-key-openstack-networker\") pod \"validate-network-openstack-openstack-networker-jl46h\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.330456 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-inventory\") pod \"validate-network-openstack-openstack-networker-jl46h\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.331402 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsps2\" (UniqueName: \"kubernetes.io/projected/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-kube-api-access-tsps2\") pod \"validate-network-openstack-openstack-networker-jl46h\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.339338 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-ssh-key-openstack-networker\") pod \"validate-network-openstack-openstack-networker-jl46h\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.339664 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-inventory\") pod \"validate-network-openstack-openstack-networker-jl46h\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.350789 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsps2\" (UniqueName: \"kubernetes.io/projected/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-kube-api-access-tsps2\") pod \"validate-network-openstack-openstack-networker-jl46h\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:36 crc kubenswrapper[4978]: I0225 09:11:36.468765 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:37 crc kubenswrapper[4978]: I0225 09:11:37.032156 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-networker-jl46h"] Feb 25 09:11:37 crc kubenswrapper[4978]: I0225 09:11:37.064512 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-jl46h" event={"ID":"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71","Type":"ContainerStarted","Data":"ce3c10c18d6d09a61b53d7d41d22862a538e661807cbc86fb82e2f095f0d35d6"} Feb 25 09:11:38 crc kubenswrapper[4978]: I0225 09:11:38.084975 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-jl46h" event={"ID":"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71","Type":"ContainerStarted","Data":"d6bea8265d13fa6b5fedc251b21c0e3c8e75236b86f249a342300b925855b351"} Feb 25 09:11:38 crc kubenswrapper[4978]: I0225 09:11:38.115610 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-networker-jl46h" podStartSLOduration=1.722290168 podStartE2EDuration="2.115583889s" podCreationTimestamp="2026-02-25 09:11:36 +0000 UTC" firstStartedPulling="2026-02-25 09:11:37.043773066 +0000 UTC m=+8790.483029525" lastFinishedPulling="2026-02-25 09:11:37.437066787 +0000 UTC m=+8790.876323246" observedRunningTime="2026-02-25 09:11:38.114629941 +0000 UTC m=+8791.553886440" watchObservedRunningTime="2026-02-25 09:11:38.115583889 +0000 UTC m=+8791.554840358" Feb 25 09:11:43 crc kubenswrapper[4978]: I0225 09:11:43.178284 4978 generic.go:334] "Generic (PLEG): container finished" podID="4ccacd71-8214-46a5-b6a4-d31ec3b9fa71" containerID="d6bea8265d13fa6b5fedc251b21c0e3c8e75236b86f249a342300b925855b351" exitCode=0 Feb 25 09:11:43 crc kubenswrapper[4978]: I0225 09:11:43.178397 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-jl46h" event={"ID":"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71","Type":"ContainerDied","Data":"d6bea8265d13fa6b5fedc251b21c0e3c8e75236b86f249a342300b925855b351"} Feb 25 09:11:44 crc kubenswrapper[4978]: I0225 09:11:44.733450 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:44 crc kubenswrapper[4978]: I0225 09:11:44.833539 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-ssh-key-openstack-networker\") pod \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " Feb 25 09:11:44 crc kubenswrapper[4978]: I0225 09:11:44.833706 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsps2\" (UniqueName: \"kubernetes.io/projected/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-kube-api-access-tsps2\") pod \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " Feb 25 09:11:44 crc kubenswrapper[4978]: I0225 09:11:44.833823 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-inventory\") pod \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\" (UID: \"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71\") " Feb 25 09:11:44 crc kubenswrapper[4978]: I0225 09:11:44.839664 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-kube-api-access-tsps2" (OuterVolumeSpecName: "kube-api-access-tsps2") pod "4ccacd71-8214-46a5-b6a4-d31ec3b9fa71" (UID: "4ccacd71-8214-46a5-b6a4-d31ec3b9fa71"). InnerVolumeSpecName "kube-api-access-tsps2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:11:44 crc kubenswrapper[4978]: I0225 09:11:44.871622 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-inventory" (OuterVolumeSpecName: "inventory") pod "4ccacd71-8214-46a5-b6a4-d31ec3b9fa71" (UID: "4ccacd71-8214-46a5-b6a4-d31ec3b9fa71"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:11:44 crc kubenswrapper[4978]: I0225 09:11:44.896819 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "4ccacd71-8214-46a5-b6a4-d31ec3b9fa71" (UID: "4ccacd71-8214-46a5-b6a4-d31ec3b9fa71"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:11:44 crc kubenswrapper[4978]: I0225 09:11:44.936994 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:11:44 crc kubenswrapper[4978]: I0225 09:11:44.937038 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:11:44 crc kubenswrapper[4978]: I0225 09:11:44.937060 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsps2\" (UniqueName: \"kubernetes.io/projected/4ccacd71-8214-46a5-b6a4-d31ec3b9fa71-kube-api-access-tsps2\") on node \"crc\" DevicePath \"\"" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.205158 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-networker-jl46h" event={"ID":"4ccacd71-8214-46a5-b6a4-d31ec3b9fa71","Type":"ContainerDied","Data":"ce3c10c18d6d09a61b53d7d41d22862a538e661807cbc86fb82e2f095f0d35d6"} Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.205635 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce3c10c18d6d09a61b53d7d41d22862a538e661807cbc86fb82e2f095f0d35d6" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.205276 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-networker-jl46h" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.350571 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-networker-fwglm"] Feb 25 09:11:45 crc kubenswrapper[4978]: E0225 09:11:45.350982 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ccacd71-8214-46a5-b6a4-d31ec3b9fa71" containerName="validate-network-openstack-openstack-networker" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.351003 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ccacd71-8214-46a5-b6a4-d31ec3b9fa71" containerName="validate-network-openstack-openstack-networker" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.351259 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ccacd71-8214-46a5-b6a4-d31ec3b9fa71" containerName="validate-network-openstack-openstack-networker" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.352138 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-networker-fwglm"] Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.352241 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.361805 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.362768 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.448821 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-inventory\") pod \"install-os-openstack-openstack-networker-fwglm\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.448870 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw9hs\" (UniqueName: \"kubernetes.io/projected/50114eec-786f-467f-b235-24829d7c3e63-kube-api-access-rw9hs\") pod \"install-os-openstack-openstack-networker-fwglm\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.448991 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-ssh-key-openstack-networker\") pod \"install-os-openstack-openstack-networker-fwglm\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.550152 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-ssh-key-openstack-networker\") pod \"install-os-openstack-openstack-networker-fwglm\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.550289 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-inventory\") pod \"install-os-openstack-openstack-networker-fwglm\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.550314 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw9hs\" (UniqueName: \"kubernetes.io/projected/50114eec-786f-467f-b235-24829d7c3e63-kube-api-access-rw9hs\") pod \"install-os-openstack-openstack-networker-fwglm\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.555489 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-inventory\") pod \"install-os-openstack-openstack-networker-fwglm\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.557056 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-ssh-key-openstack-networker\") pod \"install-os-openstack-openstack-networker-fwglm\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.567220 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw9hs\" (UniqueName: \"kubernetes.io/projected/50114eec-786f-467f-b235-24829d7c3e63-kube-api-access-rw9hs\") pod \"install-os-openstack-openstack-networker-fwglm\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:45 crc kubenswrapper[4978]: I0225 09:11:45.674768 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:11:46 crc kubenswrapper[4978]: I0225 09:11:46.306398 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-networker-fwglm"] Feb 25 09:11:46 crc kubenswrapper[4978]: I0225 09:11:46.320342 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:11:47 crc kubenswrapper[4978]: I0225 09:11:47.233722 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-fwglm" event={"ID":"50114eec-786f-467f-b235-24829d7c3e63","Type":"ContainerStarted","Data":"ff83f1bff8ecf1dbb7cba733d671cecf66129c05923d7d2c5fd39aedb49f52a6"} Feb 25 09:11:47 crc kubenswrapper[4978]: I0225 09:11:47.234349 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-fwglm" event={"ID":"50114eec-786f-467f-b235-24829d7c3e63","Type":"ContainerStarted","Data":"09a4a62d04a06adec20adcb466e6897403075a0aeb8fc5e606e1fb0d5c89a009"} Feb 25 09:11:47 crc kubenswrapper[4978]: I0225 09:11:47.264620 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-networker-fwglm" podStartSLOduration=1.660846974 podStartE2EDuration="2.264601852s" podCreationTimestamp="2026-02-25 09:11:45 +0000 UTC" firstStartedPulling="2026-02-25 09:11:46.319854074 +0000 UTC m=+8799.759110543" lastFinishedPulling="2026-02-25 09:11:46.923608922 +0000 UTC m=+8800.362865421" observedRunningTime="2026-02-25 09:11:47.259271607 +0000 UTC m=+8800.698528076" watchObservedRunningTime="2026-02-25 09:11:47.264601852 +0000 UTC m=+8800.703858311" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.444166 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9ghzj"] Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.446545 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.454325 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9ghzj"] Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.492374 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-utilities\") pod \"community-operators-9ghzj\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.492436 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-catalog-content\") pod \"community-operators-9ghzj\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.492512 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpfsh\" (UniqueName: \"kubernetes.io/projected/6637433e-0d18-4385-941b-2bbe0df9e670-kube-api-access-rpfsh\") pod \"community-operators-9ghzj\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.594851 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-utilities\") pod \"community-operators-9ghzj\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.594918 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-catalog-content\") pod \"community-operators-9ghzj\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.594949 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpfsh\" (UniqueName: \"kubernetes.io/projected/6637433e-0d18-4385-941b-2bbe0df9e670-kube-api-access-rpfsh\") pod \"community-operators-9ghzj\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.595382 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-utilities\") pod \"community-operators-9ghzj\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.595736 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-catalog-content\") pod \"community-operators-9ghzj\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.616545 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpfsh\" (UniqueName: \"kubernetes.io/projected/6637433e-0d18-4385-941b-2bbe0df9e670-kube-api-access-rpfsh\") pod \"community-operators-9ghzj\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:11:59 crc kubenswrapper[4978]: I0225 09:11:59.764023 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.140785 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533512-8j5tw"] Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.142233 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533512-8j5tw" Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.145387 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.145611 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.145911 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.154736 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533512-8j5tw"] Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.208785 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mvkh\" (UniqueName: \"kubernetes.io/projected/5fc8fde2-869b-4730-85c5-e1d6250e5edc-kube-api-access-7mvkh\") pod \"auto-csr-approver-29533512-8j5tw\" (UID: \"5fc8fde2-869b-4730-85c5-e1d6250e5edc\") " pod="openshift-infra/auto-csr-approver-29533512-8j5tw" Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.301246 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9ghzj"] Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.314218 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mvkh\" (UniqueName: \"kubernetes.io/projected/5fc8fde2-869b-4730-85c5-e1d6250e5edc-kube-api-access-7mvkh\") pod \"auto-csr-approver-29533512-8j5tw\" (UID: \"5fc8fde2-869b-4730-85c5-e1d6250e5edc\") " pod="openshift-infra/auto-csr-approver-29533512-8j5tw" Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.339120 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mvkh\" (UniqueName: \"kubernetes.io/projected/5fc8fde2-869b-4730-85c5-e1d6250e5edc-kube-api-access-7mvkh\") pod \"auto-csr-approver-29533512-8j5tw\" (UID: \"5fc8fde2-869b-4730-85c5-e1d6250e5edc\") " pod="openshift-infra/auto-csr-approver-29533512-8j5tw" Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.369492 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ghzj" event={"ID":"6637433e-0d18-4385-941b-2bbe0df9e670","Type":"ContainerStarted","Data":"2527b41f61e2378a9059c5f3086c5d6faa434aa6607021b6045661e158efe273"} Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.463878 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533512-8j5tw" Feb 25 09:12:00 crc kubenswrapper[4978]: I0225 09:12:00.925804 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533512-8j5tw"] Feb 25 09:12:01 crc kubenswrapper[4978]: I0225 09:12:01.384676 4978 generic.go:334] "Generic (PLEG): container finished" podID="6637433e-0d18-4385-941b-2bbe0df9e670" containerID="1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57" exitCode=0 Feb 25 09:12:01 crc kubenswrapper[4978]: I0225 09:12:01.384764 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ghzj" event={"ID":"6637433e-0d18-4385-941b-2bbe0df9e670","Type":"ContainerDied","Data":"1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57"} Feb 25 09:12:01 crc kubenswrapper[4978]: I0225 09:12:01.388075 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533512-8j5tw" event={"ID":"5fc8fde2-869b-4730-85c5-e1d6250e5edc","Type":"ContainerStarted","Data":"1e6124885a01552ede85529f1c55f2d404d3f51e1fff83de02246954f3d653e3"} Feb 25 09:12:02 crc kubenswrapper[4978]: I0225 09:12:02.400265 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533512-8j5tw" event={"ID":"5fc8fde2-869b-4730-85c5-e1d6250e5edc","Type":"ContainerStarted","Data":"6702ec1416d897a3573139b7470967f7a191ec9ca2551c45103df6f8ef6ddfca"} Feb 25 09:12:02 crc kubenswrapper[4978]: I0225 09:12:02.405036 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ghzj" event={"ID":"6637433e-0d18-4385-941b-2bbe0df9e670","Type":"ContainerStarted","Data":"7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1"} Feb 25 09:12:02 crc kubenswrapper[4978]: I0225 09:12:02.422309 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533512-8j5tw" podStartSLOduration=1.497317644 podStartE2EDuration="2.42228646s" podCreationTimestamp="2026-02-25 09:12:00 +0000 UTC" firstStartedPulling="2026-02-25 09:12:00.925580777 +0000 UTC m=+8814.364837246" lastFinishedPulling="2026-02-25 09:12:01.850549593 +0000 UTC m=+8815.289806062" observedRunningTime="2026-02-25 09:12:02.418902965 +0000 UTC m=+8815.858159434" watchObservedRunningTime="2026-02-25 09:12:02.42228646 +0000 UTC m=+8815.861542949" Feb 25 09:12:03 crc kubenswrapper[4978]: I0225 09:12:03.417900 4978 generic.go:334] "Generic (PLEG): container finished" podID="5fc8fde2-869b-4730-85c5-e1d6250e5edc" containerID="6702ec1416d897a3573139b7470967f7a191ec9ca2551c45103df6f8ef6ddfca" exitCode=0 Feb 25 09:12:03 crc kubenswrapper[4978]: I0225 09:12:03.418010 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533512-8j5tw" event={"ID":"5fc8fde2-869b-4730-85c5-e1d6250e5edc","Type":"ContainerDied","Data":"6702ec1416d897a3573139b7470967f7a191ec9ca2551c45103df6f8ef6ddfca"} Feb 25 09:12:04 crc kubenswrapper[4978]: I0225 09:12:04.437002 4978 generic.go:334] "Generic (PLEG): container finished" podID="6637433e-0d18-4385-941b-2bbe0df9e670" containerID="7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1" exitCode=0 Feb 25 09:12:04 crc kubenswrapper[4978]: I0225 09:12:04.437208 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ghzj" event={"ID":"6637433e-0d18-4385-941b-2bbe0df9e670","Type":"ContainerDied","Data":"7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1"} Feb 25 09:12:04 crc kubenswrapper[4978]: I0225 09:12:04.883962 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533512-8j5tw" Feb 25 09:12:04 crc kubenswrapper[4978]: I0225 09:12:04.928084 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mvkh\" (UniqueName: \"kubernetes.io/projected/5fc8fde2-869b-4730-85c5-e1d6250e5edc-kube-api-access-7mvkh\") pod \"5fc8fde2-869b-4730-85c5-e1d6250e5edc\" (UID: \"5fc8fde2-869b-4730-85c5-e1d6250e5edc\") " Feb 25 09:12:04 crc kubenswrapper[4978]: I0225 09:12:04.934791 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fc8fde2-869b-4730-85c5-e1d6250e5edc-kube-api-access-7mvkh" (OuterVolumeSpecName: "kube-api-access-7mvkh") pod "5fc8fde2-869b-4730-85c5-e1d6250e5edc" (UID: "5fc8fde2-869b-4730-85c5-e1d6250e5edc"). InnerVolumeSpecName "kube-api-access-7mvkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:12:05 crc kubenswrapper[4978]: I0225 09:12:05.030668 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mvkh\" (UniqueName: \"kubernetes.io/projected/5fc8fde2-869b-4730-85c5-e1d6250e5edc-kube-api-access-7mvkh\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:05 crc kubenswrapper[4978]: E0225 09:12:05.418397 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fc8fde2_869b_4730_85c5_e1d6250e5edc.slice/crio-1e6124885a01552ede85529f1c55f2d404d3f51e1fff83de02246954f3d653e3\": RecentStats: unable to find data in memory cache]" Feb 25 09:12:05 crc kubenswrapper[4978]: I0225 09:12:05.471627 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ghzj" event={"ID":"6637433e-0d18-4385-941b-2bbe0df9e670","Type":"ContainerStarted","Data":"3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9"} Feb 25 09:12:05 crc kubenswrapper[4978]: I0225 09:12:05.478317 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533512-8j5tw" event={"ID":"5fc8fde2-869b-4730-85c5-e1d6250e5edc","Type":"ContainerDied","Data":"1e6124885a01552ede85529f1c55f2d404d3f51e1fff83de02246954f3d653e3"} Feb 25 09:12:05 crc kubenswrapper[4978]: I0225 09:12:05.478631 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e6124885a01552ede85529f1c55f2d404d3f51e1fff83de02246954f3d653e3" Feb 25 09:12:05 crc kubenswrapper[4978]: I0225 09:12:05.478685 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533512-8j5tw" Feb 25 09:12:05 crc kubenswrapper[4978]: I0225 09:12:05.490799 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9ghzj" podStartSLOduration=2.9246576490000002 podStartE2EDuration="6.4907834s" podCreationTimestamp="2026-02-25 09:11:59 +0000 UTC" firstStartedPulling="2026-02-25 09:12:01.387423161 +0000 UTC m=+8814.826679620" lastFinishedPulling="2026-02-25 09:12:04.953548912 +0000 UTC m=+8818.392805371" observedRunningTime="2026-02-25 09:12:05.485953851 +0000 UTC m=+8818.925210310" watchObservedRunningTime="2026-02-25 09:12:05.4907834 +0000 UTC m=+8818.930039859" Feb 25 09:12:05 crc kubenswrapper[4978]: I0225 09:12:05.505500 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533506-kjrl4"] Feb 25 09:12:05 crc kubenswrapper[4978]: I0225 09:12:05.513000 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533506-kjrl4"] Feb 25 09:12:07 crc kubenswrapper[4978]: I0225 09:12:07.344711 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02bb15c4-7c69-4b27-bd36-bc13eb17cdce" path="/var/lib/kubelet/pods/02bb15c4-7c69-4b27-bd36-bc13eb17cdce/volumes" Feb 25 09:12:09 crc kubenswrapper[4978]: I0225 09:12:09.764569 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:12:09 crc kubenswrapper[4978]: I0225 09:12:09.764948 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:12:09 crc kubenswrapper[4978]: I0225 09:12:09.833201 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:12:10 crc kubenswrapper[4978]: I0225 09:12:10.594903 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:12:10 crc kubenswrapper[4978]: I0225 09:12:10.657081 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9ghzj"] Feb 25 09:12:12 crc kubenswrapper[4978]: I0225 09:12:12.554065 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9ghzj" podUID="6637433e-0d18-4385-941b-2bbe0df9e670" containerName="registry-server" containerID="cri-o://3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9" gracePeriod=2 Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.077556 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.166736 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-utilities\") pod \"6637433e-0d18-4385-941b-2bbe0df9e670\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.167391 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-catalog-content\") pod \"6637433e-0d18-4385-941b-2bbe0df9e670\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.167661 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpfsh\" (UniqueName: \"kubernetes.io/projected/6637433e-0d18-4385-941b-2bbe0df9e670-kube-api-access-rpfsh\") pod \"6637433e-0d18-4385-941b-2bbe0df9e670\" (UID: \"6637433e-0d18-4385-941b-2bbe0df9e670\") " Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.168143 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-utilities" (OuterVolumeSpecName: "utilities") pod "6637433e-0d18-4385-941b-2bbe0df9e670" (UID: "6637433e-0d18-4385-941b-2bbe0df9e670"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.168403 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.176918 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6637433e-0d18-4385-941b-2bbe0df9e670-kube-api-access-rpfsh" (OuterVolumeSpecName: "kube-api-access-rpfsh") pod "6637433e-0d18-4385-941b-2bbe0df9e670" (UID: "6637433e-0d18-4385-941b-2bbe0df9e670"). InnerVolumeSpecName "kube-api-access-rpfsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.240811 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6637433e-0d18-4385-941b-2bbe0df9e670" (UID: "6637433e-0d18-4385-941b-2bbe0df9e670"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.270955 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6637433e-0d18-4385-941b-2bbe0df9e670-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.271001 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpfsh\" (UniqueName: \"kubernetes.io/projected/6637433e-0d18-4385-941b-2bbe0df9e670-kube-api-access-rpfsh\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.566417 4978 generic.go:334] "Generic (PLEG): container finished" podID="03aef306-3186-4b8a-9a1e-523ebca2e8fc" containerID="e06689688035bd4659e9538531ebe3c19a074cb02ea7a0c5c13bb56669d22a69" exitCode=0 Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.566540 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-gc692" event={"ID":"03aef306-3186-4b8a-9a1e-523ebca2e8fc","Type":"ContainerDied","Data":"e06689688035bd4659e9538531ebe3c19a074cb02ea7a0c5c13bb56669d22a69"} Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.571826 4978 generic.go:334] "Generic (PLEG): container finished" podID="6637433e-0d18-4385-941b-2bbe0df9e670" containerID="3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9" exitCode=0 Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.571877 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ghzj" event={"ID":"6637433e-0d18-4385-941b-2bbe0df9e670","Type":"ContainerDied","Data":"3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9"} Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.571907 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9ghzj" event={"ID":"6637433e-0d18-4385-941b-2bbe0df9e670","Type":"ContainerDied","Data":"2527b41f61e2378a9059c5f3086c5d6faa434aa6607021b6045661e158efe273"} Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.571914 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9ghzj" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.571928 4978 scope.go:117] "RemoveContainer" containerID="3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.610726 4978 scope.go:117] "RemoveContainer" containerID="7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.625976 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9ghzj"] Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.643273 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9ghzj"] Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.647278 4978 scope.go:117] "RemoveContainer" containerID="1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.718824 4978 scope.go:117] "RemoveContainer" containerID="3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9" Feb 25 09:12:13 crc kubenswrapper[4978]: E0225 09:12:13.719580 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9\": container with ID starting with 3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9 not found: ID does not exist" containerID="3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.719631 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9"} err="failed to get container status \"3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9\": rpc error: code = NotFound desc = could not find container \"3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9\": container with ID starting with 3b62eac0ebd05c128d6ca6b202951f128e6793bfb9ac7954453d7fb5bbd59ed9 not found: ID does not exist" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.719665 4978 scope.go:117] "RemoveContainer" containerID="7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1" Feb 25 09:12:13 crc kubenswrapper[4978]: E0225 09:12:13.720171 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1\": container with ID starting with 7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1 not found: ID does not exist" containerID="7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.720230 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1"} err="failed to get container status \"7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1\": rpc error: code = NotFound desc = could not find container \"7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1\": container with ID starting with 7dfd996187de6e1dfb33bdffc30a7dd8777b8512f0b6683ec072e42348bddfd1 not found: ID does not exist" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.720261 4978 scope.go:117] "RemoveContainer" containerID="1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57" Feb 25 09:12:13 crc kubenswrapper[4978]: E0225 09:12:13.720619 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57\": container with ID starting with 1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57 not found: ID does not exist" containerID="1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57" Feb 25 09:12:13 crc kubenswrapper[4978]: I0225 09:12:13.720640 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57"} err="failed to get container status \"1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57\": rpc error: code = NotFound desc = could not find container \"1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57\": container with ID starting with 1340c6929b0a63afaf00c3d3296ab3e976b3757b4ce7f51da51db4cffa361a57 not found: ID does not exist" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.187003 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.326611 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-ssh-key-openstack-cell1\") pod \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.327540 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-inventory\") pod \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.327585 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbhmh\" (UniqueName: \"kubernetes.io/projected/03aef306-3186-4b8a-9a1e-523ebca2e8fc-kube-api-access-kbhmh\") pod \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\" (UID: \"03aef306-3186-4b8a-9a1e-523ebca2e8fc\") " Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.337676 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03aef306-3186-4b8a-9a1e-523ebca2e8fc-kube-api-access-kbhmh" (OuterVolumeSpecName: "kube-api-access-kbhmh") pod "03aef306-3186-4b8a-9a1e-523ebca2e8fc" (UID: "03aef306-3186-4b8a-9a1e-523ebca2e8fc"). InnerVolumeSpecName "kube-api-access-kbhmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.342883 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6637433e-0d18-4385-941b-2bbe0df9e670" path="/var/lib/kubelet/pods/6637433e-0d18-4385-941b-2bbe0df9e670/volumes" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.364489 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-inventory" (OuterVolumeSpecName: "inventory") pod "03aef306-3186-4b8a-9a1e-523ebca2e8fc" (UID: "03aef306-3186-4b8a-9a1e-523ebca2e8fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.375188 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "03aef306-3186-4b8a-9a1e-523ebca2e8fc" (UID: "03aef306-3186-4b8a-9a1e-523ebca2e8fc"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.431047 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.431090 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbhmh\" (UniqueName: \"kubernetes.io/projected/03aef306-3186-4b8a-9a1e-523ebca2e8fc-kube-api-access-kbhmh\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.431105 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/03aef306-3186-4b8a-9a1e-523ebca2e8fc-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.949584 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-gc692" event={"ID":"03aef306-3186-4b8a-9a1e-523ebca2e8fc","Type":"ContainerDied","Data":"47e32296ca32e4b858fea65f506e1cbfc05187414bb15d71e10f21eb610ad623"} Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.950187 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47e32296ca32e4b858fea65f506e1cbfc05187414bb15d71e10f21eb610ad623" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.949654 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-gc692" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.953915 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-czcmg"] Feb 25 09:12:15 crc kubenswrapper[4978]: E0225 09:12:15.954480 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6637433e-0d18-4385-941b-2bbe0df9e670" containerName="registry-server" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.954501 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6637433e-0d18-4385-941b-2bbe0df9e670" containerName="registry-server" Feb 25 09:12:15 crc kubenswrapper[4978]: E0225 09:12:15.954533 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc8fde2-869b-4730-85c5-e1d6250e5edc" containerName="oc" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.954541 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc8fde2-869b-4730-85c5-e1d6250e5edc" containerName="oc" Feb 25 09:12:15 crc kubenswrapper[4978]: E0225 09:12:15.954557 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6637433e-0d18-4385-941b-2bbe0df9e670" containerName="extract-content" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.954565 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6637433e-0d18-4385-941b-2bbe0df9e670" containerName="extract-content" Feb 25 09:12:15 crc kubenswrapper[4978]: E0225 09:12:15.954578 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03aef306-3186-4b8a-9a1e-523ebca2e8fc" containerName="configure-network-openstack-openstack-cell1" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.954587 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="03aef306-3186-4b8a-9a1e-523ebca2e8fc" containerName="configure-network-openstack-openstack-cell1" Feb 25 09:12:15 crc kubenswrapper[4978]: E0225 09:12:15.954619 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6637433e-0d18-4385-941b-2bbe0df9e670" containerName="extract-utilities" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.954628 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6637433e-0d18-4385-941b-2bbe0df9e670" containerName="extract-utilities" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.954883 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc8fde2-869b-4730-85c5-e1d6250e5edc" containerName="oc" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.954902 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="03aef306-3186-4b8a-9a1e-523ebca2e8fc" containerName="configure-network-openstack-openstack-cell1" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.954915 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6637433e-0d18-4385-941b-2bbe0df9e670" containerName="registry-server" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.956777 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.961675 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.961907 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:12:15 crc kubenswrapper[4978]: I0225 09:12:15.962621 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-czcmg"] Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.116703 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-inventory\") pod \"validate-network-openstack-openstack-cell1-czcmg\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.117086 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbgdp\" (UniqueName: \"kubernetes.io/projected/ad13325e-8670-43ca-9461-aec47779e55e-kube-api-access-mbgdp\") pod \"validate-network-openstack-openstack-cell1-czcmg\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.117161 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-czcmg\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.218804 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbgdp\" (UniqueName: \"kubernetes.io/projected/ad13325e-8670-43ca-9461-aec47779e55e-kube-api-access-mbgdp\") pod \"validate-network-openstack-openstack-cell1-czcmg\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.218941 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-czcmg\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.219226 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-inventory\") pod \"validate-network-openstack-openstack-cell1-czcmg\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.225085 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-inventory\") pod \"validate-network-openstack-openstack-cell1-czcmg\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.231205 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-czcmg\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.241169 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbgdp\" (UniqueName: \"kubernetes.io/projected/ad13325e-8670-43ca-9461-aec47779e55e-kube-api-access-mbgdp\") pod \"validate-network-openstack-openstack-cell1-czcmg\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.275919 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:16 crc kubenswrapper[4978]: W0225 09:12:16.896799 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad13325e_8670_43ca_9461_aec47779e55e.slice/crio-926ec2d23e264d402fdcf56ad12a8c833d70fb91550a98df0906d54a564225ed WatchSource:0}: Error finding container 926ec2d23e264d402fdcf56ad12a8c833d70fb91550a98df0906d54a564225ed: Status 404 returned error can't find the container with id 926ec2d23e264d402fdcf56ad12a8c833d70fb91550a98df0906d54a564225ed Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.897414 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-czcmg"] Feb 25 09:12:16 crc kubenswrapper[4978]: I0225 09:12:16.966013 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-czcmg" event={"ID":"ad13325e-8670-43ca-9461-aec47779e55e","Type":"ContainerStarted","Data":"926ec2d23e264d402fdcf56ad12a8c833d70fb91550a98df0906d54a564225ed"} Feb 25 09:12:17 crc kubenswrapper[4978]: I0225 09:12:17.977706 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-czcmg" event={"ID":"ad13325e-8670-43ca-9461-aec47779e55e","Type":"ContainerStarted","Data":"432377501aabf68358ce888bc832cee2ce2c61fd45062119266ceb6c3bceeda2"} Feb 25 09:12:18 crc kubenswrapper[4978]: I0225 09:12:18.012753 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-czcmg" podStartSLOduration=2.475008047 podStartE2EDuration="3.01272685s" podCreationTimestamp="2026-02-25 09:12:15 +0000 UTC" firstStartedPulling="2026-02-25 09:12:16.900974389 +0000 UTC m=+8830.340230888" lastFinishedPulling="2026-02-25 09:12:17.438693202 +0000 UTC m=+8830.877949691" observedRunningTime="2026-02-25 09:12:17.997657284 +0000 UTC m=+8831.436913763" watchObservedRunningTime="2026-02-25 09:12:18.01272685 +0000 UTC m=+8831.451983349" Feb 25 09:12:24 crc kubenswrapper[4978]: I0225 09:12:24.050443 4978 generic.go:334] "Generic (PLEG): container finished" podID="ad13325e-8670-43ca-9461-aec47779e55e" containerID="432377501aabf68358ce888bc832cee2ce2c61fd45062119266ceb6c3bceeda2" exitCode=0 Feb 25 09:12:24 crc kubenswrapper[4978]: I0225 09:12:24.050569 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-czcmg" event={"ID":"ad13325e-8670-43ca-9461-aec47779e55e","Type":"ContainerDied","Data":"432377501aabf68358ce888bc832cee2ce2c61fd45062119266ceb6c3bceeda2"} Feb 25 09:12:25 crc kubenswrapper[4978]: I0225 09:12:25.517543 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:25 crc kubenswrapper[4978]: I0225 09:12:25.671075 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbgdp\" (UniqueName: \"kubernetes.io/projected/ad13325e-8670-43ca-9461-aec47779e55e-kube-api-access-mbgdp\") pod \"ad13325e-8670-43ca-9461-aec47779e55e\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " Feb 25 09:12:25 crc kubenswrapper[4978]: I0225 09:12:25.671254 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-ssh-key-openstack-cell1\") pod \"ad13325e-8670-43ca-9461-aec47779e55e\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " Feb 25 09:12:25 crc kubenswrapper[4978]: I0225 09:12:25.671464 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-inventory\") pod \"ad13325e-8670-43ca-9461-aec47779e55e\" (UID: \"ad13325e-8670-43ca-9461-aec47779e55e\") " Feb 25 09:12:25 crc kubenswrapper[4978]: I0225 09:12:25.679875 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad13325e-8670-43ca-9461-aec47779e55e-kube-api-access-mbgdp" (OuterVolumeSpecName: "kube-api-access-mbgdp") pod "ad13325e-8670-43ca-9461-aec47779e55e" (UID: "ad13325e-8670-43ca-9461-aec47779e55e"). InnerVolumeSpecName "kube-api-access-mbgdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:12:25 crc kubenswrapper[4978]: I0225 09:12:25.721720 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "ad13325e-8670-43ca-9461-aec47779e55e" (UID: "ad13325e-8670-43ca-9461-aec47779e55e"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:12:25 crc kubenswrapper[4978]: I0225 09:12:25.722966 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-inventory" (OuterVolumeSpecName: "inventory") pod "ad13325e-8670-43ca-9461-aec47779e55e" (UID: "ad13325e-8670-43ca-9461-aec47779e55e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:12:25 crc kubenswrapper[4978]: I0225 09:12:25.775492 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:25 crc kubenswrapper[4978]: I0225 09:12:25.775542 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ad13325e-8670-43ca-9461-aec47779e55e-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:25 crc kubenswrapper[4978]: I0225 09:12:25.775563 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbgdp\" (UniqueName: \"kubernetes.io/projected/ad13325e-8670-43ca-9461-aec47779e55e-kube-api-access-mbgdp\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.077612 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-czcmg" event={"ID":"ad13325e-8670-43ca-9461-aec47779e55e","Type":"ContainerDied","Data":"926ec2d23e264d402fdcf56ad12a8c833d70fb91550a98df0906d54a564225ed"} Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.077659 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="926ec2d23e264d402fdcf56ad12a8c833d70fb91550a98df0906d54a564225ed" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.077726 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-czcmg" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.227810 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-97vfl"] Feb 25 09:12:26 crc kubenswrapper[4978]: E0225 09:12:26.228327 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad13325e-8670-43ca-9461-aec47779e55e" containerName="validate-network-openstack-openstack-cell1" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.228352 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad13325e-8670-43ca-9461-aec47779e55e" containerName="validate-network-openstack-openstack-cell1" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.228636 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad13325e-8670-43ca-9461-aec47779e55e" containerName="validate-network-openstack-openstack-cell1" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.229492 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.233833 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.233998 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.243537 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-97vfl"] Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.291983 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-inventory\") pod \"install-os-openstack-openstack-cell1-97vfl\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.292045 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-97vfl\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.292106 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95mrn\" (UniqueName: \"kubernetes.io/projected/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-kube-api-access-95mrn\") pod \"install-os-openstack-openstack-cell1-97vfl\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:26 crc kubenswrapper[4978]: E0225 09:12:26.362315 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad13325e_8670_43ca_9461_aec47779e55e.slice\": RecentStats: unable to find data in memory cache]" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.394202 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-inventory\") pod \"install-os-openstack-openstack-cell1-97vfl\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.394626 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-97vfl\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.394713 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95mrn\" (UniqueName: \"kubernetes.io/projected/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-kube-api-access-95mrn\") pod \"install-os-openstack-openstack-cell1-97vfl\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.400058 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-inventory\") pod \"install-os-openstack-openstack-cell1-97vfl\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.400109 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-97vfl\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.409755 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95mrn\" (UniqueName: \"kubernetes.io/projected/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-kube-api-access-95mrn\") pod \"install-os-openstack-openstack-cell1-97vfl\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:26 crc kubenswrapper[4978]: I0225 09:12:26.555976 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:12:27 crc kubenswrapper[4978]: I0225 09:12:27.157734 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-97vfl"] Feb 25 09:12:28 crc kubenswrapper[4978]: I0225 09:12:28.107006 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-97vfl" event={"ID":"0aaa93ca-3269-4ade-bb1d-ac862e9738c5","Type":"ContainerStarted","Data":"08782adac0dcfcc94f490dd32c241bcde6d9513c5b0bd5a84340a7ed5743fb42"} Feb 25 09:12:28 crc kubenswrapper[4978]: I0225 09:12:28.107405 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-97vfl" event={"ID":"0aaa93ca-3269-4ade-bb1d-ac862e9738c5","Type":"ContainerStarted","Data":"a0242979fa3684840fce73a8614735db3ebf0b1766701e8c666c6d22ce37bda5"} Feb 25 09:12:28 crc kubenswrapper[4978]: I0225 09:12:28.134266 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-97vfl" podStartSLOduration=1.740813594 podStartE2EDuration="2.134240848s" podCreationTimestamp="2026-02-25 09:12:26 +0000 UTC" firstStartedPulling="2026-02-25 09:12:27.166509428 +0000 UTC m=+8840.605765897" lastFinishedPulling="2026-02-25 09:12:27.559936652 +0000 UTC m=+8840.999193151" observedRunningTime="2026-02-25 09:12:28.124265809 +0000 UTC m=+8841.563522298" watchObservedRunningTime="2026-02-25 09:12:28.134240848 +0000 UTC m=+8841.573497307" Feb 25 09:12:35 crc kubenswrapper[4978]: I0225 09:12:35.659681 4978 scope.go:117] "RemoveContainer" containerID="307a04b86c55772b23c949535a67b84987c7219eada766a80ac3faadd1ca6ef5" Feb 25 09:12:36 crc kubenswrapper[4978]: I0225 09:12:36.201654 4978 generic.go:334] "Generic (PLEG): container finished" podID="50114eec-786f-467f-b235-24829d7c3e63" containerID="ff83f1bff8ecf1dbb7cba733d671cecf66129c05923d7d2c5fd39aedb49f52a6" exitCode=0 Feb 25 09:12:36 crc kubenswrapper[4978]: I0225 09:12:36.201750 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-fwglm" event={"ID":"50114eec-786f-467f-b235-24829d7c3e63","Type":"ContainerDied","Data":"ff83f1bff8ecf1dbb7cba733d671cecf66129c05923d7d2c5fd39aedb49f52a6"} Feb 25 09:12:37 crc kubenswrapper[4978]: I0225 09:12:37.680425 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:12:37 crc kubenswrapper[4978]: I0225 09:12:37.759324 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-inventory\") pod \"50114eec-786f-467f-b235-24829d7c3e63\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " Feb 25 09:12:37 crc kubenswrapper[4978]: I0225 09:12:37.759533 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-ssh-key-openstack-networker\") pod \"50114eec-786f-467f-b235-24829d7c3e63\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " Feb 25 09:12:37 crc kubenswrapper[4978]: I0225 09:12:37.759674 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw9hs\" (UniqueName: \"kubernetes.io/projected/50114eec-786f-467f-b235-24829d7c3e63-kube-api-access-rw9hs\") pod \"50114eec-786f-467f-b235-24829d7c3e63\" (UID: \"50114eec-786f-467f-b235-24829d7c3e63\") " Feb 25 09:12:37 crc kubenswrapper[4978]: I0225 09:12:37.768826 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50114eec-786f-467f-b235-24829d7c3e63-kube-api-access-rw9hs" (OuterVolumeSpecName: "kube-api-access-rw9hs") pod "50114eec-786f-467f-b235-24829d7c3e63" (UID: "50114eec-786f-467f-b235-24829d7c3e63"). InnerVolumeSpecName "kube-api-access-rw9hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:12:37 crc kubenswrapper[4978]: I0225 09:12:37.796822 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-inventory" (OuterVolumeSpecName: "inventory") pod "50114eec-786f-467f-b235-24829d7c3e63" (UID: "50114eec-786f-467f-b235-24829d7c3e63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:12:37 crc kubenswrapper[4978]: I0225 09:12:37.798221 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "50114eec-786f-467f-b235-24829d7c3e63" (UID: "50114eec-786f-467f-b235-24829d7c3e63"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:12:37 crc kubenswrapper[4978]: I0225 09:12:37.862879 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:37 crc kubenswrapper[4978]: I0225 09:12:37.862918 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/50114eec-786f-467f-b235-24829d7c3e63-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:37 crc kubenswrapper[4978]: I0225 09:12:37.862935 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw9hs\" (UniqueName: \"kubernetes.io/projected/50114eec-786f-467f-b235-24829d7c3e63-kube-api-access-rw9hs\") on node \"crc\" DevicePath \"\"" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.227098 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-networker-fwglm" event={"ID":"50114eec-786f-467f-b235-24829d7c3e63","Type":"ContainerDied","Data":"09a4a62d04a06adec20adcb466e6897403075a0aeb8fc5e606e1fb0d5c89a009"} Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.227157 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09a4a62d04a06adec20adcb466e6897403075a0aeb8fc5e606e1fb0d5c89a009" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.227666 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-networker-fwglm" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.333193 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-networker-lgws7"] Feb 25 09:12:38 crc kubenswrapper[4978]: E0225 09:12:38.333890 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50114eec-786f-467f-b235-24829d7c3e63" containerName="install-os-openstack-openstack-networker" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.333995 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="50114eec-786f-467f-b235-24829d7c3e63" containerName="install-os-openstack-openstack-networker" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.334399 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="50114eec-786f-467f-b235-24829d7c3e63" containerName="install-os-openstack-openstack-networker" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.335330 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.337940 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.337941 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.345645 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-lgws7"] Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.475177 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-ssh-key-openstack-networker\") pod \"configure-os-openstack-openstack-networker-lgws7\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.475499 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-inventory\") pod \"configure-os-openstack-openstack-networker-lgws7\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.475720 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlcmz\" (UniqueName: \"kubernetes.io/projected/ec14586b-e00a-4d26-9afc-87ef605ddfc4-kube-api-access-nlcmz\") pod \"configure-os-openstack-openstack-networker-lgws7\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.577365 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-ssh-key-openstack-networker\") pod \"configure-os-openstack-openstack-networker-lgws7\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.577506 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-inventory\") pod \"configure-os-openstack-openstack-networker-lgws7\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.577621 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlcmz\" (UniqueName: \"kubernetes.io/projected/ec14586b-e00a-4d26-9afc-87ef605ddfc4-kube-api-access-nlcmz\") pod \"configure-os-openstack-openstack-networker-lgws7\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.584010 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-ssh-key-openstack-networker\") pod \"configure-os-openstack-openstack-networker-lgws7\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.589133 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-inventory\") pod \"configure-os-openstack-openstack-networker-lgws7\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.598247 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlcmz\" (UniqueName: \"kubernetes.io/projected/ec14586b-e00a-4d26-9afc-87ef605ddfc4-kube-api-access-nlcmz\") pod \"configure-os-openstack-openstack-networker-lgws7\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:38 crc kubenswrapper[4978]: I0225 09:12:38.663243 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:12:39 crc kubenswrapper[4978]: I0225 09:12:39.195571 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-networker-lgws7"] Feb 25 09:12:39 crc kubenswrapper[4978]: W0225 09:12:39.198320 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec14586b_e00a_4d26_9afc_87ef605ddfc4.slice/crio-bf8aae067ce15fd55d9a9560dd73e9917f3e04c41e08af3a090dc23ccdf8dc7f WatchSource:0}: Error finding container bf8aae067ce15fd55d9a9560dd73e9917f3e04c41e08af3a090dc23ccdf8dc7f: Status 404 returned error can't find the container with id bf8aae067ce15fd55d9a9560dd73e9917f3e04c41e08af3a090dc23ccdf8dc7f Feb 25 09:12:39 crc kubenswrapper[4978]: I0225 09:12:39.238180 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-lgws7" event={"ID":"ec14586b-e00a-4d26-9afc-87ef605ddfc4","Type":"ContainerStarted","Data":"bf8aae067ce15fd55d9a9560dd73e9917f3e04c41e08af3a090dc23ccdf8dc7f"} Feb 25 09:12:40 crc kubenswrapper[4978]: I0225 09:12:40.247637 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-lgws7" event={"ID":"ec14586b-e00a-4d26-9afc-87ef605ddfc4","Type":"ContainerStarted","Data":"f4f19cd6ad2ae3a2e620908132f2c11c3aaec47c5f84f9ea5419a7b51378b5ca"} Feb 25 09:12:40 crc kubenswrapper[4978]: I0225 09:12:40.267433 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-networker-lgws7" podStartSLOduration=1.506414812 podStartE2EDuration="2.267412219s" podCreationTimestamp="2026-02-25 09:12:38 +0000 UTC" firstStartedPulling="2026-02-25 09:12:39.200566269 +0000 UTC m=+8852.639822728" lastFinishedPulling="2026-02-25 09:12:39.961563666 +0000 UTC m=+8853.400820135" observedRunningTime="2026-02-25 09:12:40.261432183 +0000 UTC m=+8853.700688652" watchObservedRunningTime="2026-02-25 09:12:40.267412219 +0000 UTC m=+8853.706668698" Feb 25 09:13:19 crc kubenswrapper[4978]: I0225 09:13:19.673014 4978 generic.go:334] "Generic (PLEG): container finished" podID="0aaa93ca-3269-4ade-bb1d-ac862e9738c5" containerID="08782adac0dcfcc94f490dd32c241bcde6d9513c5b0bd5a84340a7ed5743fb42" exitCode=0 Feb 25 09:13:19 crc kubenswrapper[4978]: I0225 09:13:19.673094 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-97vfl" event={"ID":"0aaa93ca-3269-4ade-bb1d-ac862e9738c5","Type":"ContainerDied","Data":"08782adac0dcfcc94f490dd32c241bcde6d9513c5b0bd5a84340a7ed5743fb42"} Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.130729 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.204594 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-inventory\") pod \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.204668 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95mrn\" (UniqueName: \"kubernetes.io/projected/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-kube-api-access-95mrn\") pod \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.205062 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-ssh-key-openstack-cell1\") pod \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\" (UID: \"0aaa93ca-3269-4ade-bb1d-ac862e9738c5\") " Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.214680 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-kube-api-access-95mrn" (OuterVolumeSpecName: "kube-api-access-95mrn") pod "0aaa93ca-3269-4ade-bb1d-ac862e9738c5" (UID: "0aaa93ca-3269-4ade-bb1d-ac862e9738c5"). InnerVolumeSpecName "kube-api-access-95mrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.237610 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "0aaa93ca-3269-4ade-bb1d-ac862e9738c5" (UID: "0aaa93ca-3269-4ade-bb1d-ac862e9738c5"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.240816 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-inventory" (OuterVolumeSpecName: "inventory") pod "0aaa93ca-3269-4ade-bb1d-ac862e9738c5" (UID: "0aaa93ca-3269-4ade-bb1d-ac862e9738c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.308585 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.308626 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95mrn\" (UniqueName: \"kubernetes.io/projected/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-kube-api-access-95mrn\") on node \"crc\" DevicePath \"\"" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.308640 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/0aaa93ca-3269-4ade-bb1d-ac862e9738c5-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.696312 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-97vfl" event={"ID":"0aaa93ca-3269-4ade-bb1d-ac862e9738c5","Type":"ContainerDied","Data":"a0242979fa3684840fce73a8614735db3ebf0b1766701e8c666c6d22ce37bda5"} Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.696386 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-97vfl" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.696421 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0242979fa3684840fce73a8614735db3ebf0b1766701e8c666c6d22ce37bda5" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.824635 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-xbb8f"] Feb 25 09:13:21 crc kubenswrapper[4978]: E0225 09:13:21.825106 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aaa93ca-3269-4ade-bb1d-ac862e9738c5" containerName="install-os-openstack-openstack-cell1" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.825125 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aaa93ca-3269-4ade-bb1d-ac862e9738c5" containerName="install-os-openstack-openstack-cell1" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.825317 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aaa93ca-3269-4ade-bb1d-ac862e9738c5" containerName="install-os-openstack-openstack-cell1" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.826183 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.830977 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.831716 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.849538 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-xbb8f"] Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.918788 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjd85\" (UniqueName: \"kubernetes.io/projected/07f0da72-81c7-4452-8662-bc63584cad91-kube-api-access-tjd85\") pod \"configure-os-openstack-openstack-cell1-xbb8f\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.919101 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-inventory\") pod \"configure-os-openstack-openstack-cell1-xbb8f\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:21 crc kubenswrapper[4978]: I0225 09:13:21.919258 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-xbb8f\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:22 crc kubenswrapper[4978]: I0225 09:13:22.020121 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-inventory\") pod \"configure-os-openstack-openstack-cell1-xbb8f\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:22 crc kubenswrapper[4978]: I0225 09:13:22.020193 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-xbb8f\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:22 crc kubenswrapper[4978]: I0225 09:13:22.020274 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjd85\" (UniqueName: \"kubernetes.io/projected/07f0da72-81c7-4452-8662-bc63584cad91-kube-api-access-tjd85\") pod \"configure-os-openstack-openstack-cell1-xbb8f\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:22 crc kubenswrapper[4978]: I0225 09:13:22.024244 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-xbb8f\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:22 crc kubenswrapper[4978]: I0225 09:13:22.024427 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-inventory\") pod \"configure-os-openstack-openstack-cell1-xbb8f\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:22 crc kubenswrapper[4978]: I0225 09:13:22.043814 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjd85\" (UniqueName: \"kubernetes.io/projected/07f0da72-81c7-4452-8662-bc63584cad91-kube-api-access-tjd85\") pod \"configure-os-openstack-openstack-cell1-xbb8f\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:22 crc kubenswrapper[4978]: I0225 09:13:22.159334 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:13:22 crc kubenswrapper[4978]: I0225 09:13:22.769605 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-xbb8f"] Feb 25 09:13:23 crc kubenswrapper[4978]: I0225 09:13:23.722909 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" event={"ID":"07f0da72-81c7-4452-8662-bc63584cad91","Type":"ContainerStarted","Data":"bd1bf89277289572a19976a84d3b738d4ec01f9ac57b48ddf2d347378c849d5b"} Feb 25 09:13:23 crc kubenswrapper[4978]: I0225 09:13:23.725173 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" event={"ID":"07f0da72-81c7-4452-8662-bc63584cad91","Type":"ContainerStarted","Data":"9a5d23555ac6e160c18cc18892b4184552ed19e4061ced11c0ffe1bbfa6e0b56"} Feb 25 09:13:23 crc kubenswrapper[4978]: I0225 09:13:23.749925 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" podStartSLOduration=2.295749401 podStartE2EDuration="2.749893276s" podCreationTimestamp="2026-02-25 09:13:21 +0000 UTC" firstStartedPulling="2026-02-25 09:13:22.773564049 +0000 UTC m=+8896.212820508" lastFinishedPulling="2026-02-25 09:13:23.227707924 +0000 UTC m=+8896.666964383" observedRunningTime="2026-02-25 09:13:23.744830349 +0000 UTC m=+8897.184086818" watchObservedRunningTime="2026-02-25 09:13:23.749893276 +0000 UTC m=+8897.189149745" Feb 25 09:13:33 crc kubenswrapper[4978]: I0225 09:13:33.866202 4978 generic.go:334] "Generic (PLEG): container finished" podID="ec14586b-e00a-4d26-9afc-87ef605ddfc4" containerID="f4f19cd6ad2ae3a2e620908132f2c11c3aaec47c5f84f9ea5419a7b51378b5ca" exitCode=0 Feb 25 09:13:33 crc kubenswrapper[4978]: I0225 09:13:33.866783 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-lgws7" event={"ID":"ec14586b-e00a-4d26-9afc-87ef605ddfc4","Type":"ContainerDied","Data":"f4f19cd6ad2ae3a2e620908132f2c11c3aaec47c5f84f9ea5419a7b51378b5ca"} Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.402809 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.544541 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-ssh-key-openstack-networker\") pod \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.544645 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-inventory\") pod \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.544811 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlcmz\" (UniqueName: \"kubernetes.io/projected/ec14586b-e00a-4d26-9afc-87ef605ddfc4-kube-api-access-nlcmz\") pod \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\" (UID: \"ec14586b-e00a-4d26-9afc-87ef605ddfc4\") " Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.550438 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec14586b-e00a-4d26-9afc-87ef605ddfc4-kube-api-access-nlcmz" (OuterVolumeSpecName: "kube-api-access-nlcmz") pod "ec14586b-e00a-4d26-9afc-87ef605ddfc4" (UID: "ec14586b-e00a-4d26-9afc-87ef605ddfc4"). InnerVolumeSpecName "kube-api-access-nlcmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.574743 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "ec14586b-e00a-4d26-9afc-87ef605ddfc4" (UID: "ec14586b-e00a-4d26-9afc-87ef605ddfc4"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.574795 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-inventory" (OuterVolumeSpecName: "inventory") pod "ec14586b-e00a-4d26-9afc-87ef605ddfc4" (UID: "ec14586b-e00a-4d26-9afc-87ef605ddfc4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.647645 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.647701 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec14586b-e00a-4d26-9afc-87ef605ddfc4-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.647724 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlcmz\" (UniqueName: \"kubernetes.io/projected/ec14586b-e00a-4d26-9afc-87ef605ddfc4-kube-api-access-nlcmz\") on node \"crc\" DevicePath \"\"" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.900241 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-networker-lgws7" event={"ID":"ec14586b-e00a-4d26-9afc-87ef605ddfc4","Type":"ContainerDied","Data":"bf8aae067ce15fd55d9a9560dd73e9917f3e04c41e08af3a090dc23ccdf8dc7f"} Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.900897 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf8aae067ce15fd55d9a9560dd73e9917f3e04c41e08af3a090dc23ccdf8dc7f" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.900592 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-networker-lgws7" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.986435 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-networker-2mwbf"] Feb 25 09:13:35 crc kubenswrapper[4978]: E0225 09:13:35.987139 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec14586b-e00a-4d26-9afc-87ef605ddfc4" containerName="configure-os-openstack-openstack-networker" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.987171 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec14586b-e00a-4d26-9afc-87ef605ddfc4" containerName="configure-os-openstack-openstack-networker" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.987567 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec14586b-e00a-4d26-9afc-87ef605ddfc4" containerName="configure-os-openstack-openstack-networker" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.988651 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.992134 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.992328 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 09:13:35 crc kubenswrapper[4978]: I0225 09:13:35.997292 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-networker-2mwbf"] Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.054995 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-ssh-key-openstack-networker\") pod \"run-os-openstack-openstack-networker-2mwbf\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.055055 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4q4f\" (UniqueName: \"kubernetes.io/projected/ce547a55-9fd6-4692-b7e8-5e6303cd4293-kube-api-access-v4q4f\") pod \"run-os-openstack-openstack-networker-2mwbf\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.055110 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-inventory\") pod \"run-os-openstack-openstack-networker-2mwbf\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.156508 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-ssh-key-openstack-networker\") pod \"run-os-openstack-openstack-networker-2mwbf\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.156575 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4q4f\" (UniqueName: \"kubernetes.io/projected/ce547a55-9fd6-4692-b7e8-5e6303cd4293-kube-api-access-v4q4f\") pod \"run-os-openstack-openstack-networker-2mwbf\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.156627 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-inventory\") pod \"run-os-openstack-openstack-networker-2mwbf\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.162329 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-inventory\") pod \"run-os-openstack-openstack-networker-2mwbf\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.163245 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-ssh-key-openstack-networker\") pod \"run-os-openstack-openstack-networker-2mwbf\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.178706 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4q4f\" (UniqueName: \"kubernetes.io/projected/ce547a55-9fd6-4692-b7e8-5e6303cd4293-kube-api-access-v4q4f\") pod \"run-os-openstack-openstack-networker-2mwbf\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.315844 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:36 crc kubenswrapper[4978]: I0225 09:13:36.981840 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-networker-2mwbf"] Feb 25 09:13:37 crc kubenswrapper[4978]: I0225 09:13:37.924275 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-2mwbf" event={"ID":"ce547a55-9fd6-4692-b7e8-5e6303cd4293","Type":"ContainerStarted","Data":"a173bd595d8a73b46fe6347a77d8652d027e95117720913e048bb9335cef3d5a"} Feb 25 09:13:37 crc kubenswrapper[4978]: I0225 09:13:37.924802 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-2mwbf" event={"ID":"ce547a55-9fd6-4692-b7e8-5e6303cd4293","Type":"ContainerStarted","Data":"e3601f595cb0f03380b07ac8a4d8f0db7225a41cd2b6a0c6c3c259bcd6f0101b"} Feb 25 09:13:37 crc kubenswrapper[4978]: I0225 09:13:37.964649 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-networker-2mwbf" podStartSLOduration=2.5602115359999997 podStartE2EDuration="2.964629872s" podCreationTimestamp="2026-02-25 09:13:35 +0000 UTC" firstStartedPulling="2026-02-25 09:13:36.994038972 +0000 UTC m=+8910.433295471" lastFinishedPulling="2026-02-25 09:13:37.398457338 +0000 UTC m=+8910.837713807" observedRunningTime="2026-02-25 09:13:37.949166113 +0000 UTC m=+8911.388422592" watchObservedRunningTime="2026-02-25 09:13:37.964629872 +0000 UTC m=+8911.403886331" Feb 25 09:13:46 crc kubenswrapper[4978]: I0225 09:13:46.540034 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:13:46 crc kubenswrapper[4978]: I0225 09:13:46.540611 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:13:48 crc kubenswrapper[4978]: E0225 09:13:48.637453 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce547a55_9fd6_4692_b7e8_5e6303cd4293.slice/crio-a173bd595d8a73b46fe6347a77d8652d027e95117720913e048bb9335cef3d5a.scope\": RecentStats: unable to find data in memory cache]" Feb 25 09:13:49 crc kubenswrapper[4978]: I0225 09:13:49.046448 4978 generic.go:334] "Generic (PLEG): container finished" podID="ce547a55-9fd6-4692-b7e8-5e6303cd4293" containerID="a173bd595d8a73b46fe6347a77d8652d027e95117720913e048bb9335cef3d5a" exitCode=0 Feb 25 09:13:49 crc kubenswrapper[4978]: I0225 09:13:49.046544 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-2mwbf" event={"ID":"ce547a55-9fd6-4692-b7e8-5e6303cd4293","Type":"ContainerDied","Data":"a173bd595d8a73b46fe6347a77d8652d027e95117720913e048bb9335cef3d5a"} Feb 25 09:13:50 crc kubenswrapper[4978]: I0225 09:13:50.618443 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:50 crc kubenswrapper[4978]: I0225 09:13:50.707043 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4q4f\" (UniqueName: \"kubernetes.io/projected/ce547a55-9fd6-4692-b7e8-5e6303cd4293-kube-api-access-v4q4f\") pod \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " Feb 25 09:13:50 crc kubenswrapper[4978]: I0225 09:13:50.707403 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-inventory\") pod \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " Feb 25 09:13:50 crc kubenswrapper[4978]: I0225 09:13:50.707554 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-ssh-key-openstack-networker\") pod \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\" (UID: \"ce547a55-9fd6-4692-b7e8-5e6303cd4293\") " Feb 25 09:13:50 crc kubenswrapper[4978]: I0225 09:13:50.712086 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce547a55-9fd6-4692-b7e8-5e6303cd4293-kube-api-access-v4q4f" (OuterVolumeSpecName: "kube-api-access-v4q4f") pod "ce547a55-9fd6-4692-b7e8-5e6303cd4293" (UID: "ce547a55-9fd6-4692-b7e8-5e6303cd4293"). InnerVolumeSpecName "kube-api-access-v4q4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:13:50 crc kubenswrapper[4978]: I0225 09:13:50.734974 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-inventory" (OuterVolumeSpecName: "inventory") pod "ce547a55-9fd6-4692-b7e8-5e6303cd4293" (UID: "ce547a55-9fd6-4692-b7e8-5e6303cd4293"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:13:50 crc kubenswrapper[4978]: I0225 09:13:50.739736 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "ce547a55-9fd6-4692-b7e8-5e6303cd4293" (UID: "ce547a55-9fd6-4692-b7e8-5e6303cd4293"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:13:50 crc kubenswrapper[4978]: I0225 09:13:50.810901 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:13:50 crc kubenswrapper[4978]: I0225 09:13:50.810960 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4q4f\" (UniqueName: \"kubernetes.io/projected/ce547a55-9fd6-4692-b7e8-5e6303cd4293-kube-api-access-v4q4f\") on node \"crc\" DevicePath \"\"" Feb 25 09:13:50 crc kubenswrapper[4978]: I0225 09:13:50.810981 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce547a55-9fd6-4692-b7e8-5e6303cd4293-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.077293 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-networker-2mwbf" event={"ID":"ce547a55-9fd6-4692-b7e8-5e6303cd4293","Type":"ContainerDied","Data":"e3601f595cb0f03380b07ac8a4d8f0db7225a41cd2b6a0c6c3c259bcd6f0101b"} Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.077430 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3601f595cb0f03380b07ac8a4d8f0db7225a41cd2b6a0c6c3c259bcd6f0101b" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.077553 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-networker-2mwbf" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.187870 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-ddfpk"] Feb 25 09:13:51 crc kubenswrapper[4978]: E0225 09:13:51.188338 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce547a55-9fd6-4692-b7e8-5e6303cd4293" containerName="run-os-openstack-openstack-networker" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.188360 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce547a55-9fd6-4692-b7e8-5e6303cd4293" containerName="run-os-openstack-openstack-networker" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.188621 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce547a55-9fd6-4692-b7e8-5e6303cd4293" containerName="run-os-openstack-openstack-networker" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.189431 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.194075 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.194497 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.213552 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-ddfpk"] Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.329220 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-ssh-key-openstack-networker\") pod \"reboot-os-openstack-openstack-networker-ddfpk\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.329275 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxpf6\" (UniqueName: \"kubernetes.io/projected/87856568-5e2c-40f3-99cd-a159a8ae96e9-kube-api-access-sxpf6\") pod \"reboot-os-openstack-openstack-networker-ddfpk\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.329325 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-inventory\") pod \"reboot-os-openstack-openstack-networker-ddfpk\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.431294 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-ssh-key-openstack-networker\") pod \"reboot-os-openstack-openstack-networker-ddfpk\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.431394 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxpf6\" (UniqueName: \"kubernetes.io/projected/87856568-5e2c-40f3-99cd-a159a8ae96e9-kube-api-access-sxpf6\") pod \"reboot-os-openstack-openstack-networker-ddfpk\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.431496 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-inventory\") pod \"reboot-os-openstack-openstack-networker-ddfpk\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.435803 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-inventory\") pod \"reboot-os-openstack-openstack-networker-ddfpk\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.436563 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-ssh-key-openstack-networker\") pod \"reboot-os-openstack-openstack-networker-ddfpk\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.453517 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxpf6\" (UniqueName: \"kubernetes.io/projected/87856568-5e2c-40f3-99cd-a159a8ae96e9-kube-api-access-sxpf6\") pod \"reboot-os-openstack-openstack-networker-ddfpk\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:51 crc kubenswrapper[4978]: I0225 09:13:51.560157 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:13:52 crc kubenswrapper[4978]: I0225 09:13:52.122473 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-networker-ddfpk"] Feb 25 09:13:52 crc kubenswrapper[4978]: W0225 09:13:52.129854 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87856568_5e2c_40f3_99cd_a159a8ae96e9.slice/crio-4579dad824acf2df67f2bb6421fdf5eb5a66ad300cba123fd83ded1fcfd6d646 WatchSource:0}: Error finding container 4579dad824acf2df67f2bb6421fdf5eb5a66ad300cba123fd83ded1fcfd6d646: Status 404 returned error can't find the container with id 4579dad824acf2df67f2bb6421fdf5eb5a66ad300cba123fd83ded1fcfd6d646 Feb 25 09:13:53 crc kubenswrapper[4978]: I0225 09:13:53.104299 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" event={"ID":"87856568-5e2c-40f3-99cd-a159a8ae96e9","Type":"ContainerStarted","Data":"d6538069a351992b3ac515d6a7b7974f63fbf6f89648a6b961001917884413d8"} Feb 25 09:13:53 crc kubenswrapper[4978]: I0225 09:13:53.104842 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" event={"ID":"87856568-5e2c-40f3-99cd-a159a8ae96e9","Type":"ContainerStarted","Data":"4579dad824acf2df67f2bb6421fdf5eb5a66ad300cba123fd83ded1fcfd6d646"} Feb 25 09:13:53 crc kubenswrapper[4978]: I0225 09:13:53.140595 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" podStartSLOduration=1.650763285 podStartE2EDuration="2.140576465s" podCreationTimestamp="2026-02-25 09:13:51 +0000 UTC" firstStartedPulling="2026-02-25 09:13:52.136358534 +0000 UTC m=+8925.575615023" lastFinishedPulling="2026-02-25 09:13:52.626171704 +0000 UTC m=+8926.065428203" observedRunningTime="2026-02-25 09:13:53.136437867 +0000 UTC m=+8926.575694416" watchObservedRunningTime="2026-02-25 09:13:53.140576465 +0000 UTC m=+8926.579832934" Feb 25 09:14:00 crc kubenswrapper[4978]: I0225 09:14:00.183138 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533514-w7wkw"] Feb 25 09:14:00 crc kubenswrapper[4978]: I0225 09:14:00.185462 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533514-w7wkw" Feb 25 09:14:00 crc kubenswrapper[4978]: I0225 09:14:00.187442 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:14:00 crc kubenswrapper[4978]: I0225 09:14:00.188136 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:14:00 crc kubenswrapper[4978]: I0225 09:14:00.188411 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:14:00 crc kubenswrapper[4978]: I0225 09:14:00.198984 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533514-w7wkw"] Feb 25 09:14:00 crc kubenswrapper[4978]: I0225 09:14:00.247177 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zpv7\" (UniqueName: \"kubernetes.io/projected/0ae641dc-1ef2-464e-b1da-6d7341626b34-kube-api-access-4zpv7\") pod \"auto-csr-approver-29533514-w7wkw\" (UID: \"0ae641dc-1ef2-464e-b1da-6d7341626b34\") " pod="openshift-infra/auto-csr-approver-29533514-w7wkw" Feb 25 09:14:00 crc kubenswrapper[4978]: I0225 09:14:00.349275 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zpv7\" (UniqueName: \"kubernetes.io/projected/0ae641dc-1ef2-464e-b1da-6d7341626b34-kube-api-access-4zpv7\") pod \"auto-csr-approver-29533514-w7wkw\" (UID: \"0ae641dc-1ef2-464e-b1da-6d7341626b34\") " pod="openshift-infra/auto-csr-approver-29533514-w7wkw" Feb 25 09:14:00 crc kubenswrapper[4978]: I0225 09:14:00.369576 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zpv7\" (UniqueName: \"kubernetes.io/projected/0ae641dc-1ef2-464e-b1da-6d7341626b34-kube-api-access-4zpv7\") pod \"auto-csr-approver-29533514-w7wkw\" (UID: \"0ae641dc-1ef2-464e-b1da-6d7341626b34\") " pod="openshift-infra/auto-csr-approver-29533514-w7wkw" Feb 25 09:14:00 crc kubenswrapper[4978]: I0225 09:14:00.523570 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533514-w7wkw" Feb 25 09:14:01 crc kubenswrapper[4978]: I0225 09:14:01.027657 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533514-w7wkw"] Feb 25 09:14:01 crc kubenswrapper[4978]: I0225 09:14:01.200222 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533514-w7wkw" event={"ID":"0ae641dc-1ef2-464e-b1da-6d7341626b34","Type":"ContainerStarted","Data":"062cf3a7ebff8a94c5a74e3b8958ecc98c0278f792189209320ba2b63311cecc"} Feb 25 09:14:02 crc kubenswrapper[4978]: I0225 09:14:02.214791 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533514-w7wkw" event={"ID":"0ae641dc-1ef2-464e-b1da-6d7341626b34","Type":"ContainerStarted","Data":"d4fdd675d37c0f972b6bd25dd903d0c351d3a314487fc067a3ba615e54ac475e"} Feb 25 09:14:02 crc kubenswrapper[4978]: I0225 09:14:02.231236 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533514-w7wkw" podStartSLOduration=1.341111532 podStartE2EDuration="2.231216668s" podCreationTimestamp="2026-02-25 09:14:00 +0000 UTC" firstStartedPulling="2026-02-25 09:14:01.057468538 +0000 UTC m=+8934.496724997" lastFinishedPulling="2026-02-25 09:14:01.947573674 +0000 UTC m=+8935.386830133" observedRunningTime="2026-02-25 09:14:02.228885625 +0000 UTC m=+8935.668142104" watchObservedRunningTime="2026-02-25 09:14:02.231216668 +0000 UTC m=+8935.670473137" Feb 25 09:14:03 crc kubenswrapper[4978]: I0225 09:14:03.229353 4978 generic.go:334] "Generic (PLEG): container finished" podID="0ae641dc-1ef2-464e-b1da-6d7341626b34" containerID="d4fdd675d37c0f972b6bd25dd903d0c351d3a314487fc067a3ba615e54ac475e" exitCode=0 Feb 25 09:14:03 crc kubenswrapper[4978]: I0225 09:14:03.230007 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533514-w7wkw" event={"ID":"0ae641dc-1ef2-464e-b1da-6d7341626b34","Type":"ContainerDied","Data":"d4fdd675d37c0f972b6bd25dd903d0c351d3a314487fc067a3ba615e54ac475e"} Feb 25 09:14:04 crc kubenswrapper[4978]: I0225 09:14:04.667837 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533514-w7wkw" Feb 25 09:14:04 crc kubenswrapper[4978]: I0225 09:14:04.751312 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zpv7\" (UniqueName: \"kubernetes.io/projected/0ae641dc-1ef2-464e-b1da-6d7341626b34-kube-api-access-4zpv7\") pod \"0ae641dc-1ef2-464e-b1da-6d7341626b34\" (UID: \"0ae641dc-1ef2-464e-b1da-6d7341626b34\") " Feb 25 09:14:04 crc kubenswrapper[4978]: I0225 09:14:04.761584 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ae641dc-1ef2-464e-b1da-6d7341626b34-kube-api-access-4zpv7" (OuterVolumeSpecName: "kube-api-access-4zpv7") pod "0ae641dc-1ef2-464e-b1da-6d7341626b34" (UID: "0ae641dc-1ef2-464e-b1da-6d7341626b34"). InnerVolumeSpecName "kube-api-access-4zpv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:14:04 crc kubenswrapper[4978]: I0225 09:14:04.854310 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zpv7\" (UniqueName: \"kubernetes.io/projected/0ae641dc-1ef2-464e-b1da-6d7341626b34-kube-api-access-4zpv7\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:05 crc kubenswrapper[4978]: I0225 09:14:05.259543 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533514-w7wkw" event={"ID":"0ae641dc-1ef2-464e-b1da-6d7341626b34","Type":"ContainerDied","Data":"062cf3a7ebff8a94c5a74e3b8958ecc98c0278f792189209320ba2b63311cecc"} Feb 25 09:14:05 crc kubenswrapper[4978]: I0225 09:14:05.259889 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="062cf3a7ebff8a94c5a74e3b8958ecc98c0278f792189209320ba2b63311cecc" Feb 25 09:14:05 crc kubenswrapper[4978]: I0225 09:14:05.259704 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533514-w7wkw" Feb 25 09:14:05 crc kubenswrapper[4978]: I0225 09:14:05.343566 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533508-28cvw"] Feb 25 09:14:05 crc kubenswrapper[4978]: I0225 09:14:05.343609 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533508-28cvw"] Feb 25 09:14:07 crc kubenswrapper[4978]: I0225 09:14:07.285823 4978 generic.go:334] "Generic (PLEG): container finished" podID="87856568-5e2c-40f3-99cd-a159a8ae96e9" containerID="d6538069a351992b3ac515d6a7b7974f63fbf6f89648a6b961001917884413d8" exitCode=0 Feb 25 09:14:07 crc kubenswrapper[4978]: I0225 09:14:07.285918 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" event={"ID":"87856568-5e2c-40f3-99cd-a159a8ae96e9","Type":"ContainerDied","Data":"d6538069a351992b3ac515d6a7b7974f63fbf6f89648a6b961001917884413d8"} Feb 25 09:14:07 crc kubenswrapper[4978]: I0225 09:14:07.340234 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d7e4fd9-5585-44d8-9498-dd11cec75fa8" path="/var/lib/kubelet/pods/4d7e4fd9-5585-44d8-9498-dd11cec75fa8/volumes" Feb 25 09:14:08 crc kubenswrapper[4978]: I0225 09:14:08.835688 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:14:08 crc kubenswrapper[4978]: I0225 09:14:08.941831 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxpf6\" (UniqueName: \"kubernetes.io/projected/87856568-5e2c-40f3-99cd-a159a8ae96e9-kube-api-access-sxpf6\") pod \"87856568-5e2c-40f3-99cd-a159a8ae96e9\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " Feb 25 09:14:08 crc kubenswrapper[4978]: I0225 09:14:08.941939 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-inventory\") pod \"87856568-5e2c-40f3-99cd-a159a8ae96e9\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " Feb 25 09:14:08 crc kubenswrapper[4978]: I0225 09:14:08.942107 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-ssh-key-openstack-networker\") pod \"87856568-5e2c-40f3-99cd-a159a8ae96e9\" (UID: \"87856568-5e2c-40f3-99cd-a159a8ae96e9\") " Feb 25 09:14:08 crc kubenswrapper[4978]: I0225 09:14:08.953232 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87856568-5e2c-40f3-99cd-a159a8ae96e9-kube-api-access-sxpf6" (OuterVolumeSpecName: "kube-api-access-sxpf6") pod "87856568-5e2c-40f3-99cd-a159a8ae96e9" (UID: "87856568-5e2c-40f3-99cd-a159a8ae96e9"). InnerVolumeSpecName "kube-api-access-sxpf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:14:08 crc kubenswrapper[4978]: I0225 09:14:08.976899 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-inventory" (OuterVolumeSpecName: "inventory") pod "87856568-5e2c-40f3-99cd-a159a8ae96e9" (UID: "87856568-5e2c-40f3-99cd-a159a8ae96e9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:08 crc kubenswrapper[4978]: I0225 09:14:08.995845 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "87856568-5e2c-40f3-99cd-a159a8ae96e9" (UID: "87856568-5e2c-40f3-99cd-a159a8ae96e9"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.044122 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxpf6\" (UniqueName: \"kubernetes.io/projected/87856568-5e2c-40f3-99cd-a159a8ae96e9-kube-api-access-sxpf6\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.044171 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.044185 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/87856568-5e2c-40f3-99cd-a159a8ae96e9-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.306083 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" event={"ID":"87856568-5e2c-40f3-99cd-a159a8ae96e9","Type":"ContainerDied","Data":"4579dad824acf2df67f2bb6421fdf5eb5a66ad300cba123fd83ded1fcfd6d646"} Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.306128 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4579dad824acf2df67f2bb6421fdf5eb5a66ad300cba123fd83ded1fcfd6d646" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.306176 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-networker-ddfpk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.419290 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-networker-q99kk"] Feb 25 09:14:09 crc kubenswrapper[4978]: E0225 09:14:09.420128 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87856568-5e2c-40f3-99cd-a159a8ae96e9" containerName="reboot-os-openstack-openstack-networker" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.420171 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="87856568-5e2c-40f3-99cd-a159a8ae96e9" containerName="reboot-os-openstack-openstack-networker" Feb 25 09:14:09 crc kubenswrapper[4978]: E0225 09:14:09.420210 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ae641dc-1ef2-464e-b1da-6d7341626b34" containerName="oc" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.420220 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ae641dc-1ef2-464e-b1da-6d7341626b34" containerName="oc" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.420579 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ae641dc-1ef2-464e-b1da-6d7341626b34" containerName="oc" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.420603 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="87856568-5e2c-40f3-99cd-a159a8ae96e9" containerName="reboot-os-openstack-openstack-networker" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.421656 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.427032 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-neutron-metadata-default-certs-0" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.427212 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-networker-q99kk"] Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.427321 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.427387 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.427640 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-ovn-default-certs-0" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.561819 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.561955 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-inventory\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.562011 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.562075 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.562125 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.562147 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sx2h\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-kube-api-access-4sx2h\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.562260 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.562304 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ssh-key-openstack-networker\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.663638 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.663690 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.663709 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sx2h\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-kube-api-access-4sx2h\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.663766 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.663791 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ssh-key-openstack-networker\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.663834 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.663902 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-inventory\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.663935 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.670292 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.670421 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.670475 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-inventory\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.671305 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.671834 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.672575 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.675414 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ssh-key-openstack-networker\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.693517 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sx2h\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-kube-api-access-4sx2h\") pod \"install-certs-openstack-openstack-networker-q99kk\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:09 crc kubenswrapper[4978]: I0225 09:14:09.741005 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:10 crc kubenswrapper[4978]: I0225 09:14:10.502015 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-networker-q99kk"] Feb 25 09:14:10 crc kubenswrapper[4978]: W0225 09:14:10.502519 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4555309b_6018_4eb3_ae3b_0f7ab8d79e10.slice/crio-b3e6e4d1368ed054ba5cb0faa07586c5df3f5ca90ae12b264e76b4cbfaba66bb WatchSource:0}: Error finding container b3e6e4d1368ed054ba5cb0faa07586c5df3f5ca90ae12b264e76b4cbfaba66bb: Status 404 returned error can't find the container with id b3e6e4d1368ed054ba5cb0faa07586c5df3f5ca90ae12b264e76b4cbfaba66bb Feb 25 09:14:11 crc kubenswrapper[4978]: I0225 09:14:11.321100 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-q99kk" event={"ID":"4555309b-6018-4eb3-ae3b-0f7ab8d79e10","Type":"ContainerStarted","Data":"4146e0f62d131d97267626c7a46308dadbbd801a15adc08405695beabf7b72bd"} Feb 25 09:14:11 crc kubenswrapper[4978]: I0225 09:14:11.321622 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-q99kk" event={"ID":"4555309b-6018-4eb3-ae3b-0f7ab8d79e10","Type":"ContainerStarted","Data":"b3e6e4d1368ed054ba5cb0faa07586c5df3f5ca90ae12b264e76b4cbfaba66bb"} Feb 25 09:14:11 crc kubenswrapper[4978]: I0225 09:14:11.348684 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-networker-q99kk" podStartSLOduration=1.925689642 podStartE2EDuration="2.348663102s" podCreationTimestamp="2026-02-25 09:14:09 +0000 UTC" firstStartedPulling="2026-02-25 09:14:10.506899492 +0000 UTC m=+8943.946155971" lastFinishedPulling="2026-02-25 09:14:10.929872972 +0000 UTC m=+8944.369129431" observedRunningTime="2026-02-25 09:14:11.343219624 +0000 UTC m=+8944.782476073" watchObservedRunningTime="2026-02-25 09:14:11.348663102 +0000 UTC m=+8944.787919561" Feb 25 09:14:16 crc kubenswrapper[4978]: I0225 09:14:16.540252 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:14:16 crc kubenswrapper[4978]: I0225 09:14:16.540686 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:14:16 crc kubenswrapper[4978]: I0225 09:14:16.561163 4978 generic.go:334] "Generic (PLEG): container finished" podID="07f0da72-81c7-4452-8662-bc63584cad91" containerID="bd1bf89277289572a19976a84d3b738d4ec01f9ac57b48ddf2d347378c849d5b" exitCode=0 Feb 25 09:14:16 crc kubenswrapper[4978]: I0225 09:14:16.561216 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" event={"ID":"07f0da72-81c7-4452-8662-bc63584cad91","Type":"ContainerDied","Data":"bd1bf89277289572a19976a84d3b738d4ec01f9ac57b48ddf2d347378c849d5b"} Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.043752 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.150135 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjd85\" (UniqueName: \"kubernetes.io/projected/07f0da72-81c7-4452-8662-bc63584cad91-kube-api-access-tjd85\") pod \"07f0da72-81c7-4452-8662-bc63584cad91\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.150308 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-inventory\") pod \"07f0da72-81c7-4452-8662-bc63584cad91\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.150380 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-ssh-key-openstack-cell1\") pod \"07f0da72-81c7-4452-8662-bc63584cad91\" (UID: \"07f0da72-81c7-4452-8662-bc63584cad91\") " Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.156114 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07f0da72-81c7-4452-8662-bc63584cad91-kube-api-access-tjd85" (OuterVolumeSpecName: "kube-api-access-tjd85") pod "07f0da72-81c7-4452-8662-bc63584cad91" (UID: "07f0da72-81c7-4452-8662-bc63584cad91"). InnerVolumeSpecName "kube-api-access-tjd85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.180540 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-inventory" (OuterVolumeSpecName: "inventory") pod "07f0da72-81c7-4452-8662-bc63584cad91" (UID: "07f0da72-81c7-4452-8662-bc63584cad91"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.182017 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "07f0da72-81c7-4452-8662-bc63584cad91" (UID: "07f0da72-81c7-4452-8662-bc63584cad91"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.252434 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjd85\" (UniqueName: \"kubernetes.io/projected/07f0da72-81c7-4452-8662-bc63584cad91-kube-api-access-tjd85\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.252563 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.252625 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/07f0da72-81c7-4452-8662-bc63584cad91-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.578167 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" event={"ID":"07f0da72-81c7-4452-8662-bc63584cad91","Type":"ContainerDied","Data":"9a5d23555ac6e160c18cc18892b4184552ed19e4061ced11c0ffe1bbfa6e0b56"} Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.578471 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a5d23555ac6e160c18cc18892b4184552ed19e4061ced11c0ffe1bbfa6e0b56" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.578241 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-xbb8f" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.709224 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-9fl9m"] Feb 25 09:14:18 crc kubenswrapper[4978]: E0225 09:14:18.709714 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07f0da72-81c7-4452-8662-bc63584cad91" containerName="configure-os-openstack-openstack-cell1" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.709732 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="07f0da72-81c7-4452-8662-bc63584cad91" containerName="configure-os-openstack-openstack-cell1" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.709947 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="07f0da72-81c7-4452-8662-bc63584cad91" containerName="configure-os-openstack-openstack-cell1" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.710712 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.714951 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.716311 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.722569 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-9fl9m"] Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.762282 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lp6z\" (UniqueName: \"kubernetes.io/projected/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-kube-api-access-4lp6z\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.762487 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.762565 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-1\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.762845 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-0\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.762890 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.865175 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.865341 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lp6z\" (UniqueName: \"kubernetes.io/projected/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-kube-api-access-4lp6z\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.865475 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.865515 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-1\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.865624 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-0\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.873037 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-0\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.873194 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.873741 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-networker\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.874190 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-1\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:18 crc kubenswrapper[4978]: I0225 09:14:18.886613 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lp6z\" (UniqueName: \"kubernetes.io/projected/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-kube-api-access-4lp6z\") pod \"ssh-known-hosts-openstack-9fl9m\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:19 crc kubenswrapper[4978]: I0225 09:14:19.027921 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:19 crc kubenswrapper[4978]: I0225 09:14:19.647707 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-9fl9m"] Feb 25 09:14:20 crc kubenswrapper[4978]: I0225 09:14:20.599429 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-9fl9m" event={"ID":"06f07aaf-4685-457d-b8e5-da1b4f2daa9b","Type":"ContainerStarted","Data":"a4742fd3831552b0f1cb2d57e0896cd768f254db36d31f48ad501483426b33a7"} Feb 25 09:14:20 crc kubenswrapper[4978]: I0225 09:14:20.599795 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-9fl9m" event={"ID":"06f07aaf-4685-457d-b8e5-da1b4f2daa9b","Type":"ContainerStarted","Data":"a658d37fc38d6179be619b4d2a432bfd82f7704f260013733f59cfec1a9412c8"} Feb 25 09:14:20 crc kubenswrapper[4978]: I0225 09:14:20.631432 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-9fl9m" podStartSLOduration=2.235133403 podStartE2EDuration="2.631412396s" podCreationTimestamp="2026-02-25 09:14:18 +0000 UTC" firstStartedPulling="2026-02-25 09:14:19.651180108 +0000 UTC m=+8953.090436607" lastFinishedPulling="2026-02-25 09:14:20.047459101 +0000 UTC m=+8953.486715600" observedRunningTime="2026-02-25 09:14:20.62381523 +0000 UTC m=+8954.063071729" watchObservedRunningTime="2026-02-25 09:14:20.631412396 +0000 UTC m=+8954.070668865" Feb 25 09:14:31 crc kubenswrapper[4978]: I0225 09:14:31.722030 4978 generic.go:334] "Generic (PLEG): container finished" podID="4555309b-6018-4eb3-ae3b-0f7ab8d79e10" containerID="4146e0f62d131d97267626c7a46308dadbbd801a15adc08405695beabf7b72bd" exitCode=0 Feb 25 09:14:31 crc kubenswrapper[4978]: I0225 09:14:31.722164 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-q99kk" event={"ID":"4555309b-6018-4eb3-ae3b-0f7ab8d79e10","Type":"ContainerDied","Data":"4146e0f62d131d97267626c7a46308dadbbd801a15adc08405695beabf7b72bd"} Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.184271 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.303687 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-bootstrap-combined-ca-bundle\") pod \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.304015 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-neutron-metadata-default-certs-0\") pod \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.304056 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-ovn-default-certs-0\") pod \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.304088 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-inventory\") pod \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.304257 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sx2h\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-kube-api-access-4sx2h\") pod \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.304277 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-neutron-metadata-combined-ca-bundle\") pod \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.304293 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ssh-key-openstack-networker\") pod \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.304346 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ovn-combined-ca-bundle\") pod \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\" (UID: \"4555309b-6018-4eb3-ae3b-0f7ab8d79e10\") " Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.310944 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4555309b-6018-4eb3-ae3b-0f7ab8d79e10" (UID: "4555309b-6018-4eb3-ae3b-0f7ab8d79e10"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.311682 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4555309b-6018-4eb3-ae3b-0f7ab8d79e10" (UID: "4555309b-6018-4eb3-ae3b-0f7ab8d79e10"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.312705 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-kube-api-access-4sx2h" (OuterVolumeSpecName: "kube-api-access-4sx2h") pod "4555309b-6018-4eb3-ae3b-0f7ab8d79e10" (UID: "4555309b-6018-4eb3-ae3b-0f7ab8d79e10"). InnerVolumeSpecName "kube-api-access-4sx2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.314336 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-networker-ovn-default-certs-0") pod "4555309b-6018-4eb3-ae3b-0f7ab8d79e10" (UID: "4555309b-6018-4eb3-ae3b-0f7ab8d79e10"). InnerVolumeSpecName "openstack-networker-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.314700 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-networker-neutron-metadata-default-certs-0") pod "4555309b-6018-4eb3-ae3b-0f7ab8d79e10" (UID: "4555309b-6018-4eb3-ae3b-0f7ab8d79e10"). InnerVolumeSpecName "openstack-networker-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.315787 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4555309b-6018-4eb3-ae3b-0f7ab8d79e10" (UID: "4555309b-6018-4eb3-ae3b-0f7ab8d79e10"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.351000 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "4555309b-6018-4eb3-ae3b-0f7ab8d79e10" (UID: "4555309b-6018-4eb3-ae3b-0f7ab8d79e10"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.357793 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-inventory" (OuterVolumeSpecName: "inventory") pod "4555309b-6018-4eb3-ae3b-0f7ab8d79e10" (UID: "4555309b-6018-4eb3-ae3b-0f7ab8d79e10"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.408970 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sx2h\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-kube-api-access-4sx2h\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.409027 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.409052 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.409074 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.409094 4978 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.409114 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-networker-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.409135 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-networker-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-openstack-networker-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.409158 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4555309b-6018-4eb3-ae3b-0f7ab8d79e10-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.745357 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-networker-q99kk" event={"ID":"4555309b-6018-4eb3-ae3b-0f7ab8d79e10","Type":"ContainerDied","Data":"b3e6e4d1368ed054ba5cb0faa07586c5df3f5ca90ae12b264e76b4cbfaba66bb"} Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.745427 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3e6e4d1368ed054ba5cb0faa07586c5df3f5ca90ae12b264e76b4cbfaba66bb" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.745468 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-networker-q99kk" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.930323 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-networker-fxwqp"] Feb 25 09:14:33 crc kubenswrapper[4978]: E0225 09:14:33.930730 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4555309b-6018-4eb3-ae3b-0f7ab8d79e10" containerName="install-certs-openstack-openstack-networker" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.930749 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4555309b-6018-4eb3-ae3b-0f7ab8d79e10" containerName="install-certs-openstack-openstack-networker" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.930924 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4555309b-6018-4eb3-ae3b-0f7ab8d79e10" containerName="install-certs-openstack-openstack-networker" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.933337 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.939135 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.939563 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:14:33 crc kubenswrapper[4978]: I0225 09:14:33.946423 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-networker-fxwqp"] Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.024427 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ssh-key-openstack-networker\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.024799 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn69z\" (UniqueName: \"kubernetes.io/projected/35bf1547-15c8-4591-945f-f185c31b9660-kube-api-access-pn69z\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.024981 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.025104 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-inventory\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.025221 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/35bf1547-15c8-4591-945f-f185c31b9660-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.127189 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn69z\" (UniqueName: \"kubernetes.io/projected/35bf1547-15c8-4591-945f-f185c31b9660-kube-api-access-pn69z\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.127326 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.127392 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-inventory\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.127415 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/35bf1547-15c8-4591-945f-f185c31b9660-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.127454 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ssh-key-openstack-networker\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.129157 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/35bf1547-15c8-4591-945f-f185c31b9660-ovncontroller-config-0\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.130659 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.131046 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-inventory\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.131740 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ssh-key-openstack-networker\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.145243 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn69z\" (UniqueName: \"kubernetes.io/projected/35bf1547-15c8-4591-945f-f185c31b9660-kube-api-access-pn69z\") pod \"ovn-openstack-openstack-networker-fxwqp\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.259147 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:14:34 crc kubenswrapper[4978]: I0225 09:14:34.819741 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-networker-fxwqp"] Feb 25 09:14:34 crc kubenswrapper[4978]: W0225 09:14:34.825715 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35bf1547_15c8_4591_945f_f185c31b9660.slice/crio-c1c55638c09a153e8fd72f32fefec221168b1482c4a940e7e069e5c7364497cf WatchSource:0}: Error finding container c1c55638c09a153e8fd72f32fefec221168b1482c4a940e7e069e5c7364497cf: Status 404 returned error can't find the container with id c1c55638c09a153e8fd72f32fefec221168b1482c4a940e7e069e5c7364497cf Feb 25 09:14:35 crc kubenswrapper[4978]: I0225 09:14:35.776277 4978 generic.go:334] "Generic (PLEG): container finished" podID="06f07aaf-4685-457d-b8e5-da1b4f2daa9b" containerID="a4742fd3831552b0f1cb2d57e0896cd768f254db36d31f48ad501483426b33a7" exitCode=0 Feb 25 09:14:35 crc kubenswrapper[4978]: I0225 09:14:35.776462 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-9fl9m" event={"ID":"06f07aaf-4685-457d-b8e5-da1b4f2daa9b","Type":"ContainerDied","Data":"a4742fd3831552b0f1cb2d57e0896cd768f254db36d31f48ad501483426b33a7"} Feb 25 09:14:35 crc kubenswrapper[4978]: I0225 09:14:35.779519 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-fxwqp" event={"ID":"35bf1547-15c8-4591-945f-f185c31b9660","Type":"ContainerStarted","Data":"c1c55638c09a153e8fd72f32fefec221168b1482c4a940e7e069e5c7364497cf"} Feb 25 09:14:35 crc kubenswrapper[4978]: I0225 09:14:35.789797 4978 scope.go:117] "RemoveContainer" containerID="dbd06cbbc2d528d2b04a966950f772bc7c45d2eac157596ed7d04f560338dd56" Feb 25 09:14:36 crc kubenswrapper[4978]: I0225 09:14:36.793296 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-fxwqp" event={"ID":"35bf1547-15c8-4591-945f-f185c31b9660","Type":"ContainerStarted","Data":"184bd6febc1f0114f6c6be8701fdac5e7f1a3befe743ffc55922ef9f2f2bc0bf"} Feb 25 09:14:36 crc kubenswrapper[4978]: I0225 09:14:36.845206 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-networker-fxwqp" podStartSLOduration=2.602884568 podStartE2EDuration="3.845179941s" podCreationTimestamp="2026-02-25 09:14:33 +0000 UTC" firstStartedPulling="2026-02-25 09:14:34.830269799 +0000 UTC m=+8968.269526268" lastFinishedPulling="2026-02-25 09:14:36.072565182 +0000 UTC m=+8969.511821641" observedRunningTime="2026-02-25 09:14:36.823743057 +0000 UTC m=+8970.262999516" watchObservedRunningTime="2026-02-25 09:14:36.845179941 +0000 UTC m=+8970.284436430" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.347037 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.509099 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-networker\") pod \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.509320 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lp6z\" (UniqueName: \"kubernetes.io/projected/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-kube-api-access-4lp6z\") pod \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.509449 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-cell1\") pod \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.509503 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-1\") pod \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.509643 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-0\") pod \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\" (UID: \"06f07aaf-4685-457d-b8e5-da1b4f2daa9b\") " Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.522688 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-kube-api-access-4lp6z" (OuterVolumeSpecName: "kube-api-access-4lp6z") pod "06f07aaf-4685-457d-b8e5-da1b4f2daa9b" (UID: "06f07aaf-4685-457d-b8e5-da1b4f2daa9b"). InnerVolumeSpecName "kube-api-access-4lp6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.539735 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "06f07aaf-4685-457d-b8e5-da1b4f2daa9b" (UID: "06f07aaf-4685-457d-b8e5-da1b4f2daa9b"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.552772 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-1" (OuterVolumeSpecName: "inventory-1") pod "06f07aaf-4685-457d-b8e5-da1b4f2daa9b" (UID: "06f07aaf-4685-457d-b8e5-da1b4f2daa9b"). InnerVolumeSpecName "inventory-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.565352 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "06f07aaf-4685-457d-b8e5-da1b4f2daa9b" (UID: "06f07aaf-4685-457d-b8e5-da1b4f2daa9b"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.588792 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "06f07aaf-4685-457d-b8e5-da1b4f2daa9b" (UID: "06f07aaf-4685-457d-b8e5-da1b4f2daa9b"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.612425 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.612461 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lp6z\" (UniqueName: \"kubernetes.io/projected/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-kube-api-access-4lp6z\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.612474 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.612488 4978 reconciler_common.go:293] "Volume detached for volume \"inventory-1\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-1\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.612503 4978 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06f07aaf-4685-457d-b8e5-da1b4f2daa9b-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.807267 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-9fl9m" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.807280 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-9fl9m" event={"ID":"06f07aaf-4685-457d-b8e5-da1b4f2daa9b","Type":"ContainerDied","Data":"a658d37fc38d6179be619b4d2a432bfd82f7704f260013733f59cfec1a9412c8"} Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.807356 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a658d37fc38d6179be619b4d2a432bfd82f7704f260013733f59cfec1a9412c8" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.943481 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-tkbrw"] Feb 25 09:14:37 crc kubenswrapper[4978]: E0225 09:14:37.944211 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f07aaf-4685-457d-b8e5-da1b4f2daa9b" containerName="ssh-known-hosts-openstack" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.944245 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f07aaf-4685-457d-b8e5-da1b4f2daa9b" containerName="ssh-known-hosts-openstack" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.955429 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f07aaf-4685-457d-b8e5-da1b4f2daa9b" containerName="ssh-known-hosts-openstack" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.956924 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.958942 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-tkbrw"] Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.960923 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:14:37 crc kubenswrapper[4978]: I0225 09:14:37.961656 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.123086 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-inventory\") pod \"run-os-openstack-openstack-cell1-tkbrw\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.123349 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7ch9\" (UniqueName: \"kubernetes.io/projected/1ca465d1-d912-48fc-984f-62f645da23cd-kube-api-access-w7ch9\") pod \"run-os-openstack-openstack-cell1-tkbrw\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.123423 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-tkbrw\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.226183 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-inventory\") pod \"run-os-openstack-openstack-cell1-tkbrw\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.226274 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7ch9\" (UniqueName: \"kubernetes.io/projected/1ca465d1-d912-48fc-984f-62f645da23cd-kube-api-access-w7ch9\") pod \"run-os-openstack-openstack-cell1-tkbrw\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.226331 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-tkbrw\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.233338 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-inventory\") pod \"run-os-openstack-openstack-cell1-tkbrw\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.239462 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-tkbrw\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.246121 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7ch9\" (UniqueName: \"kubernetes.io/projected/1ca465d1-d912-48fc-984f-62f645da23cd-kube-api-access-w7ch9\") pod \"run-os-openstack-openstack-cell1-tkbrw\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.280607 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:38 crc kubenswrapper[4978]: I0225 09:14:38.944970 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-tkbrw"] Feb 25 09:14:39 crc kubenswrapper[4978]: I0225 09:14:39.836983 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tkbrw" event={"ID":"1ca465d1-d912-48fc-984f-62f645da23cd","Type":"ContainerStarted","Data":"241dfc4bf21c7e6b97f317837a44c3e5624458f6ef8d3fafb068192063901484"} Feb 25 09:14:39 crc kubenswrapper[4978]: I0225 09:14:39.837475 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tkbrw" event={"ID":"1ca465d1-d912-48fc-984f-62f645da23cd","Type":"ContainerStarted","Data":"65ea9bd91d58b22bc9718e2104b3dbfa74a20e8fe326d7b7fc4ef5db87f75a8a"} Feb 25 09:14:39 crc kubenswrapper[4978]: I0225 09:14:39.876398 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-tkbrw" podStartSLOduration=2.4121815189999998 podStartE2EDuration="2.876352185s" podCreationTimestamp="2026-02-25 09:14:37 +0000 UTC" firstStartedPulling="2026-02-25 09:14:38.956608761 +0000 UTC m=+8972.395865230" lastFinishedPulling="2026-02-25 09:14:39.420779397 +0000 UTC m=+8972.860035896" observedRunningTime="2026-02-25 09:14:39.86262746 +0000 UTC m=+8973.301883919" watchObservedRunningTime="2026-02-25 09:14:39.876352185 +0000 UTC m=+8973.315608644" Feb 25 09:14:46 crc kubenswrapper[4978]: I0225 09:14:46.539932 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:14:46 crc kubenswrapper[4978]: I0225 09:14:46.540527 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:14:46 crc kubenswrapper[4978]: I0225 09:14:46.540573 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 09:14:46 crc kubenswrapper[4978]: I0225 09:14:46.541412 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6e09bc70e0ad7207309daaa14091dbcb18dfd0d3b48bf2f83a1437dfb4535285"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 09:14:46 crc kubenswrapper[4978]: I0225 09:14:46.541470 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://6e09bc70e0ad7207309daaa14091dbcb18dfd0d3b48bf2f83a1437dfb4535285" gracePeriod=600 Feb 25 09:14:46 crc kubenswrapper[4978]: I0225 09:14:46.940444 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="6e09bc70e0ad7207309daaa14091dbcb18dfd0d3b48bf2f83a1437dfb4535285" exitCode=0 Feb 25 09:14:46 crc kubenswrapper[4978]: I0225 09:14:46.941922 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"6e09bc70e0ad7207309daaa14091dbcb18dfd0d3b48bf2f83a1437dfb4535285"} Feb 25 09:14:46 crc kubenswrapper[4978]: I0225 09:14:46.942174 4978 scope.go:117] "RemoveContainer" containerID="8a3657bf1bf82966846ca3d6f433da0a5099cb2b2acfa52a807185fe778b3abe" Feb 25 09:14:47 crc kubenswrapper[4978]: I0225 09:14:47.957125 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4"} Feb 25 09:14:48 crc kubenswrapper[4978]: I0225 09:14:48.970213 4978 generic.go:334] "Generic (PLEG): container finished" podID="1ca465d1-d912-48fc-984f-62f645da23cd" containerID="241dfc4bf21c7e6b97f317837a44c3e5624458f6ef8d3fafb068192063901484" exitCode=0 Feb 25 09:14:48 crc kubenswrapper[4978]: I0225 09:14:48.970309 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tkbrw" event={"ID":"1ca465d1-d912-48fc-984f-62f645da23cd","Type":"ContainerDied","Data":"241dfc4bf21c7e6b97f317837a44c3e5624458f6ef8d3fafb068192063901484"} Feb 25 09:14:50 crc kubenswrapper[4978]: I0225 09:14:50.552736 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:50 crc kubenswrapper[4978]: I0225 09:14:50.589686 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-inventory\") pod \"1ca465d1-d912-48fc-984f-62f645da23cd\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " Feb 25 09:14:50 crc kubenswrapper[4978]: I0225 09:14:50.590207 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-ssh-key-openstack-cell1\") pod \"1ca465d1-d912-48fc-984f-62f645da23cd\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " Feb 25 09:14:50 crc kubenswrapper[4978]: I0225 09:14:50.590261 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7ch9\" (UniqueName: \"kubernetes.io/projected/1ca465d1-d912-48fc-984f-62f645da23cd-kube-api-access-w7ch9\") pod \"1ca465d1-d912-48fc-984f-62f645da23cd\" (UID: \"1ca465d1-d912-48fc-984f-62f645da23cd\") " Feb 25 09:14:50 crc kubenswrapper[4978]: I0225 09:14:50.609006 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca465d1-d912-48fc-984f-62f645da23cd-kube-api-access-w7ch9" (OuterVolumeSpecName: "kube-api-access-w7ch9") pod "1ca465d1-d912-48fc-984f-62f645da23cd" (UID: "1ca465d1-d912-48fc-984f-62f645da23cd"). InnerVolumeSpecName "kube-api-access-w7ch9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:14:50 crc kubenswrapper[4978]: I0225 09:14:50.635694 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-inventory" (OuterVolumeSpecName: "inventory") pod "1ca465d1-d912-48fc-984f-62f645da23cd" (UID: "1ca465d1-d912-48fc-984f-62f645da23cd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:50 crc kubenswrapper[4978]: I0225 09:14:50.640842 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "1ca465d1-d912-48fc-984f-62f645da23cd" (UID: "1ca465d1-d912-48fc-984f-62f645da23cd"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:14:50 crc kubenswrapper[4978]: I0225 09:14:50.693398 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:50 crc kubenswrapper[4978]: I0225 09:14:50.693473 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/1ca465d1-d912-48fc-984f-62f645da23cd-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:50 crc kubenswrapper[4978]: I0225 09:14:50.693494 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7ch9\" (UniqueName: \"kubernetes.io/projected/1ca465d1-d912-48fc-984f-62f645da23cd-kube-api-access-w7ch9\") on node \"crc\" DevicePath \"\"" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.004424 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-tkbrw" event={"ID":"1ca465d1-d912-48fc-984f-62f645da23cd","Type":"ContainerDied","Data":"65ea9bd91d58b22bc9718e2104b3dbfa74a20e8fe326d7b7fc4ef5db87f75a8a"} Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.004570 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65ea9bd91d58b22bc9718e2104b3dbfa74a20e8fe326d7b7fc4ef5db87f75a8a" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.004743 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-tkbrw" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.161727 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mbr4t"] Feb 25 09:14:51 crc kubenswrapper[4978]: E0225 09:14:51.162527 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca465d1-d912-48fc-984f-62f645da23cd" containerName="run-os-openstack-openstack-cell1" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.162547 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca465d1-d912-48fc-984f-62f645da23cd" containerName="run-os-openstack-openstack-cell1" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.162831 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca465d1-d912-48fc-984f-62f645da23cd" containerName="run-os-openstack-openstack-cell1" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.163725 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.166177 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.169047 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.180674 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mbr4t"] Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.205056 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-mbr4t\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.205147 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-inventory\") pod \"reboot-os-openstack-openstack-cell1-mbr4t\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.205197 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzpcs\" (UniqueName: \"kubernetes.io/projected/356ba162-9c45-4a68-b891-455d1e99065f-kube-api-access-gzpcs\") pod \"reboot-os-openstack-openstack-cell1-mbr4t\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.307085 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-mbr4t\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.307161 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-inventory\") pod \"reboot-os-openstack-openstack-cell1-mbr4t\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.307186 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzpcs\" (UniqueName: \"kubernetes.io/projected/356ba162-9c45-4a68-b891-455d1e99065f-kube-api-access-gzpcs\") pod \"reboot-os-openstack-openstack-cell1-mbr4t\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.325505 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-mbr4t\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.331870 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-inventory\") pod \"reboot-os-openstack-openstack-cell1-mbr4t\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.336464 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzpcs\" (UniqueName: \"kubernetes.io/projected/356ba162-9c45-4a68-b891-455d1e99065f-kube-api-access-gzpcs\") pod \"reboot-os-openstack-openstack-cell1-mbr4t\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:51 crc kubenswrapper[4978]: I0225 09:14:51.502855 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:14:52 crc kubenswrapper[4978]: I0225 09:14:52.092201 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-mbr4t"] Feb 25 09:14:52 crc kubenswrapper[4978]: W0225 09:14:52.097154 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod356ba162_9c45_4a68_b891_455d1e99065f.slice/crio-18fdb7e8b4c5ade7676075b68a4c2249a711c5eada49c533c22d00f74830ea9a WatchSource:0}: Error finding container 18fdb7e8b4c5ade7676075b68a4c2249a711c5eada49c533c22d00f74830ea9a: Status 404 returned error can't find the container with id 18fdb7e8b4c5ade7676075b68a4c2249a711c5eada49c533c22d00f74830ea9a Feb 25 09:14:53 crc kubenswrapper[4978]: I0225 09:14:53.030256 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" event={"ID":"356ba162-9c45-4a68-b891-455d1e99065f","Type":"ContainerStarted","Data":"a793dad786051915623c473d13de1e4317ff06c1c246535516a8932cd3593ca3"} Feb 25 09:14:53 crc kubenswrapper[4978]: I0225 09:14:53.030639 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" event={"ID":"356ba162-9c45-4a68-b891-455d1e99065f","Type":"ContainerStarted","Data":"18fdb7e8b4c5ade7676075b68a4c2249a711c5eada49c533c22d00f74830ea9a"} Feb 25 09:14:53 crc kubenswrapper[4978]: I0225 09:14:53.070972 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" podStartSLOduration=1.7061933200000001 podStartE2EDuration="2.070945707s" podCreationTimestamp="2026-02-25 09:14:51 +0000 UTC" firstStartedPulling="2026-02-25 09:14:52.104004301 +0000 UTC m=+8985.543260770" lastFinishedPulling="2026-02-25 09:14:52.468756698 +0000 UTC m=+8985.908013157" observedRunningTime="2026-02-25 09:14:53.057333066 +0000 UTC m=+8986.496589545" watchObservedRunningTime="2026-02-25 09:14:53.070945707 +0000 UTC m=+8986.510202176" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.200518 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv"] Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.214155 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv"] Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.214315 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.217061 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.217435 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.318789 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a03ab396-abf8-41a4-a30d-b4698113a688-config-volume\") pod \"collect-profiles-29533515-tfbtv\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.318943 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqxls\" (UniqueName: \"kubernetes.io/projected/a03ab396-abf8-41a4-a30d-b4698113a688-kube-api-access-xqxls\") pod \"collect-profiles-29533515-tfbtv\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.319018 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a03ab396-abf8-41a4-a30d-b4698113a688-secret-volume\") pod \"collect-profiles-29533515-tfbtv\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.421520 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a03ab396-abf8-41a4-a30d-b4698113a688-config-volume\") pod \"collect-profiles-29533515-tfbtv\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.421699 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqxls\" (UniqueName: \"kubernetes.io/projected/a03ab396-abf8-41a4-a30d-b4698113a688-kube-api-access-xqxls\") pod \"collect-profiles-29533515-tfbtv\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.422609 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a03ab396-abf8-41a4-a30d-b4698113a688-secret-volume\") pod \"collect-profiles-29533515-tfbtv\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.423188 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a03ab396-abf8-41a4-a30d-b4698113a688-config-volume\") pod \"collect-profiles-29533515-tfbtv\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.438118 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a03ab396-abf8-41a4-a30d-b4698113a688-secret-volume\") pod \"collect-profiles-29533515-tfbtv\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.441809 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqxls\" (UniqueName: \"kubernetes.io/projected/a03ab396-abf8-41a4-a30d-b4698113a688-kube-api-access-xqxls\") pod \"collect-profiles-29533515-tfbtv\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:00 crc kubenswrapper[4978]: I0225 09:15:00.537269 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:01 crc kubenswrapper[4978]: I0225 09:15:01.000610 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv"] Feb 25 09:15:01 crc kubenswrapper[4978]: I0225 09:15:01.131181 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" event={"ID":"a03ab396-abf8-41a4-a30d-b4698113a688","Type":"ContainerStarted","Data":"3c36c2a50bb66cd0a634fe23fc35dc22e0bd99d73e6e364d8451113b0ccabce6"} Feb 25 09:15:02 crc kubenswrapper[4978]: I0225 09:15:02.144595 4978 generic.go:334] "Generic (PLEG): container finished" podID="a03ab396-abf8-41a4-a30d-b4698113a688" containerID="f02b504744df9f5d676874c9f2715fe0811c7c258ddc559d2e4aba9b09dd7e0b" exitCode=0 Feb 25 09:15:02 crc kubenswrapper[4978]: I0225 09:15:02.144657 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" event={"ID":"a03ab396-abf8-41a4-a30d-b4698113a688","Type":"ContainerDied","Data":"f02b504744df9f5d676874c9f2715fe0811c7c258ddc559d2e4aba9b09dd7e0b"} Feb 25 09:15:03 crc kubenswrapper[4978]: I0225 09:15:03.589477 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:03 crc kubenswrapper[4978]: I0225 09:15:03.698888 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a03ab396-abf8-41a4-a30d-b4698113a688-secret-volume\") pod \"a03ab396-abf8-41a4-a30d-b4698113a688\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " Feb 25 09:15:03 crc kubenswrapper[4978]: I0225 09:15:03.698969 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqxls\" (UniqueName: \"kubernetes.io/projected/a03ab396-abf8-41a4-a30d-b4698113a688-kube-api-access-xqxls\") pod \"a03ab396-abf8-41a4-a30d-b4698113a688\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " Feb 25 09:15:03 crc kubenswrapper[4978]: I0225 09:15:03.699081 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a03ab396-abf8-41a4-a30d-b4698113a688-config-volume\") pod \"a03ab396-abf8-41a4-a30d-b4698113a688\" (UID: \"a03ab396-abf8-41a4-a30d-b4698113a688\") " Feb 25 09:15:03 crc kubenswrapper[4978]: I0225 09:15:03.700027 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a03ab396-abf8-41a4-a30d-b4698113a688-config-volume" (OuterVolumeSpecName: "config-volume") pod "a03ab396-abf8-41a4-a30d-b4698113a688" (UID: "a03ab396-abf8-41a4-a30d-b4698113a688"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:15:03 crc kubenswrapper[4978]: I0225 09:15:03.705550 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a03ab396-abf8-41a4-a30d-b4698113a688-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a03ab396-abf8-41a4-a30d-b4698113a688" (UID: "a03ab396-abf8-41a4-a30d-b4698113a688"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:03 crc kubenswrapper[4978]: I0225 09:15:03.705982 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a03ab396-abf8-41a4-a30d-b4698113a688-kube-api-access-xqxls" (OuterVolumeSpecName: "kube-api-access-xqxls") pod "a03ab396-abf8-41a4-a30d-b4698113a688" (UID: "a03ab396-abf8-41a4-a30d-b4698113a688"). InnerVolumeSpecName "kube-api-access-xqxls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:15:03 crc kubenswrapper[4978]: I0225 09:15:03.801599 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a03ab396-abf8-41a4-a30d-b4698113a688-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:03 crc kubenswrapper[4978]: I0225 09:15:03.801639 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqxls\" (UniqueName: \"kubernetes.io/projected/a03ab396-abf8-41a4-a30d-b4698113a688-kube-api-access-xqxls\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:03 crc kubenswrapper[4978]: I0225 09:15:03.801656 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a03ab396-abf8-41a4-a30d-b4698113a688-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:04 crc kubenswrapper[4978]: I0225 09:15:04.171072 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" event={"ID":"a03ab396-abf8-41a4-a30d-b4698113a688","Type":"ContainerDied","Data":"3c36c2a50bb66cd0a634fe23fc35dc22e0bd99d73e6e364d8451113b0ccabce6"} Feb 25 09:15:04 crc kubenswrapper[4978]: I0225 09:15:04.171119 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv" Feb 25 09:15:04 crc kubenswrapper[4978]: I0225 09:15:04.171130 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c36c2a50bb66cd0a634fe23fc35dc22e0bd99d73e6e364d8451113b0ccabce6" Feb 25 09:15:04 crc kubenswrapper[4978]: I0225 09:15:04.685594 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p"] Feb 25 09:15:04 crc kubenswrapper[4978]: I0225 09:15:04.696697 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533470-9w64p"] Feb 25 09:15:05 crc kubenswrapper[4978]: I0225 09:15:05.348193 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59fb3f4d-8672-48ea-946e-5b768b1cb62a" path="/var/lib/kubelet/pods/59fb3f4d-8672-48ea-946e-5b768b1cb62a/volumes" Feb 25 09:15:08 crc kubenswrapper[4978]: I0225 09:15:08.221043 4978 generic.go:334] "Generic (PLEG): container finished" podID="356ba162-9c45-4a68-b891-455d1e99065f" containerID="a793dad786051915623c473d13de1e4317ff06c1c246535516a8932cd3593ca3" exitCode=0 Feb 25 09:15:08 crc kubenswrapper[4978]: I0225 09:15:08.221127 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" event={"ID":"356ba162-9c45-4a68-b891-455d1e99065f","Type":"ContainerDied","Data":"a793dad786051915623c473d13de1e4317ff06c1c246535516a8932cd3593ca3"} Feb 25 09:15:09 crc kubenswrapper[4978]: I0225 09:15:09.727429 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:15:09 crc kubenswrapper[4978]: I0225 09:15:09.841204 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-ssh-key-openstack-cell1\") pod \"356ba162-9c45-4a68-b891-455d1e99065f\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " Feb 25 09:15:09 crc kubenswrapper[4978]: I0225 09:15:09.841342 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzpcs\" (UniqueName: \"kubernetes.io/projected/356ba162-9c45-4a68-b891-455d1e99065f-kube-api-access-gzpcs\") pod \"356ba162-9c45-4a68-b891-455d1e99065f\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " Feb 25 09:15:09 crc kubenswrapper[4978]: I0225 09:15:09.841620 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-inventory\") pod \"356ba162-9c45-4a68-b891-455d1e99065f\" (UID: \"356ba162-9c45-4a68-b891-455d1e99065f\") " Feb 25 09:15:09 crc kubenswrapper[4978]: I0225 09:15:09.847562 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/356ba162-9c45-4a68-b891-455d1e99065f-kube-api-access-gzpcs" (OuterVolumeSpecName: "kube-api-access-gzpcs") pod "356ba162-9c45-4a68-b891-455d1e99065f" (UID: "356ba162-9c45-4a68-b891-455d1e99065f"). InnerVolumeSpecName "kube-api-access-gzpcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:15:09 crc kubenswrapper[4978]: I0225 09:15:09.880923 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-inventory" (OuterVolumeSpecName: "inventory") pod "356ba162-9c45-4a68-b891-455d1e99065f" (UID: "356ba162-9c45-4a68-b891-455d1e99065f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:09 crc kubenswrapper[4978]: I0225 09:15:09.886548 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "356ba162-9c45-4a68-b891-455d1e99065f" (UID: "356ba162-9c45-4a68-b891-455d1e99065f"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:09 crc kubenswrapper[4978]: I0225 09:15:09.944453 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:09 crc kubenswrapper[4978]: I0225 09:15:09.944490 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/356ba162-9c45-4a68-b891-455d1e99065f-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:09 crc kubenswrapper[4978]: I0225 09:15:09.944502 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzpcs\" (UniqueName: \"kubernetes.io/projected/356ba162-9c45-4a68-b891-455d1e99065f-kube-api-access-gzpcs\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.247769 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" event={"ID":"356ba162-9c45-4a68-b891-455d1e99065f","Type":"ContainerDied","Data":"18fdb7e8b4c5ade7676075b68a4c2249a711c5eada49c533c22d00f74830ea9a"} Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.247843 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18fdb7e8b4c5ade7676075b68a4c2249a711c5eada49c533c22d00f74830ea9a" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.247796 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-mbr4t" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.338504 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-7fcwt"] Feb 25 09:15:10 crc kubenswrapper[4978]: E0225 09:15:10.339220 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a03ab396-abf8-41a4-a30d-b4698113a688" containerName="collect-profiles" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.339314 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a03ab396-abf8-41a4-a30d-b4698113a688" containerName="collect-profiles" Feb 25 09:15:10 crc kubenswrapper[4978]: E0225 09:15:10.339417 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356ba162-9c45-4a68-b891-455d1e99065f" containerName="reboot-os-openstack-openstack-cell1" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.339499 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="356ba162-9c45-4a68-b891-455d1e99065f" containerName="reboot-os-openstack-openstack-cell1" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.339854 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a03ab396-abf8-41a4-a30d-b4698113a688" containerName="collect-profiles" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.339953 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="356ba162-9c45-4a68-b891-455d1e99065f" containerName="reboot-os-openstack-openstack-cell1" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.340797 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.342275 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.343649 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-libvirt-default-certs-0" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.344009 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.344302 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-neutron-metadata-default-certs-0" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.344341 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-telemetry-default-certs-0" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.344780 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-ovn-default-certs-0" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.358991 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-7fcwt"] Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.453970 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454086 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454127 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t94jj\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-kube-api-access-t94jj\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454196 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454340 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454474 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454551 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454604 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454657 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454706 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454786 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-inventory\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.454994 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.455085 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.455152 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.455197 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.558274 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-inventory\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.558434 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.558554 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.558636 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.558694 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.558784 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.558869 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.558907 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t94jj\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-kube-api-access-t94jj\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.558978 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.559055 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.559130 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.559209 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.559250 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.559288 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.559331 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.564496 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.566216 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.566523 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.566899 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.567204 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-inventory\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.567267 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.568883 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.570409 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.571294 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.571804 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.572521 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.573905 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.575064 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.576898 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.601540 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t94jj\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-kube-api-access-t94jj\") pod \"install-certs-openstack-openstack-cell1-7fcwt\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:10 crc kubenswrapper[4978]: I0225 09:15:10.678554 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:11 crc kubenswrapper[4978]: I0225 09:15:11.283159 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-7fcwt"] Feb 25 09:15:11 crc kubenswrapper[4978]: W0225 09:15:11.295803 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb722f89d_c453_4235_8d4c_8d9990f0551b.slice/crio-f340685706fa9edf9eeee6091fdf28749d5ca12c1af0463b0da99319180ce75d WatchSource:0}: Error finding container f340685706fa9edf9eeee6091fdf28749d5ca12c1af0463b0da99319180ce75d: Status 404 returned error can't find the container with id f340685706fa9edf9eeee6091fdf28749d5ca12c1af0463b0da99319180ce75d Feb 25 09:15:12 crc kubenswrapper[4978]: I0225 09:15:12.265356 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" event={"ID":"b722f89d-c453-4235-8d4c-8d9990f0551b","Type":"ContainerStarted","Data":"49f712a79e9450674b70464faa970d26338dc46a03cc44b7624dd08eccbed959"} Feb 25 09:15:12 crc kubenswrapper[4978]: I0225 09:15:12.267059 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" event={"ID":"b722f89d-c453-4235-8d4c-8d9990f0551b","Type":"ContainerStarted","Data":"f340685706fa9edf9eeee6091fdf28749d5ca12c1af0463b0da99319180ce75d"} Feb 25 09:15:12 crc kubenswrapper[4978]: I0225 09:15:12.310177 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" podStartSLOduration=1.8181568129999999 podStartE2EDuration="2.310157199s" podCreationTimestamp="2026-02-25 09:15:10 +0000 UTC" firstStartedPulling="2026-02-25 09:15:11.298980853 +0000 UTC m=+9004.738237312" lastFinishedPulling="2026-02-25 09:15:11.790981239 +0000 UTC m=+9005.230237698" observedRunningTime="2026-02-25 09:15:12.29889186 +0000 UTC m=+9005.738148329" watchObservedRunningTime="2026-02-25 09:15:12.310157199 +0000 UTC m=+9005.749413658" Feb 25 09:15:35 crc kubenswrapper[4978]: I0225 09:15:35.888545 4978 scope.go:117] "RemoveContainer" containerID="30549a3b33f6ec3dda56bbc9285e86c7d29f4a96d6f67e5b1a2257630e608d6e" Feb 25 09:15:35 crc kubenswrapper[4978]: I0225 09:15:35.917663 4978 scope.go:117] "RemoveContainer" containerID="975cdd422cdb7a0dbff5cbd8d5a90ea8450a4015b33bd7cf34b78e556318dc99" Feb 25 09:15:35 crc kubenswrapper[4978]: I0225 09:15:35.983230 4978 scope.go:117] "RemoveContainer" containerID="f37eaa18f90b4dd2d338d71b81ca4b16e44541947a22845767a9f6b1ece4b1a4" Feb 25 09:15:36 crc kubenswrapper[4978]: I0225 09:15:36.011184 4978 scope.go:117] "RemoveContainer" containerID="80cb84fc321be9bdaf63738d7cd737a060a425c1bc6c7009a312571b625f4536" Feb 25 09:15:47 crc kubenswrapper[4978]: I0225 09:15:47.704263 4978 generic.go:334] "Generic (PLEG): container finished" podID="35bf1547-15c8-4591-945f-f185c31b9660" containerID="184bd6febc1f0114f6c6be8701fdac5e7f1a3befe743ffc55922ef9f2f2bc0bf" exitCode=0 Feb 25 09:15:47 crc kubenswrapper[4978]: I0225 09:15:47.704397 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-fxwqp" event={"ID":"35bf1547-15c8-4591-945f-f185c31b9660","Type":"ContainerDied","Data":"184bd6febc1f0114f6c6be8701fdac5e7f1a3befe743ffc55922ef9f2f2bc0bf"} Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.221624 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.319228 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-inventory\") pod \"35bf1547-15c8-4591-945f-f185c31b9660\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.319456 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn69z\" (UniqueName: \"kubernetes.io/projected/35bf1547-15c8-4591-945f-f185c31b9660-kube-api-access-pn69z\") pod \"35bf1547-15c8-4591-945f-f185c31b9660\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.319533 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ovn-combined-ca-bundle\") pod \"35bf1547-15c8-4591-945f-f185c31b9660\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.319645 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/35bf1547-15c8-4591-945f-f185c31b9660-ovncontroller-config-0\") pod \"35bf1547-15c8-4591-945f-f185c31b9660\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.319669 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ssh-key-openstack-networker\") pod \"35bf1547-15c8-4591-945f-f185c31b9660\" (UID: \"35bf1547-15c8-4591-945f-f185c31b9660\") " Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.324482 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35bf1547-15c8-4591-945f-f185c31b9660-kube-api-access-pn69z" (OuterVolumeSpecName: "kube-api-access-pn69z") pod "35bf1547-15c8-4591-945f-f185c31b9660" (UID: "35bf1547-15c8-4591-945f-f185c31b9660"). InnerVolumeSpecName "kube-api-access-pn69z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.327009 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "35bf1547-15c8-4591-945f-f185c31b9660" (UID: "35bf1547-15c8-4591-945f-f185c31b9660"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.349253 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-inventory" (OuterVolumeSpecName: "inventory") pod "35bf1547-15c8-4591-945f-f185c31b9660" (UID: "35bf1547-15c8-4591-945f-f185c31b9660"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.349414 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "35bf1547-15c8-4591-945f-f185c31b9660" (UID: "35bf1547-15c8-4591-945f-f185c31b9660"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.362028 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35bf1547-15c8-4591-945f-f185c31b9660-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "35bf1547-15c8-4591-945f-f185c31b9660" (UID: "35bf1547-15c8-4591-945f-f185c31b9660"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.421750 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn69z\" (UniqueName: \"kubernetes.io/projected/35bf1547-15c8-4591-945f-f185c31b9660-kube-api-access-pn69z\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.421782 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.421792 4978 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/35bf1547-15c8-4591-945f-f185c31b9660-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.421801 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.421809 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35bf1547-15c8-4591-945f-f185c31b9660-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.730067 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-networker-fxwqp" event={"ID":"35bf1547-15c8-4591-945f-f185c31b9660","Type":"ContainerDied","Data":"c1c55638c09a153e8fd72f32fefec221168b1482c4a940e7e069e5c7364497cf"} Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.730115 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1c55638c09a153e8fd72f32fefec221168b1482c4a940e7e069e5c7364497cf" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.730142 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-networker-fxwqp" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.853086 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-wvtwt"] Feb 25 09:15:49 crc kubenswrapper[4978]: E0225 09:15:49.853631 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35bf1547-15c8-4591-945f-f185c31b9660" containerName="ovn-openstack-openstack-networker" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.853655 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="35bf1547-15c8-4591-945f-f185c31b9660" containerName="ovn-openstack-openstack-networker" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.853968 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="35bf1547-15c8-4591-945f-f185c31b9660" containerName="ovn-openstack-openstack-networker" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.855336 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.858651 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-networker-dockercfg-srplt" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.859081 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.859157 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.859395 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-networker" Feb 25 09:15:49 crc kubenswrapper[4978]: I0225 09:15:49.865113 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-wvtwt"] Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.033341 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.033672 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-ssh-key-openstack-networker\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.033725 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-inventory\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.034104 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjq67\" (UniqueName: \"kubernetes.io/projected/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-kube-api-access-fjq67\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.034347 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.034561 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.136533 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjq67\" (UniqueName: \"kubernetes.io/projected/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-kube-api-access-fjq67\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.136695 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.136837 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.137015 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.137111 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-ssh-key-openstack-networker\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.137204 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-inventory\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.142485 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.143301 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.144114 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-ssh-key-openstack-networker\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.145028 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-inventory\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.146679 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.164846 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjq67\" (UniqueName: \"kubernetes.io/projected/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-kube-api-access-fjq67\") pod \"neutron-metadata-openstack-openstack-networker-wvtwt\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.176320 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:15:50 crc kubenswrapper[4978]: I0225 09:15:50.775197 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-networker-wvtwt"] Feb 25 09:15:51 crc kubenswrapper[4978]: I0225 09:15:51.755976 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" event={"ID":"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d","Type":"ContainerStarted","Data":"e5293b7e01b90c56f1865ca0f311c4ff1a5354a014f25558e8a0e6bf8e7ca77e"} Feb 25 09:15:51 crc kubenswrapper[4978]: I0225 09:15:51.756606 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" event={"ID":"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d","Type":"ContainerStarted","Data":"876e6b9e12c8f2d821dbc96b33d2589c8d812823ba8ab440e9029f245db9e57b"} Feb 25 09:15:51 crc kubenswrapper[4978]: I0225 09:15:51.786634 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" podStartSLOduration=2.362393621 podStartE2EDuration="2.786599479s" podCreationTimestamp="2026-02-25 09:15:49 +0000 UTC" firstStartedPulling="2026-02-25 09:15:50.774570076 +0000 UTC m=+9044.213826575" lastFinishedPulling="2026-02-25 09:15:51.198775964 +0000 UTC m=+9044.638032433" observedRunningTime="2026-02-25 09:15:51.775733751 +0000 UTC m=+9045.214990280" watchObservedRunningTime="2026-02-25 09:15:51.786599479 +0000 UTC m=+9045.225856038" Feb 25 09:15:52 crc kubenswrapper[4978]: I0225 09:15:52.775065 4978 generic.go:334] "Generic (PLEG): container finished" podID="b722f89d-c453-4235-8d4c-8d9990f0551b" containerID="49f712a79e9450674b70464faa970d26338dc46a03cc44b7624dd08eccbed959" exitCode=0 Feb 25 09:15:52 crc kubenswrapper[4978]: I0225 09:15:52.775524 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" event={"ID":"b722f89d-c453-4235-8d4c-8d9990f0551b","Type":"ContainerDied","Data":"49f712a79e9450674b70464faa970d26338dc46a03cc44b7624dd08eccbed959"} Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.267304 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.447682 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ovn-combined-ca-bundle\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.447814 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-dhcp-combined-ca-bundle\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.447936 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-libvirt-default-certs-0\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448004 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-telemetry-combined-ca-bundle\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448041 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-sriov-combined-ca-bundle\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448119 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ssh-key-openstack-cell1\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448268 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-nova-combined-ca-bundle\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448308 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-inventory\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448348 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-ovn-default-certs-0\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448414 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t94jj\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-kube-api-access-t94jj\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448455 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-neutron-metadata-default-certs-0\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448584 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-telemetry-default-certs-0\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448627 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-metadata-combined-ca-bundle\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448667 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-libvirt-combined-ca-bundle\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.448708 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-bootstrap-combined-ca-bundle\") pod \"b722f89d-c453-4235-8d4c-8d9990f0551b\" (UID: \"b722f89d-c453-4235-8d4c-8d9990f0551b\") " Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.459800 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.460528 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.460765 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.462757 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-ovn-default-certs-0") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "openstack-cell1-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.463575 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-kube-api-access-t94jj" (OuterVolumeSpecName: "kube-api-access-t94jj") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "kube-api-access-t94jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.463708 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.463822 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.463871 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.464182 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-telemetry-default-certs-0") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "openstack-cell1-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.465898 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-neutron-metadata-default-certs-0") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "openstack-cell1-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.467483 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-libvirt-default-certs-0") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "openstack-cell1-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.483517 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.483667 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.498541 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-inventory" (OuterVolumeSpecName: "inventory") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.502554 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "b722f89d-c453-4235-8d4c-8d9990f0551b" (UID: "b722f89d-c453-4235-8d4c-8d9990f0551b"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.551821 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.551875 4978 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.551895 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.551920 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.551943 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t94jj\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-kube-api-access-t94jj\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.551964 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.551984 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.552002 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.552020 4978 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.552039 4978 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.552057 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.552076 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.552096 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b722f89d-c453-4235-8d4c-8d9990f0551b-openstack-cell1-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.552115 4978 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.552132 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b722f89d-c453-4235-8d4c-8d9990f0551b-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.803176 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" event={"ID":"b722f89d-c453-4235-8d4c-8d9990f0551b","Type":"ContainerDied","Data":"f340685706fa9edf9eeee6091fdf28749d5ca12c1af0463b0da99319180ce75d"} Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.803235 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f340685706fa9edf9eeee6091fdf28749d5ca12c1af0463b0da99319180ce75d" Feb 25 09:15:54 crc kubenswrapper[4978]: I0225 09:15:54.803254 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-7fcwt" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.040989 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-rq28r"] Feb 25 09:15:55 crc kubenswrapper[4978]: E0225 09:15:55.041643 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b722f89d-c453-4235-8d4c-8d9990f0551b" containerName="install-certs-openstack-openstack-cell1" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.041672 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b722f89d-c453-4235-8d4c-8d9990f0551b" containerName="install-certs-openstack-openstack-cell1" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.042083 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b722f89d-c453-4235-8d4c-8d9990f0551b" containerName="install-certs-openstack-openstack-cell1" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.043339 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.049662 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.049951 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.049988 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.057625 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-rq28r"] Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.168843 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.169177 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.169200 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpprl\" (UniqueName: \"kubernetes.io/projected/7fdbbb5e-0829-45c0-8995-42267e75fa07-kube-api-access-xpprl\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.169270 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-inventory\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.169326 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.270974 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.271077 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.271111 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpprl\" (UniqueName: \"kubernetes.io/projected/7fdbbb5e-0829-45c0-8995-42267e75fa07-kube-api-access-xpprl\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.271211 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-inventory\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.271291 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.272832 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.274759 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.286267 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-inventory\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.287099 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.295338 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpprl\" (UniqueName: \"kubernetes.io/projected/7fdbbb5e-0829-45c0-8995-42267e75fa07-kube-api-access-xpprl\") pod \"ovn-openstack-openstack-cell1-rq28r\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.374018 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:15:55 crc kubenswrapper[4978]: I0225 09:15:55.970306 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-rq28r"] Feb 25 09:15:56 crc kubenswrapper[4978]: I0225 09:15:56.824385 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rq28r" event={"ID":"7fdbbb5e-0829-45c0-8995-42267e75fa07","Type":"ContainerStarted","Data":"1cceeb8b6ae049c4c26df4ea65392830886048b3b3302e315c4c8c13e57dbbe1"} Feb 25 09:15:56 crc kubenswrapper[4978]: I0225 09:15:56.824751 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rq28r" event={"ID":"7fdbbb5e-0829-45c0-8995-42267e75fa07","Type":"ContainerStarted","Data":"ed406508e1772509688dfd9c780b6acf8708bde94b2ceaea42d8426b6019dbd8"} Feb 25 09:15:56 crc kubenswrapper[4978]: I0225 09:15:56.855152 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-rq28r" podStartSLOduration=1.316146577 podStartE2EDuration="1.855130459s" podCreationTimestamp="2026-02-25 09:15:55 +0000 UTC" firstStartedPulling="2026-02-25 09:15:55.975093074 +0000 UTC m=+9049.414349533" lastFinishedPulling="2026-02-25 09:15:56.514076936 +0000 UTC m=+9049.953333415" observedRunningTime="2026-02-25 09:15:56.846565563 +0000 UTC m=+9050.285822022" watchObservedRunningTime="2026-02-25 09:15:56.855130459 +0000 UTC m=+9050.294386928" Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.148712 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533516-lnrdv"] Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.151027 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533516-lnrdv" Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.154554 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.154604 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.162340 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.167315 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533516-lnrdv"] Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.286707 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx498\" (UniqueName: \"kubernetes.io/projected/bd4afe50-4f1e-4522-bb21-9fb4208b350e-kube-api-access-dx498\") pod \"auto-csr-approver-29533516-lnrdv\" (UID: \"bd4afe50-4f1e-4522-bb21-9fb4208b350e\") " pod="openshift-infra/auto-csr-approver-29533516-lnrdv" Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.390474 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx498\" (UniqueName: \"kubernetes.io/projected/bd4afe50-4f1e-4522-bb21-9fb4208b350e-kube-api-access-dx498\") pod \"auto-csr-approver-29533516-lnrdv\" (UID: \"bd4afe50-4f1e-4522-bb21-9fb4208b350e\") " pod="openshift-infra/auto-csr-approver-29533516-lnrdv" Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.414638 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx498\" (UniqueName: \"kubernetes.io/projected/bd4afe50-4f1e-4522-bb21-9fb4208b350e-kube-api-access-dx498\") pod \"auto-csr-approver-29533516-lnrdv\" (UID: \"bd4afe50-4f1e-4522-bb21-9fb4208b350e\") " pod="openshift-infra/auto-csr-approver-29533516-lnrdv" Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.478168 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533516-lnrdv" Feb 25 09:16:00 crc kubenswrapper[4978]: I0225 09:16:00.979669 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533516-lnrdv"] Feb 25 09:16:01 crc kubenswrapper[4978]: I0225 09:16:01.889392 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533516-lnrdv" event={"ID":"bd4afe50-4f1e-4522-bb21-9fb4208b350e","Type":"ContainerStarted","Data":"160ea1c9019ce5d8db357cc4bb364014e09f3b958899b8ac8f5ceb3426099473"} Feb 25 09:16:02 crc kubenswrapper[4978]: I0225 09:16:02.906325 4978 generic.go:334] "Generic (PLEG): container finished" podID="bd4afe50-4f1e-4522-bb21-9fb4208b350e" containerID="3818c5f3d610ca0d306ccdc828b4f4567c653ff242893ba9092078dac3838fff" exitCode=0 Feb 25 09:16:02 crc kubenswrapper[4978]: I0225 09:16:02.906450 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533516-lnrdv" event={"ID":"bd4afe50-4f1e-4522-bb21-9fb4208b350e","Type":"ContainerDied","Data":"3818c5f3d610ca0d306ccdc828b4f4567c653ff242893ba9092078dac3838fff"} Feb 25 09:16:04 crc kubenswrapper[4978]: I0225 09:16:04.394504 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533516-lnrdv" Feb 25 09:16:04 crc kubenswrapper[4978]: I0225 09:16:04.585990 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx498\" (UniqueName: \"kubernetes.io/projected/bd4afe50-4f1e-4522-bb21-9fb4208b350e-kube-api-access-dx498\") pod \"bd4afe50-4f1e-4522-bb21-9fb4208b350e\" (UID: \"bd4afe50-4f1e-4522-bb21-9fb4208b350e\") " Feb 25 09:16:04 crc kubenswrapper[4978]: I0225 09:16:04.598686 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd4afe50-4f1e-4522-bb21-9fb4208b350e-kube-api-access-dx498" (OuterVolumeSpecName: "kube-api-access-dx498") pod "bd4afe50-4f1e-4522-bb21-9fb4208b350e" (UID: "bd4afe50-4f1e-4522-bb21-9fb4208b350e"). InnerVolumeSpecName "kube-api-access-dx498". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:16:04 crc kubenswrapper[4978]: I0225 09:16:04.689545 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx498\" (UniqueName: \"kubernetes.io/projected/bd4afe50-4f1e-4522-bb21-9fb4208b350e-kube-api-access-dx498\") on node \"crc\" DevicePath \"\"" Feb 25 09:16:04 crc kubenswrapper[4978]: I0225 09:16:04.955019 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533516-lnrdv" event={"ID":"bd4afe50-4f1e-4522-bb21-9fb4208b350e","Type":"ContainerDied","Data":"160ea1c9019ce5d8db357cc4bb364014e09f3b958899b8ac8f5ceb3426099473"} Feb 25 09:16:04 crc kubenswrapper[4978]: I0225 09:16:04.955084 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="160ea1c9019ce5d8db357cc4bb364014e09f3b958899b8ac8f5ceb3426099473" Feb 25 09:16:04 crc kubenswrapper[4978]: I0225 09:16:04.955155 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533516-lnrdv" Feb 25 09:16:05 crc kubenswrapper[4978]: I0225 09:16:05.476933 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533510-9jj6q"] Feb 25 09:16:05 crc kubenswrapper[4978]: I0225 09:16:05.484935 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533510-9jj6q"] Feb 25 09:16:07 crc kubenswrapper[4978]: I0225 09:16:07.348531 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37ae8605-e2a4-41dd-b5bc-5275301ed3be" path="/var/lib/kubelet/pods/37ae8605-e2a4-41dd-b5bc-5275301ed3be/volumes" Feb 25 09:16:36 crc kubenswrapper[4978]: I0225 09:16:36.143664 4978 scope.go:117] "RemoveContainer" containerID="aee1214b39366db4378d9a34da91be3f704cd7816af2f78422e056aab38d3a27" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.528057 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k6jln"] Feb 25 09:16:45 crc kubenswrapper[4978]: E0225 09:16:45.529179 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd4afe50-4f1e-4522-bb21-9fb4208b350e" containerName="oc" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.529196 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd4afe50-4f1e-4522-bb21-9fb4208b350e" containerName="oc" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.529617 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd4afe50-4f1e-4522-bb21-9fb4208b350e" containerName="oc" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.532241 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.559597 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k6jln"] Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.668753 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h99df\" (UniqueName: \"kubernetes.io/projected/56b56675-83a1-4e7d-906f-aa1dffe4ce97-kube-api-access-h99df\") pod \"redhat-operators-k6jln\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.668996 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-utilities\") pod \"redhat-operators-k6jln\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.669172 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-catalog-content\") pod \"redhat-operators-k6jln\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.770673 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-utilities\") pod \"redhat-operators-k6jln\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.770819 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-catalog-content\") pod \"redhat-operators-k6jln\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.770917 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h99df\" (UniqueName: \"kubernetes.io/projected/56b56675-83a1-4e7d-906f-aa1dffe4ce97-kube-api-access-h99df\") pod \"redhat-operators-k6jln\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.771285 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-utilities\") pod \"redhat-operators-k6jln\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.771356 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-catalog-content\") pod \"redhat-operators-k6jln\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.793211 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h99df\" (UniqueName: \"kubernetes.io/projected/56b56675-83a1-4e7d-906f-aa1dffe4ce97-kube-api-access-h99df\") pod \"redhat-operators-k6jln\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:45 crc kubenswrapper[4978]: I0225 09:16:45.850790 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:46 crc kubenswrapper[4978]: I0225 09:16:46.356510 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k6jln"] Feb 25 09:16:46 crc kubenswrapper[4978]: W0225 09:16:46.357406 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56b56675_83a1_4e7d_906f_aa1dffe4ce97.slice/crio-5b9b153cc923cc04a1f12fe8a3834e76b8aca58e1ae1dc9a07ef767bb8213e44 WatchSource:0}: Error finding container 5b9b153cc923cc04a1f12fe8a3834e76b8aca58e1ae1dc9a07ef767bb8213e44: Status 404 returned error can't find the container with id 5b9b153cc923cc04a1f12fe8a3834e76b8aca58e1ae1dc9a07ef767bb8213e44 Feb 25 09:16:46 crc kubenswrapper[4978]: I0225 09:16:46.456231 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6jln" event={"ID":"56b56675-83a1-4e7d-906f-aa1dffe4ce97","Type":"ContainerStarted","Data":"5b9b153cc923cc04a1f12fe8a3834e76b8aca58e1ae1dc9a07ef767bb8213e44"} Feb 25 09:16:46 crc kubenswrapper[4978]: I0225 09:16:46.540488 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:16:46 crc kubenswrapper[4978]: I0225 09:16:46.540551 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:16:47 crc kubenswrapper[4978]: I0225 09:16:47.468462 4978 generic.go:334] "Generic (PLEG): container finished" podID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerID="6f9cd96ff28ac8215611e077e03323f6ef74739fb249be2b4219a84991cb9fd4" exitCode=0 Feb 25 09:16:47 crc kubenswrapper[4978]: I0225 09:16:47.468785 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6jln" event={"ID":"56b56675-83a1-4e7d-906f-aa1dffe4ce97","Type":"ContainerDied","Data":"6f9cd96ff28ac8215611e077e03323f6ef74739fb249be2b4219a84991cb9fd4"} Feb 25 09:16:47 crc kubenswrapper[4978]: I0225 09:16:47.471412 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:16:48 crc kubenswrapper[4978]: I0225 09:16:48.480089 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6jln" event={"ID":"56b56675-83a1-4e7d-906f-aa1dffe4ce97","Type":"ContainerStarted","Data":"a25c7ce3712e6237c92de30c69539954d10022aabae4d2e2f433a4211ad0fe5f"} Feb 25 09:16:49 crc kubenswrapper[4978]: I0225 09:16:49.494530 4978 generic.go:334] "Generic (PLEG): container finished" podID="29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d" containerID="e5293b7e01b90c56f1865ca0f311c4ff1a5354a014f25558e8a0e6bf8e7ca77e" exitCode=0 Feb 25 09:16:49 crc kubenswrapper[4978]: I0225 09:16:49.494596 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" event={"ID":"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d","Type":"ContainerDied","Data":"e5293b7e01b90c56f1865ca0f311c4ff1a5354a014f25558e8a0e6bf8e7ca77e"} Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.125522 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.198160 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-ovn-metadata-agent-neutron-config-0\") pod \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.198244 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-ssh-key-openstack-networker\") pod \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.198390 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-metadata-combined-ca-bundle\") pod \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.198563 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-nova-metadata-neutron-config-0\") pod \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.198645 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-inventory\") pod \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.198744 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjq67\" (UniqueName: \"kubernetes.io/projected/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-kube-api-access-fjq67\") pod \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\" (UID: \"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d\") " Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.207157 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-kube-api-access-fjq67" (OuterVolumeSpecName: "kube-api-access-fjq67") pod "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d" (UID: "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d"). InnerVolumeSpecName "kube-api-access-fjq67". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.208047 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d" (UID: "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.242607 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-ssh-key-openstack-networker" (OuterVolumeSpecName: "ssh-key-openstack-networker") pod "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d" (UID: "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d"). InnerVolumeSpecName "ssh-key-openstack-networker". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.248745 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d" (UID: "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.253229 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d" (UID: "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.254611 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-inventory" (OuterVolumeSpecName: "inventory") pod "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d" (UID: "29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.300959 4978 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.301000 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.301015 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjq67\" (UniqueName: \"kubernetes.io/projected/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-kube-api-access-fjq67\") on node \"crc\" DevicePath \"\"" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.301031 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.301043 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-networker\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-ssh-key-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.301056 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.520953 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" event={"ID":"29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d","Type":"ContainerDied","Data":"876e6b9e12c8f2d821dbc96b33d2589c8d812823ba8ab440e9029f245db9e57b"} Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.520974 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-networker-wvtwt" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.520988 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="876e6b9e12c8f2d821dbc96b33d2589c8d812823ba8ab440e9029f245db9e57b" Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.524436 4978 generic.go:334] "Generic (PLEG): container finished" podID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerID="a25c7ce3712e6237c92de30c69539954d10022aabae4d2e2f433a4211ad0fe5f" exitCode=0 Feb 25 09:16:51 crc kubenswrapper[4978]: I0225 09:16:51.524463 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6jln" event={"ID":"56b56675-83a1-4e7d-906f-aa1dffe4ce97","Type":"ContainerDied","Data":"a25c7ce3712e6237c92de30c69539954d10022aabae4d2e2f433a4211ad0fe5f"} Feb 25 09:16:52 crc kubenswrapper[4978]: I0225 09:16:52.537847 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6jln" event={"ID":"56b56675-83a1-4e7d-906f-aa1dffe4ce97","Type":"ContainerStarted","Data":"7ebdb4db4ffd4a468c3f0a9d653f7bb02a06a35410041d0f9245f3bea129cdb9"} Feb 25 09:16:52 crc kubenswrapper[4978]: I0225 09:16:52.571131 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k6jln" podStartSLOduration=2.872593482 podStartE2EDuration="7.571106583s" podCreationTimestamp="2026-02-25 09:16:45 +0000 UTC" firstStartedPulling="2026-02-25 09:16:47.471059746 +0000 UTC m=+9100.910316215" lastFinishedPulling="2026-02-25 09:16:52.169572847 +0000 UTC m=+9105.608829316" observedRunningTime="2026-02-25 09:16:52.558242044 +0000 UTC m=+9105.997498523" watchObservedRunningTime="2026-02-25 09:16:52.571106583 +0000 UTC m=+9106.010363052" Feb 25 09:16:55 crc kubenswrapper[4978]: I0225 09:16:55.852639 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:55 crc kubenswrapper[4978]: I0225 09:16:55.853619 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:16:56 crc kubenswrapper[4978]: I0225 09:16:56.910989 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k6jln" podUID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerName="registry-server" probeResult="failure" output=< Feb 25 09:16:56 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 09:16:56 crc kubenswrapper[4978]: > Feb 25 09:17:05 crc kubenswrapper[4978]: I0225 09:17:05.918141 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:17:05 crc kubenswrapper[4978]: I0225 09:17:05.968960 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:17:06 crc kubenswrapper[4978]: I0225 09:17:06.156905 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k6jln"] Feb 25 09:17:07 crc kubenswrapper[4978]: I0225 09:17:07.719213 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k6jln" podUID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerName="registry-server" containerID="cri-o://7ebdb4db4ffd4a468c3f0a9d653f7bb02a06a35410041d0f9245f3bea129cdb9" gracePeriod=2 Feb 25 09:17:08 crc kubenswrapper[4978]: I0225 09:17:08.737007 4978 generic.go:334] "Generic (PLEG): container finished" podID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerID="7ebdb4db4ffd4a468c3f0a9d653f7bb02a06a35410041d0f9245f3bea129cdb9" exitCode=0 Feb 25 09:17:08 crc kubenswrapper[4978]: I0225 09:17:08.737091 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6jln" event={"ID":"56b56675-83a1-4e7d-906f-aa1dffe4ce97","Type":"ContainerDied","Data":"7ebdb4db4ffd4a468c3f0a9d653f7bb02a06a35410041d0f9245f3bea129cdb9"} Feb 25 09:17:08 crc kubenswrapper[4978]: I0225 09:17:08.927534 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.064561 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-catalog-content\") pod \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.064735 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-utilities\") pod \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.064896 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h99df\" (UniqueName: \"kubernetes.io/projected/56b56675-83a1-4e7d-906f-aa1dffe4ce97-kube-api-access-h99df\") pod \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\" (UID: \"56b56675-83a1-4e7d-906f-aa1dffe4ce97\") " Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.076294 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-utilities" (OuterVolumeSpecName: "utilities") pod "56b56675-83a1-4e7d-906f-aa1dffe4ce97" (UID: "56b56675-83a1-4e7d-906f-aa1dffe4ce97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.083824 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56b56675-83a1-4e7d-906f-aa1dffe4ce97-kube-api-access-h99df" (OuterVolumeSpecName: "kube-api-access-h99df") pod "56b56675-83a1-4e7d-906f-aa1dffe4ce97" (UID: "56b56675-83a1-4e7d-906f-aa1dffe4ce97"). InnerVolumeSpecName "kube-api-access-h99df". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.167253 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h99df\" (UniqueName: \"kubernetes.io/projected/56b56675-83a1-4e7d-906f-aa1dffe4ce97-kube-api-access-h99df\") on node \"crc\" DevicePath \"\"" Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.167285 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.223762 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56b56675-83a1-4e7d-906f-aa1dffe4ce97" (UID: "56b56675-83a1-4e7d-906f-aa1dffe4ce97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.270350 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56b56675-83a1-4e7d-906f-aa1dffe4ce97-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.751428 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k6jln" event={"ID":"56b56675-83a1-4e7d-906f-aa1dffe4ce97","Type":"ContainerDied","Data":"5b9b153cc923cc04a1f12fe8a3834e76b8aca58e1ae1dc9a07ef767bb8213e44"} Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.751479 4978 scope.go:117] "RemoveContainer" containerID="7ebdb4db4ffd4a468c3f0a9d653f7bb02a06a35410041d0f9245f3bea129cdb9" Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.751546 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k6jln" Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.785461 4978 scope.go:117] "RemoveContainer" containerID="a25c7ce3712e6237c92de30c69539954d10022aabae4d2e2f433a4211ad0fe5f" Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.791414 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k6jln"] Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.805297 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k6jln"] Feb 25 09:17:09 crc kubenswrapper[4978]: I0225 09:17:09.820068 4978 scope.go:117] "RemoveContainer" containerID="6f9cd96ff28ac8215611e077e03323f6ef74739fb249be2b4219a84991cb9fd4" Feb 25 09:17:11 crc kubenswrapper[4978]: I0225 09:17:11.343739 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" path="/var/lib/kubelet/pods/56b56675-83a1-4e7d-906f-aa1dffe4ce97/volumes" Feb 25 09:17:11 crc kubenswrapper[4978]: I0225 09:17:11.778990 4978 generic.go:334] "Generic (PLEG): container finished" podID="7fdbbb5e-0829-45c0-8995-42267e75fa07" containerID="1cceeb8b6ae049c4c26df4ea65392830886048b3b3302e315c4c8c13e57dbbe1" exitCode=0 Feb 25 09:17:11 crc kubenswrapper[4978]: I0225 09:17:11.779061 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rq28r" event={"ID":"7fdbbb5e-0829-45c0-8995-42267e75fa07","Type":"ContainerDied","Data":"1cceeb8b6ae049c4c26df4ea65392830886048b3b3302e315c4c8c13e57dbbe1"} Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.316498 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.477337 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ssh-key-openstack-cell1\") pod \"7fdbbb5e-0829-45c0-8995-42267e75fa07\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.477768 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-inventory\") pod \"7fdbbb5e-0829-45c0-8995-42267e75fa07\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.477850 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovn-combined-ca-bundle\") pod \"7fdbbb5e-0829-45c0-8995-42267e75fa07\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.477937 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpprl\" (UniqueName: \"kubernetes.io/projected/7fdbbb5e-0829-45c0-8995-42267e75fa07-kube-api-access-xpprl\") pod \"7fdbbb5e-0829-45c0-8995-42267e75fa07\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.478213 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovncontroller-config-0\") pod \"7fdbbb5e-0829-45c0-8995-42267e75fa07\" (UID: \"7fdbbb5e-0829-45c0-8995-42267e75fa07\") " Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.483800 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fdbbb5e-0829-45c0-8995-42267e75fa07-kube-api-access-xpprl" (OuterVolumeSpecName: "kube-api-access-xpprl") pod "7fdbbb5e-0829-45c0-8995-42267e75fa07" (UID: "7fdbbb5e-0829-45c0-8995-42267e75fa07"). InnerVolumeSpecName "kube-api-access-xpprl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.487969 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpprl\" (UniqueName: \"kubernetes.io/projected/7fdbbb5e-0829-45c0-8995-42267e75fa07-kube-api-access-xpprl\") on node \"crc\" DevicePath \"\"" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.501933 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "7fdbbb5e-0829-45c0-8995-42267e75fa07" (UID: "7fdbbb5e-0829-45c0-8995-42267e75fa07"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.505993 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "7fdbbb5e-0829-45c0-8995-42267e75fa07" (UID: "7fdbbb5e-0829-45c0-8995-42267e75fa07"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.509041 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-inventory" (OuterVolumeSpecName: "inventory") pod "7fdbbb5e-0829-45c0-8995-42267e75fa07" (UID: "7fdbbb5e-0829-45c0-8995-42267e75fa07"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.513658 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "7fdbbb5e-0829-45c0-8995-42267e75fa07" (UID: "7fdbbb5e-0829-45c0-8995-42267e75fa07"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.590734 4978 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.590794 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.590818 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.590837 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fdbbb5e-0829-45c0-8995-42267e75fa07-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.811556 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-rq28r" event={"ID":"7fdbbb5e-0829-45c0-8995-42267e75fa07","Type":"ContainerDied","Data":"ed406508e1772509688dfd9c780b6acf8708bde94b2ceaea42d8426b6019dbd8"} Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.811628 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed406508e1772509688dfd9c780b6acf8708bde94b2ceaea42d8426b6019dbd8" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.811663 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-rq28r" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.951877 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xs7rz"] Feb 25 09:17:13 crc kubenswrapper[4978]: E0225 09:17:13.952470 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerName="extract-utilities" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.952496 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerName="extract-utilities" Feb 25 09:17:13 crc kubenswrapper[4978]: E0225 09:17:13.952513 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerName="registry-server" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.952522 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerName="registry-server" Feb 25 09:17:13 crc kubenswrapper[4978]: E0225 09:17:13.952543 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdbbb5e-0829-45c0-8995-42267e75fa07" containerName="ovn-openstack-openstack-cell1" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.952556 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdbbb5e-0829-45c0-8995-42267e75fa07" containerName="ovn-openstack-openstack-cell1" Feb 25 09:17:13 crc kubenswrapper[4978]: E0225 09:17:13.952592 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerName="extract-content" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.952606 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerName="extract-content" Feb 25 09:17:13 crc kubenswrapper[4978]: E0225 09:17:13.952657 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d" containerName="neutron-metadata-openstack-openstack-networker" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.952670 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d" containerName="neutron-metadata-openstack-openstack-networker" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.952911 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fdbbb5e-0829-45c0-8995-42267e75fa07" containerName="ovn-openstack-openstack-cell1" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.952933 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d" containerName="neutron-metadata-openstack-openstack-networker" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.952973 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="56b56675-83a1-4e7d-906f-aa1dffe4ce97" containerName="registry-server" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.953886 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.961722 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.961974 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.962244 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.962437 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.962774 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.964477 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xs7rz"] Feb 25 09:17:13 crc kubenswrapper[4978]: I0225 09:17:13.968833 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.001192 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.001321 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.001423 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.001462 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p8fs\" (UniqueName: \"kubernetes.io/projected/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-kube-api-access-4p8fs\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.001503 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.001565 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.103345 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.103486 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.103563 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.103616 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.103640 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p8fs\" (UniqueName: \"kubernetes.io/projected/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-kube-api-access-4p8fs\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.103660 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.107481 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.119771 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.119979 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.119988 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.120240 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.128511 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p8fs\" (UniqueName: \"kubernetes.io/projected/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-kube-api-access-4p8fs\") pod \"neutron-metadata-openstack-openstack-cell1-xs7rz\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.288697 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:17:14 crc kubenswrapper[4978]: I0225 09:17:14.911548 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xs7rz"] Feb 25 09:17:15 crc kubenswrapper[4978]: I0225 09:17:15.842430 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" event={"ID":"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a","Type":"ContainerStarted","Data":"ada10853d988c6922272202b3843b275a7373624cb44eb0e430e25ba11f4066d"} Feb 25 09:17:15 crc kubenswrapper[4978]: I0225 09:17:15.842719 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" event={"ID":"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a","Type":"ContainerStarted","Data":"4fbe3d4c94cb419e5bdfd647fd712850ffa375c448b225d65fe969adb2dae2e2"} Feb 25 09:17:15 crc kubenswrapper[4978]: I0225 09:17:15.871867 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" podStartSLOduration=2.419986522 podStartE2EDuration="2.871841017s" podCreationTimestamp="2026-02-25 09:17:13 +0000 UTC" firstStartedPulling="2026-02-25 09:17:14.920306437 +0000 UTC m=+9128.359562886" lastFinishedPulling="2026-02-25 09:17:15.372160932 +0000 UTC m=+9128.811417381" observedRunningTime="2026-02-25 09:17:15.862851148 +0000 UTC m=+9129.302107627" watchObservedRunningTime="2026-02-25 09:17:15.871841017 +0000 UTC m=+9129.311097516" Feb 25 09:17:16 crc kubenswrapper[4978]: I0225 09:17:16.546110 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:17:16 crc kubenswrapper[4978]: I0225 09:17:16.546622 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:17:46 crc kubenswrapper[4978]: I0225 09:17:46.540135 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:17:46 crc kubenswrapper[4978]: I0225 09:17:46.540808 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:17:46 crc kubenswrapper[4978]: I0225 09:17:46.540866 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 09:17:46 crc kubenswrapper[4978]: I0225 09:17:46.541750 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 09:17:46 crc kubenswrapper[4978]: I0225 09:17:46.541809 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" gracePeriod=600 Feb 25 09:17:46 crc kubenswrapper[4978]: E0225 09:17:46.667048 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:17:47 crc kubenswrapper[4978]: I0225 09:17:47.265189 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" exitCode=0 Feb 25 09:17:47 crc kubenswrapper[4978]: I0225 09:17:47.265296 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4"} Feb 25 09:17:47 crc kubenswrapper[4978]: I0225 09:17:47.265574 4978 scope.go:117] "RemoveContainer" containerID="6e09bc70e0ad7207309daaa14091dbcb18dfd0d3b48bf2f83a1437dfb4535285" Feb 25 09:17:47 crc kubenswrapper[4978]: I0225 09:17:47.266542 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:17:47 crc kubenswrapper[4978]: E0225 09:17:47.267020 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:17:58 crc kubenswrapper[4978]: I0225 09:17:58.329017 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:17:58 crc kubenswrapper[4978]: E0225 09:17:58.330251 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.176513 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533518-ph8zb"] Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.178547 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533518-ph8zb" Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.182337 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.182501 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.182519 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.197278 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533518-ph8zb"] Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.329009 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m46zl\" (UniqueName: \"kubernetes.io/projected/ecd6d99e-ee38-47dc-887b-d5ed8b2f4588-kube-api-access-m46zl\") pod \"auto-csr-approver-29533518-ph8zb\" (UID: \"ecd6d99e-ee38-47dc-887b-d5ed8b2f4588\") " pod="openshift-infra/auto-csr-approver-29533518-ph8zb" Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.431663 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m46zl\" (UniqueName: \"kubernetes.io/projected/ecd6d99e-ee38-47dc-887b-d5ed8b2f4588-kube-api-access-m46zl\") pod \"auto-csr-approver-29533518-ph8zb\" (UID: \"ecd6d99e-ee38-47dc-887b-d5ed8b2f4588\") " pod="openshift-infra/auto-csr-approver-29533518-ph8zb" Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.454425 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m46zl\" (UniqueName: \"kubernetes.io/projected/ecd6d99e-ee38-47dc-887b-d5ed8b2f4588-kube-api-access-m46zl\") pod \"auto-csr-approver-29533518-ph8zb\" (UID: \"ecd6d99e-ee38-47dc-887b-d5ed8b2f4588\") " pod="openshift-infra/auto-csr-approver-29533518-ph8zb" Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.514517 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533518-ph8zb" Feb 25 09:18:00 crc kubenswrapper[4978]: I0225 09:18:00.839499 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533518-ph8zb"] Feb 25 09:18:00 crc kubenswrapper[4978]: W0225 09:18:00.854903 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecd6d99e_ee38_47dc_887b_d5ed8b2f4588.slice/crio-3973509a29d0817072c421c40d0f89881fb816f8a4cb09d25f02b077033fe49b WatchSource:0}: Error finding container 3973509a29d0817072c421c40d0f89881fb816f8a4cb09d25f02b077033fe49b: Status 404 returned error can't find the container with id 3973509a29d0817072c421c40d0f89881fb816f8a4cb09d25f02b077033fe49b Feb 25 09:18:01 crc kubenswrapper[4978]: I0225 09:18:01.431178 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533518-ph8zb" event={"ID":"ecd6d99e-ee38-47dc-887b-d5ed8b2f4588","Type":"ContainerStarted","Data":"3973509a29d0817072c421c40d0f89881fb816f8a4cb09d25f02b077033fe49b"} Feb 25 09:18:02 crc kubenswrapper[4978]: I0225 09:18:02.450996 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533518-ph8zb" event={"ID":"ecd6d99e-ee38-47dc-887b-d5ed8b2f4588","Type":"ContainerStarted","Data":"4ad66268b89d07953efd824b688873af6a7756a873a29b93e8c908a4dd839adc"} Feb 25 09:18:02 crc kubenswrapper[4978]: I0225 09:18:02.471646 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533518-ph8zb" podStartSLOduration=1.626721077 podStartE2EDuration="2.471626894s" podCreationTimestamp="2026-02-25 09:18:00 +0000 UTC" firstStartedPulling="2026-02-25 09:18:00.860813708 +0000 UTC m=+9174.300070167" lastFinishedPulling="2026-02-25 09:18:01.705719525 +0000 UTC m=+9175.144975984" observedRunningTime="2026-02-25 09:18:02.468332032 +0000 UTC m=+9175.907588491" watchObservedRunningTime="2026-02-25 09:18:02.471626894 +0000 UTC m=+9175.910883353" Feb 25 09:18:03 crc kubenswrapper[4978]: I0225 09:18:03.462945 4978 generic.go:334] "Generic (PLEG): container finished" podID="ecd6d99e-ee38-47dc-887b-d5ed8b2f4588" containerID="4ad66268b89d07953efd824b688873af6a7756a873a29b93e8c908a4dd839adc" exitCode=0 Feb 25 09:18:03 crc kubenswrapper[4978]: I0225 09:18:03.463032 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533518-ph8zb" event={"ID":"ecd6d99e-ee38-47dc-887b-d5ed8b2f4588","Type":"ContainerDied","Data":"4ad66268b89d07953efd824b688873af6a7756a873a29b93e8c908a4dd839adc"} Feb 25 09:18:04 crc kubenswrapper[4978]: I0225 09:18:04.858025 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533518-ph8zb" Feb 25 09:18:05 crc kubenswrapper[4978]: I0225 09:18:05.035749 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m46zl\" (UniqueName: \"kubernetes.io/projected/ecd6d99e-ee38-47dc-887b-d5ed8b2f4588-kube-api-access-m46zl\") pod \"ecd6d99e-ee38-47dc-887b-d5ed8b2f4588\" (UID: \"ecd6d99e-ee38-47dc-887b-d5ed8b2f4588\") " Feb 25 09:18:05 crc kubenswrapper[4978]: I0225 09:18:05.040858 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecd6d99e-ee38-47dc-887b-d5ed8b2f4588-kube-api-access-m46zl" (OuterVolumeSpecName: "kube-api-access-m46zl") pod "ecd6d99e-ee38-47dc-887b-d5ed8b2f4588" (UID: "ecd6d99e-ee38-47dc-887b-d5ed8b2f4588"). InnerVolumeSpecName "kube-api-access-m46zl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:18:05 crc kubenswrapper[4978]: I0225 09:18:05.137981 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m46zl\" (UniqueName: \"kubernetes.io/projected/ecd6d99e-ee38-47dc-887b-d5ed8b2f4588-kube-api-access-m46zl\") on node \"crc\" DevicePath \"\"" Feb 25 09:18:05 crc kubenswrapper[4978]: I0225 09:18:05.483084 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533518-ph8zb" event={"ID":"ecd6d99e-ee38-47dc-887b-d5ed8b2f4588","Type":"ContainerDied","Data":"3973509a29d0817072c421c40d0f89881fb816f8a4cb09d25f02b077033fe49b"} Feb 25 09:18:05 crc kubenswrapper[4978]: I0225 09:18:05.483441 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3973509a29d0817072c421c40d0f89881fb816f8a4cb09d25f02b077033fe49b" Feb 25 09:18:05 crc kubenswrapper[4978]: I0225 09:18:05.483183 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533518-ph8zb" Feb 25 09:18:05 crc kubenswrapper[4978]: I0225 09:18:05.549619 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533512-8j5tw"] Feb 25 09:18:05 crc kubenswrapper[4978]: I0225 09:18:05.563242 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533512-8j5tw"] Feb 25 09:18:07 crc kubenswrapper[4978]: I0225 09:18:07.346259 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fc8fde2-869b-4730-85c5-e1d6250e5edc" path="/var/lib/kubelet/pods/5fc8fde2-869b-4730-85c5-e1d6250e5edc/volumes" Feb 25 09:18:09 crc kubenswrapper[4978]: I0225 09:18:09.327712 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:18:09 crc kubenswrapper[4978]: E0225 09:18:09.328231 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:18:18 crc kubenswrapper[4978]: I0225 09:18:18.683798 4978 generic.go:334] "Generic (PLEG): container finished" podID="0d691a60-c2ae-48a6-8d7d-b8ea03d5178a" containerID="ada10853d988c6922272202b3843b275a7373624cb44eb0e430e25ba11f4066d" exitCode=0 Feb 25 09:18:18 crc kubenswrapper[4978]: I0225 09:18:18.683928 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" event={"ID":"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a","Type":"ContainerDied","Data":"ada10853d988c6922272202b3843b275a7373624cb44eb0e430e25ba11f4066d"} Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.397493 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.490364 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.490465 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-nova-metadata-neutron-config-0\") pod \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.490510 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p8fs\" (UniqueName: \"kubernetes.io/projected/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-kube-api-access-4p8fs\") pod \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.490618 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-inventory\") pod \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.490641 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-ssh-key-openstack-cell1\") pod \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.490760 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-metadata-combined-ca-bundle\") pod \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\" (UID: \"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a\") " Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.495705 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-kube-api-access-4p8fs" (OuterVolumeSpecName: "kube-api-access-4p8fs") pod "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a" (UID: "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a"). InnerVolumeSpecName "kube-api-access-4p8fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.498260 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a" (UID: "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.521057 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a" (UID: "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.526816 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-inventory" (OuterVolumeSpecName: "inventory") pod "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a" (UID: "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.528694 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a" (UID: "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.550027 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a" (UID: "0d691a60-c2ae-48a6-8d7d-b8ea03d5178a"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.595026 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.595070 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.595086 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.595100 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.595115 4978 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.595128 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p8fs\" (UniqueName: \"kubernetes.io/projected/0d691a60-c2ae-48a6-8d7d-b8ea03d5178a-kube-api-access-4p8fs\") on node \"crc\" DevicePath \"\"" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.729654 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" event={"ID":"0d691a60-c2ae-48a6-8d7d-b8ea03d5178a","Type":"ContainerDied","Data":"4fbe3d4c94cb419e5bdfd647fd712850ffa375c448b225d65fe969adb2dae2e2"} Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.729713 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fbe3d4c94cb419e5bdfd647fd712850ffa375c448b225d65fe969adb2dae2e2" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.729731 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xs7rz" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.849015 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-2cksr"] Feb 25 09:18:20 crc kubenswrapper[4978]: E0225 09:18:20.849787 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d691a60-c2ae-48a6-8d7d-b8ea03d5178a" containerName="neutron-metadata-openstack-openstack-cell1" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.849808 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d691a60-c2ae-48a6-8d7d-b8ea03d5178a" containerName="neutron-metadata-openstack-openstack-cell1" Feb 25 09:18:20 crc kubenswrapper[4978]: E0225 09:18:20.849823 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd6d99e-ee38-47dc-887b-d5ed8b2f4588" containerName="oc" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.849830 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd6d99e-ee38-47dc-887b-d5ed8b2f4588" containerName="oc" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.850045 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d691a60-c2ae-48a6-8d7d-b8ea03d5178a" containerName="neutron-metadata-openstack-openstack-cell1" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.850068 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecd6d99e-ee38-47dc-887b-d5ed8b2f4588" containerName="oc" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.850922 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.854506 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.854693 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.856274 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.856539 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.856589 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 09:18:20 crc kubenswrapper[4978]: I0225 09:18:20.864124 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-2cksr"] Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.004114 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.004210 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.004774 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2pc6\" (UniqueName: \"kubernetes.io/projected/41f59382-722c-42ab-89f1-0e738135edcd-kube-api-access-q2pc6\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.004845 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-inventory\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.004909 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.106899 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2pc6\" (UniqueName: \"kubernetes.io/projected/41f59382-722c-42ab-89f1-0e738135edcd-kube-api-access-q2pc6\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.106973 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-inventory\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.107001 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.107056 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.107111 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.112019 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-inventory\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.112810 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.114418 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.114799 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.126702 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2pc6\" (UniqueName: \"kubernetes.io/projected/41f59382-722c-42ab-89f1-0e738135edcd-kube-api-access-q2pc6\") pod \"libvirt-openstack-openstack-cell1-2cksr\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.173061 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:18:21 crc kubenswrapper[4978]: I0225 09:18:21.799231 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-2cksr"] Feb 25 09:18:22 crc kubenswrapper[4978]: I0225 09:18:22.760194 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2cksr" event={"ID":"41f59382-722c-42ab-89f1-0e738135edcd","Type":"ContainerStarted","Data":"db3dc0fcf3282373845b13bc106f325899d230631191085bc6c0bc34d1733313"} Feb 25 09:18:22 crc kubenswrapper[4978]: I0225 09:18:22.760238 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2cksr" event={"ID":"41f59382-722c-42ab-89f1-0e738135edcd","Type":"ContainerStarted","Data":"aa5a14c828aa28f757396831f12cc91c28106b8fb4072151f6d2aae1cd99ab6a"} Feb 25 09:18:22 crc kubenswrapper[4978]: I0225 09:18:22.787868 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-2cksr" podStartSLOduration=2.318861266 podStartE2EDuration="2.787848271s" podCreationTimestamp="2026-02-25 09:18:20 +0000 UTC" firstStartedPulling="2026-02-25 09:18:21.802702831 +0000 UTC m=+9195.241959310" lastFinishedPulling="2026-02-25 09:18:22.271689846 +0000 UTC m=+9195.710946315" observedRunningTime="2026-02-25 09:18:22.777562802 +0000 UTC m=+9196.216819271" watchObservedRunningTime="2026-02-25 09:18:22.787848271 +0000 UTC m=+9196.227104730" Feb 25 09:18:23 crc kubenswrapper[4978]: I0225 09:18:23.328470 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:18:23 crc kubenswrapper[4978]: E0225 09:18:23.329056 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:18:35 crc kubenswrapper[4978]: I0225 09:18:35.329211 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:18:35 crc kubenswrapper[4978]: E0225 09:18:35.329958 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:18:36 crc kubenswrapper[4978]: I0225 09:18:36.311232 4978 scope.go:117] "RemoveContainer" containerID="6702ec1416d897a3573139b7470967f7a191ec9ca2551c45103df6f8ef6ddfca" Feb 25 09:18:50 crc kubenswrapper[4978]: I0225 09:18:50.328780 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:18:50 crc kubenswrapper[4978]: E0225 09:18:50.329721 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.254069 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f5htw"] Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.256587 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.281145 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f5htw"] Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.357751 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc5hk\" (UniqueName: \"kubernetes.io/projected/2d352777-d907-4a43-8baf-e263e9fca045-kube-api-access-nc5hk\") pod \"certified-operators-f5htw\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.357950 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-catalog-content\") pod \"certified-operators-f5htw\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.358000 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-utilities\") pod \"certified-operators-f5htw\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.459409 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-utilities\") pod \"certified-operators-f5htw\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.459645 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc5hk\" (UniqueName: \"kubernetes.io/projected/2d352777-d907-4a43-8baf-e263e9fca045-kube-api-access-nc5hk\") pod \"certified-operators-f5htw\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.459946 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-catalog-content\") pod \"certified-operators-f5htw\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.461247 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-utilities\") pod \"certified-operators-f5htw\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.462116 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-catalog-content\") pod \"certified-operators-f5htw\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.483470 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc5hk\" (UniqueName: \"kubernetes.io/projected/2d352777-d907-4a43-8baf-e263e9fca045-kube-api-access-nc5hk\") pod \"certified-operators-f5htw\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:02 crc kubenswrapper[4978]: I0225 09:19:02.598899 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:03 crc kubenswrapper[4978]: I0225 09:19:03.118651 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f5htw"] Feb 25 09:19:03 crc kubenswrapper[4978]: I0225 09:19:03.260493 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5htw" event={"ID":"2d352777-d907-4a43-8baf-e263e9fca045","Type":"ContainerStarted","Data":"3b0d37748910183ac163293009396dd14e4f33f64711464e4bba59341a0a939e"} Feb 25 09:19:04 crc kubenswrapper[4978]: I0225 09:19:04.271595 4978 generic.go:334] "Generic (PLEG): container finished" podID="2d352777-d907-4a43-8baf-e263e9fca045" containerID="14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3" exitCode=0 Feb 25 09:19:04 crc kubenswrapper[4978]: I0225 09:19:04.271712 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5htw" event={"ID":"2d352777-d907-4a43-8baf-e263e9fca045","Type":"ContainerDied","Data":"14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3"} Feb 25 09:19:04 crc kubenswrapper[4978]: I0225 09:19:04.327668 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:19:04 crc kubenswrapper[4978]: E0225 09:19:04.328114 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:19:06 crc kubenswrapper[4978]: I0225 09:19:06.292446 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5htw" event={"ID":"2d352777-d907-4a43-8baf-e263e9fca045","Type":"ContainerStarted","Data":"eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51"} Feb 25 09:19:08 crc kubenswrapper[4978]: I0225 09:19:08.314016 4978 generic.go:334] "Generic (PLEG): container finished" podID="2d352777-d907-4a43-8baf-e263e9fca045" containerID="eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51" exitCode=0 Feb 25 09:19:08 crc kubenswrapper[4978]: I0225 09:19:08.314074 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5htw" event={"ID":"2d352777-d907-4a43-8baf-e263e9fca045","Type":"ContainerDied","Data":"eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51"} Feb 25 09:19:09 crc kubenswrapper[4978]: I0225 09:19:09.325614 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5htw" event={"ID":"2d352777-d907-4a43-8baf-e263e9fca045","Type":"ContainerStarted","Data":"2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123"} Feb 25 09:19:09 crc kubenswrapper[4978]: I0225 09:19:09.356281 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f5htw" podStartSLOduration=2.911704484 podStartE2EDuration="7.356261297s" podCreationTimestamp="2026-02-25 09:19:02 +0000 UTC" firstStartedPulling="2026-02-25 09:19:04.274294896 +0000 UTC m=+9237.713551405" lastFinishedPulling="2026-02-25 09:19:08.718851759 +0000 UTC m=+9242.158108218" observedRunningTime="2026-02-25 09:19:09.350594101 +0000 UTC m=+9242.789850560" watchObservedRunningTime="2026-02-25 09:19:09.356261297 +0000 UTC m=+9242.795517756" Feb 25 09:19:12 crc kubenswrapper[4978]: I0225 09:19:12.599624 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:12 crc kubenswrapper[4978]: I0225 09:19:12.600125 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:12 crc kubenswrapper[4978]: I0225 09:19:12.672395 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:13 crc kubenswrapper[4978]: I0225 09:19:13.465925 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:13 crc kubenswrapper[4978]: I0225 09:19:13.529735 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f5htw"] Feb 25 09:19:15 crc kubenswrapper[4978]: I0225 09:19:15.406937 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f5htw" podUID="2d352777-d907-4a43-8baf-e263e9fca045" containerName="registry-server" containerID="cri-o://2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123" gracePeriod=2 Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.294050 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.406849 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-utilities\") pod \"2d352777-d907-4a43-8baf-e263e9fca045\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.407074 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc5hk\" (UniqueName: \"kubernetes.io/projected/2d352777-d907-4a43-8baf-e263e9fca045-kube-api-access-nc5hk\") pod \"2d352777-d907-4a43-8baf-e263e9fca045\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.407105 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-catalog-content\") pod \"2d352777-d907-4a43-8baf-e263e9fca045\" (UID: \"2d352777-d907-4a43-8baf-e263e9fca045\") " Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.408058 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-utilities" (OuterVolumeSpecName: "utilities") pod "2d352777-d907-4a43-8baf-e263e9fca045" (UID: "2d352777-d907-4a43-8baf-e263e9fca045"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.416190 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d352777-d907-4a43-8baf-e263e9fca045-kube-api-access-nc5hk" (OuterVolumeSpecName: "kube-api-access-nc5hk") pod "2d352777-d907-4a43-8baf-e263e9fca045" (UID: "2d352777-d907-4a43-8baf-e263e9fca045"). InnerVolumeSpecName "kube-api-access-nc5hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.420483 4978 generic.go:334] "Generic (PLEG): container finished" podID="2d352777-d907-4a43-8baf-e263e9fca045" containerID="2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123" exitCode=0 Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.420547 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f5htw" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.420554 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5htw" event={"ID":"2d352777-d907-4a43-8baf-e263e9fca045","Type":"ContainerDied","Data":"2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123"} Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.420596 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f5htw" event={"ID":"2d352777-d907-4a43-8baf-e263e9fca045","Type":"ContainerDied","Data":"3b0d37748910183ac163293009396dd14e4f33f64711464e4bba59341a0a939e"} Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.420622 4978 scope.go:117] "RemoveContainer" containerID="2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.476466 4978 scope.go:117] "RemoveContainer" containerID="eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.490344 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d352777-d907-4a43-8baf-e263e9fca045" (UID: "2d352777-d907-4a43-8baf-e263e9fca045"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.502895 4978 scope.go:117] "RemoveContainer" containerID="14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.508957 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc5hk\" (UniqueName: \"kubernetes.io/projected/2d352777-d907-4a43-8baf-e263e9fca045-kube-api-access-nc5hk\") on node \"crc\" DevicePath \"\"" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.509007 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.509019 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d352777-d907-4a43-8baf-e263e9fca045-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.574129 4978 scope.go:117] "RemoveContainer" containerID="2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123" Feb 25 09:19:16 crc kubenswrapper[4978]: E0225 09:19:16.581173 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123\": container with ID starting with 2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123 not found: ID does not exist" containerID="2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.581227 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123"} err="failed to get container status \"2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123\": rpc error: code = NotFound desc = could not find container \"2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123\": container with ID starting with 2bbbff22d78060bf15a5c2038b51922e223989eebe8dee147d2a19c20a58a123 not found: ID does not exist" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.581256 4978 scope.go:117] "RemoveContainer" containerID="eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51" Feb 25 09:19:16 crc kubenswrapper[4978]: E0225 09:19:16.581652 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51\": container with ID starting with eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51 not found: ID does not exist" containerID="eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.581683 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51"} err="failed to get container status \"eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51\": rpc error: code = NotFound desc = could not find container \"eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51\": container with ID starting with eeba567d8d3affa6a4bd711478ea6996c03244c9d7b1a4a1e00b39f871994e51 not found: ID does not exist" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.581702 4978 scope.go:117] "RemoveContainer" containerID="14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3" Feb 25 09:19:16 crc kubenswrapper[4978]: E0225 09:19:16.582177 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3\": container with ID starting with 14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3 not found: ID does not exist" containerID="14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.582208 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3"} err="failed to get container status \"14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3\": rpc error: code = NotFound desc = could not find container \"14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3\": container with ID starting with 14d953539aed648ec51b367aac5cfe1963aa3341d2c812449974af98f7784fa3 not found: ID does not exist" Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.765622 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f5htw"] Feb 25 09:19:16 crc kubenswrapper[4978]: I0225 09:19:16.774186 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f5htw"] Feb 25 09:19:17 crc kubenswrapper[4978]: I0225 09:19:17.342444 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d352777-d907-4a43-8baf-e263e9fca045" path="/var/lib/kubelet/pods/2d352777-d907-4a43-8baf-e263e9fca045/volumes" Feb 25 09:19:19 crc kubenswrapper[4978]: I0225 09:19:19.328629 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:19:19 crc kubenswrapper[4978]: E0225 09:19:19.329315 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.106383 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6x8l8"] Feb 25 09:19:28 crc kubenswrapper[4978]: E0225 09:19:28.107434 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d352777-d907-4a43-8baf-e263e9fca045" containerName="registry-server" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.107450 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d352777-d907-4a43-8baf-e263e9fca045" containerName="registry-server" Feb 25 09:19:28 crc kubenswrapper[4978]: E0225 09:19:28.107466 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d352777-d907-4a43-8baf-e263e9fca045" containerName="extract-utilities" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.107474 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d352777-d907-4a43-8baf-e263e9fca045" containerName="extract-utilities" Feb 25 09:19:28 crc kubenswrapper[4978]: E0225 09:19:28.107497 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d352777-d907-4a43-8baf-e263e9fca045" containerName="extract-content" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.107505 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d352777-d907-4a43-8baf-e263e9fca045" containerName="extract-content" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.107775 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d352777-d907-4a43-8baf-e263e9fca045" containerName="registry-server" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.109606 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.123753 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6x8l8"] Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.175704 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-catalog-content\") pod \"redhat-marketplace-6x8l8\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.175858 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rzlf\" (UniqueName: \"kubernetes.io/projected/4c2199e7-e9a1-4efc-917d-ddf915175e85-kube-api-access-5rzlf\") pod \"redhat-marketplace-6x8l8\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.175946 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-utilities\") pod \"redhat-marketplace-6x8l8\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.277639 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-catalog-content\") pod \"redhat-marketplace-6x8l8\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.277773 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rzlf\" (UniqueName: \"kubernetes.io/projected/4c2199e7-e9a1-4efc-917d-ddf915175e85-kube-api-access-5rzlf\") pod \"redhat-marketplace-6x8l8\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.278266 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-utilities\") pod \"redhat-marketplace-6x8l8\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.279017 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-utilities\") pod \"redhat-marketplace-6x8l8\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.279009 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-catalog-content\") pod \"redhat-marketplace-6x8l8\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.299149 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rzlf\" (UniqueName: \"kubernetes.io/projected/4c2199e7-e9a1-4efc-917d-ddf915175e85-kube-api-access-5rzlf\") pod \"redhat-marketplace-6x8l8\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.444709 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:28 crc kubenswrapper[4978]: I0225 09:19:28.960249 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6x8l8"] Feb 25 09:19:29 crc kubenswrapper[4978]: I0225 09:19:29.604771 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6x8l8" event={"ID":"4c2199e7-e9a1-4efc-917d-ddf915175e85","Type":"ContainerStarted","Data":"3d1175263908876aedaa0dc10a3ba5a706cda7e54ef4f8ef78a98fb36f0c2baa"} Feb 25 09:19:30 crc kubenswrapper[4978]: I0225 09:19:30.624745 4978 generic.go:334] "Generic (PLEG): container finished" podID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerID="fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b" exitCode=0 Feb 25 09:19:30 crc kubenswrapper[4978]: I0225 09:19:30.624821 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6x8l8" event={"ID":"4c2199e7-e9a1-4efc-917d-ddf915175e85","Type":"ContainerDied","Data":"fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b"} Feb 25 09:19:32 crc kubenswrapper[4978]: I0225 09:19:32.659253 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6x8l8" event={"ID":"4c2199e7-e9a1-4efc-917d-ddf915175e85","Type":"ContainerStarted","Data":"4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc"} Feb 25 09:19:33 crc kubenswrapper[4978]: I0225 09:19:33.676805 4978 generic.go:334] "Generic (PLEG): container finished" podID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerID="4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc" exitCode=0 Feb 25 09:19:33 crc kubenswrapper[4978]: I0225 09:19:33.677064 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6x8l8" event={"ID":"4c2199e7-e9a1-4efc-917d-ddf915175e85","Type":"ContainerDied","Data":"4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc"} Feb 25 09:19:34 crc kubenswrapper[4978]: I0225 09:19:34.327363 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:19:34 crc kubenswrapper[4978]: E0225 09:19:34.327956 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:19:34 crc kubenswrapper[4978]: I0225 09:19:34.693545 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6x8l8" event={"ID":"4c2199e7-e9a1-4efc-917d-ddf915175e85","Type":"ContainerStarted","Data":"ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771"} Feb 25 09:19:34 crc kubenswrapper[4978]: I0225 09:19:34.718878 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6x8l8" podStartSLOduration=3.224391466 podStartE2EDuration="6.718851668s" podCreationTimestamp="2026-02-25 09:19:28 +0000 UTC" firstStartedPulling="2026-02-25 09:19:30.626867564 +0000 UTC m=+9264.066124033" lastFinishedPulling="2026-02-25 09:19:34.121327736 +0000 UTC m=+9267.560584235" observedRunningTime="2026-02-25 09:19:34.717896648 +0000 UTC m=+9268.157153127" watchObservedRunningTime="2026-02-25 09:19:34.718851668 +0000 UTC m=+9268.158108147" Feb 25 09:19:38 crc kubenswrapper[4978]: I0225 09:19:38.445213 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:38 crc kubenswrapper[4978]: I0225 09:19:38.445837 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:38 crc kubenswrapper[4978]: I0225 09:19:38.532057 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:48 crc kubenswrapper[4978]: I0225 09:19:48.328003 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:19:48 crc kubenswrapper[4978]: E0225 09:19:48.331314 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:19:48 crc kubenswrapper[4978]: I0225 09:19:48.496832 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:48 crc kubenswrapper[4978]: I0225 09:19:48.558689 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6x8l8"] Feb 25 09:19:48 crc kubenswrapper[4978]: I0225 09:19:48.881854 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6x8l8" podUID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerName="registry-server" containerID="cri-o://ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771" gracePeriod=2 Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.389908 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.461800 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-utilities\") pod \"4c2199e7-e9a1-4efc-917d-ddf915175e85\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.461880 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-catalog-content\") pod \"4c2199e7-e9a1-4efc-917d-ddf915175e85\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.461943 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rzlf\" (UniqueName: \"kubernetes.io/projected/4c2199e7-e9a1-4efc-917d-ddf915175e85-kube-api-access-5rzlf\") pod \"4c2199e7-e9a1-4efc-917d-ddf915175e85\" (UID: \"4c2199e7-e9a1-4efc-917d-ddf915175e85\") " Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.462870 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-utilities" (OuterVolumeSpecName: "utilities") pod "4c2199e7-e9a1-4efc-917d-ddf915175e85" (UID: "4c2199e7-e9a1-4efc-917d-ddf915175e85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.477918 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c2199e7-e9a1-4efc-917d-ddf915175e85-kube-api-access-5rzlf" (OuterVolumeSpecName: "kube-api-access-5rzlf") pod "4c2199e7-e9a1-4efc-917d-ddf915175e85" (UID: "4c2199e7-e9a1-4efc-917d-ddf915175e85"). InnerVolumeSpecName "kube-api-access-5rzlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.499822 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c2199e7-e9a1-4efc-917d-ddf915175e85" (UID: "4c2199e7-e9a1-4efc-917d-ddf915175e85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.565004 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.565053 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c2199e7-e9a1-4efc-917d-ddf915175e85-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.565068 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rzlf\" (UniqueName: \"kubernetes.io/projected/4c2199e7-e9a1-4efc-917d-ddf915175e85-kube-api-access-5rzlf\") on node \"crc\" DevicePath \"\"" Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.900325 4978 generic.go:334] "Generic (PLEG): container finished" podID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerID="ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771" exitCode=0 Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.900404 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6x8l8" event={"ID":"4c2199e7-e9a1-4efc-917d-ddf915175e85","Type":"ContainerDied","Data":"ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771"} Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.900477 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6x8l8" event={"ID":"4c2199e7-e9a1-4efc-917d-ddf915175e85","Type":"ContainerDied","Data":"3d1175263908876aedaa0dc10a3ba5a706cda7e54ef4f8ef78a98fb36f0c2baa"} Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.900503 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6x8l8" Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.900513 4978 scope.go:117] "RemoveContainer" containerID="ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771" Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.940195 4978 scope.go:117] "RemoveContainer" containerID="4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc" Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.966489 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6x8l8"] Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.979253 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6x8l8"] Feb 25 09:19:49 crc kubenswrapper[4978]: I0225 09:19:49.984413 4978 scope.go:117] "RemoveContainer" containerID="fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b" Feb 25 09:19:50 crc kubenswrapper[4978]: I0225 09:19:50.050645 4978 scope.go:117] "RemoveContainer" containerID="ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771" Feb 25 09:19:50 crc kubenswrapper[4978]: E0225 09:19:50.051256 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771\": container with ID starting with ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771 not found: ID does not exist" containerID="ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771" Feb 25 09:19:50 crc kubenswrapper[4978]: I0225 09:19:50.051308 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771"} err="failed to get container status \"ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771\": rpc error: code = NotFound desc = could not find container \"ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771\": container with ID starting with ba968c4b270781393d555af0b6766ff55352e3b53b0267f106499ce86ae60771 not found: ID does not exist" Feb 25 09:19:50 crc kubenswrapper[4978]: I0225 09:19:50.051340 4978 scope.go:117] "RemoveContainer" containerID="4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc" Feb 25 09:19:50 crc kubenswrapper[4978]: E0225 09:19:50.051868 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc\": container with ID starting with 4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc not found: ID does not exist" containerID="4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc" Feb 25 09:19:50 crc kubenswrapper[4978]: I0225 09:19:50.051930 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc"} err="failed to get container status \"4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc\": rpc error: code = NotFound desc = could not find container \"4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc\": container with ID starting with 4b50119fc4daf6e66f2f1b036f2fd61fd4a69588550dba7485e4c27436e259cc not found: ID does not exist" Feb 25 09:19:50 crc kubenswrapper[4978]: I0225 09:19:50.051962 4978 scope.go:117] "RemoveContainer" containerID="fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b" Feb 25 09:19:50 crc kubenswrapper[4978]: E0225 09:19:50.052567 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b\": container with ID starting with fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b not found: ID does not exist" containerID="fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b" Feb 25 09:19:50 crc kubenswrapper[4978]: I0225 09:19:50.052595 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b"} err="failed to get container status \"fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b\": rpc error: code = NotFound desc = could not find container \"fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b\": container with ID starting with fe8fbc183a67af0c1840e5af0b114311c475d13872adee0f98902b6f8d68304b not found: ID does not exist" Feb 25 09:19:51 crc kubenswrapper[4978]: I0225 09:19:51.348462 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c2199e7-e9a1-4efc-917d-ddf915175e85" path="/var/lib/kubelet/pods/4c2199e7-e9a1-4efc-917d-ddf915175e85/volumes" Feb 25 09:19:59 crc kubenswrapper[4978]: I0225 09:19:59.328517 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:19:59 crc kubenswrapper[4978]: E0225 09:19:59.329852 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.176875 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533520-ms2xm"] Feb 25 09:20:00 crc kubenswrapper[4978]: E0225 09:20:00.177674 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerName="registry-server" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.177707 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerName="registry-server" Feb 25 09:20:00 crc kubenswrapper[4978]: E0225 09:20:00.177775 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerName="extract-utilities" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.177790 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerName="extract-utilities" Feb 25 09:20:00 crc kubenswrapper[4978]: E0225 09:20:00.177826 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerName="extract-content" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.177838 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerName="extract-content" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.178231 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c2199e7-e9a1-4efc-917d-ddf915175e85" containerName="registry-server" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.179435 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533520-ms2xm" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.187199 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.187247 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.187431 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.190932 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533520-ms2xm"] Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.345154 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpfwj\" (UniqueName: \"kubernetes.io/projected/e8b23b30-7e9c-474c-a663-ffdf14c61085-kube-api-access-cpfwj\") pod \"auto-csr-approver-29533520-ms2xm\" (UID: \"e8b23b30-7e9c-474c-a663-ffdf14c61085\") " pod="openshift-infra/auto-csr-approver-29533520-ms2xm" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.448545 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpfwj\" (UniqueName: \"kubernetes.io/projected/e8b23b30-7e9c-474c-a663-ffdf14c61085-kube-api-access-cpfwj\") pod \"auto-csr-approver-29533520-ms2xm\" (UID: \"e8b23b30-7e9c-474c-a663-ffdf14c61085\") " pod="openshift-infra/auto-csr-approver-29533520-ms2xm" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.478824 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpfwj\" (UniqueName: \"kubernetes.io/projected/e8b23b30-7e9c-474c-a663-ffdf14c61085-kube-api-access-cpfwj\") pod \"auto-csr-approver-29533520-ms2xm\" (UID: \"e8b23b30-7e9c-474c-a663-ffdf14c61085\") " pod="openshift-infra/auto-csr-approver-29533520-ms2xm" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.501908 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533520-ms2xm" Feb 25 09:20:00 crc kubenswrapper[4978]: I0225 09:20:00.985216 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533520-ms2xm"] Feb 25 09:20:01 crc kubenswrapper[4978]: I0225 09:20:01.062161 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533520-ms2xm" event={"ID":"e8b23b30-7e9c-474c-a663-ffdf14c61085","Type":"ContainerStarted","Data":"d2d68e9c2cc65d70cbc2f618feee4268d889a86a4a69d54b7148a6e932bcd624"} Feb 25 09:20:03 crc kubenswrapper[4978]: I0225 09:20:03.093228 4978 generic.go:334] "Generic (PLEG): container finished" podID="e8b23b30-7e9c-474c-a663-ffdf14c61085" containerID="d37104d3ab4631454cffbeb4b63fb198b3ac792ebf9f978399d51c1fd526aee0" exitCode=0 Feb 25 09:20:03 crc kubenswrapper[4978]: I0225 09:20:03.093350 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533520-ms2xm" event={"ID":"e8b23b30-7e9c-474c-a663-ffdf14c61085","Type":"ContainerDied","Data":"d37104d3ab4631454cffbeb4b63fb198b3ac792ebf9f978399d51c1fd526aee0"} Feb 25 09:20:04 crc kubenswrapper[4978]: I0225 09:20:04.529165 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533520-ms2xm" Feb 25 09:20:04 crc kubenswrapper[4978]: I0225 09:20:04.579849 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpfwj\" (UniqueName: \"kubernetes.io/projected/e8b23b30-7e9c-474c-a663-ffdf14c61085-kube-api-access-cpfwj\") pod \"e8b23b30-7e9c-474c-a663-ffdf14c61085\" (UID: \"e8b23b30-7e9c-474c-a663-ffdf14c61085\") " Feb 25 09:20:04 crc kubenswrapper[4978]: I0225 09:20:04.592696 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8b23b30-7e9c-474c-a663-ffdf14c61085-kube-api-access-cpfwj" (OuterVolumeSpecName: "kube-api-access-cpfwj") pod "e8b23b30-7e9c-474c-a663-ffdf14c61085" (UID: "e8b23b30-7e9c-474c-a663-ffdf14c61085"). InnerVolumeSpecName "kube-api-access-cpfwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:20:04 crc kubenswrapper[4978]: I0225 09:20:04.682865 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpfwj\" (UniqueName: \"kubernetes.io/projected/e8b23b30-7e9c-474c-a663-ffdf14c61085-kube-api-access-cpfwj\") on node \"crc\" DevicePath \"\"" Feb 25 09:20:05 crc kubenswrapper[4978]: I0225 09:20:05.122790 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533520-ms2xm" event={"ID":"e8b23b30-7e9c-474c-a663-ffdf14c61085","Type":"ContainerDied","Data":"d2d68e9c2cc65d70cbc2f618feee4268d889a86a4a69d54b7148a6e932bcd624"} Feb 25 09:20:05 crc kubenswrapper[4978]: I0225 09:20:05.122848 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2d68e9c2cc65d70cbc2f618feee4268d889a86a4a69d54b7148a6e932bcd624" Feb 25 09:20:05 crc kubenswrapper[4978]: I0225 09:20:05.122892 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533520-ms2xm" Feb 25 09:20:05 crc kubenswrapper[4978]: I0225 09:20:05.625543 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533514-w7wkw"] Feb 25 09:20:05 crc kubenswrapper[4978]: I0225 09:20:05.637747 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533514-w7wkw"] Feb 25 09:20:07 crc kubenswrapper[4978]: I0225 09:20:07.352583 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ae641dc-1ef2-464e-b1da-6d7341626b34" path="/var/lib/kubelet/pods/0ae641dc-1ef2-464e-b1da-6d7341626b34/volumes" Feb 25 09:20:14 crc kubenswrapper[4978]: I0225 09:20:14.327418 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:20:14 crc kubenswrapper[4978]: E0225 09:20:14.328705 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:20:29 crc kubenswrapper[4978]: I0225 09:20:29.327823 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:20:29 crc kubenswrapper[4978]: E0225 09:20:29.328850 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:20:36 crc kubenswrapper[4978]: I0225 09:20:36.440436 4978 scope.go:117] "RemoveContainer" containerID="d4fdd675d37c0f972b6bd25dd903d0c351d3a314487fc067a3ba615e54ac475e" Feb 25 09:20:44 crc kubenswrapper[4978]: I0225 09:20:44.330107 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:20:44 crc kubenswrapper[4978]: E0225 09:20:44.331527 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:20:59 crc kubenswrapper[4978]: I0225 09:20:59.328255 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:20:59 crc kubenswrapper[4978]: E0225 09:20:59.329030 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:21:12 crc kubenswrapper[4978]: I0225 09:21:12.328154 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:21:12 crc kubenswrapper[4978]: E0225 09:21:12.329718 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:21:23 crc kubenswrapper[4978]: I0225 09:21:23.328759 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:21:23 crc kubenswrapper[4978]: E0225 09:21:23.330014 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:21:36 crc kubenswrapper[4978]: I0225 09:21:36.328953 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:21:36 crc kubenswrapper[4978]: E0225 09:21:36.330310 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:21:47 crc kubenswrapper[4978]: I0225 09:21:47.344121 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:21:47 crc kubenswrapper[4978]: E0225 09:21:47.345122 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:21:59 crc kubenswrapper[4978]: I0225 09:21:59.328114 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:21:59 crc kubenswrapper[4978]: E0225 09:21:59.329185 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.165028 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533522-zhsm2"] Feb 25 09:22:00 crc kubenswrapper[4978]: E0225 09:22:00.166006 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8b23b30-7e9c-474c-a663-ffdf14c61085" containerName="oc" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.166033 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8b23b30-7e9c-474c-a663-ffdf14c61085" containerName="oc" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.166418 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8b23b30-7e9c-474c-a663-ffdf14c61085" containerName="oc" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.167574 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533522-zhsm2" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.170759 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.171070 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.171862 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.177318 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533522-zhsm2"] Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.253497 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtgjd\" (UniqueName: \"kubernetes.io/projected/1159a8fb-8eca-48a1-9f7a-b01a5d027206-kube-api-access-qtgjd\") pod \"auto-csr-approver-29533522-zhsm2\" (UID: \"1159a8fb-8eca-48a1-9f7a-b01a5d027206\") " pod="openshift-infra/auto-csr-approver-29533522-zhsm2" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.355676 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtgjd\" (UniqueName: \"kubernetes.io/projected/1159a8fb-8eca-48a1-9f7a-b01a5d027206-kube-api-access-qtgjd\") pod \"auto-csr-approver-29533522-zhsm2\" (UID: \"1159a8fb-8eca-48a1-9f7a-b01a5d027206\") " pod="openshift-infra/auto-csr-approver-29533522-zhsm2" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.375706 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtgjd\" (UniqueName: \"kubernetes.io/projected/1159a8fb-8eca-48a1-9f7a-b01a5d027206-kube-api-access-qtgjd\") pod \"auto-csr-approver-29533522-zhsm2\" (UID: \"1159a8fb-8eca-48a1-9f7a-b01a5d027206\") " pod="openshift-infra/auto-csr-approver-29533522-zhsm2" Feb 25 09:22:00 crc kubenswrapper[4978]: I0225 09:22:00.529819 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533522-zhsm2" Feb 25 09:22:01 crc kubenswrapper[4978]: I0225 09:22:01.049737 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:22:01 crc kubenswrapper[4978]: I0225 09:22:01.051620 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533522-zhsm2"] Feb 25 09:22:01 crc kubenswrapper[4978]: I0225 09:22:01.551944 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533522-zhsm2" event={"ID":"1159a8fb-8eca-48a1-9f7a-b01a5d027206","Type":"ContainerStarted","Data":"21d4f561c55da6aa3b37d1a862c972ac57f85e7c567c5c31ad4661c6a5efb148"} Feb 25 09:22:02 crc kubenswrapper[4978]: I0225 09:22:02.567351 4978 generic.go:334] "Generic (PLEG): container finished" podID="1159a8fb-8eca-48a1-9f7a-b01a5d027206" containerID="6bdce94aa874c75a28a7d62b34953625d11d56ad0107a56b72908133ec3803c0" exitCode=0 Feb 25 09:22:02 crc kubenswrapper[4978]: I0225 09:22:02.567422 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533522-zhsm2" event={"ID":"1159a8fb-8eca-48a1-9f7a-b01a5d027206","Type":"ContainerDied","Data":"6bdce94aa874c75a28a7d62b34953625d11d56ad0107a56b72908133ec3803c0"} Feb 25 09:22:03 crc kubenswrapper[4978]: I0225 09:22:03.984129 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533522-zhsm2" Feb 25 09:22:04 crc kubenswrapper[4978]: I0225 09:22:04.045196 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtgjd\" (UniqueName: \"kubernetes.io/projected/1159a8fb-8eca-48a1-9f7a-b01a5d027206-kube-api-access-qtgjd\") pod \"1159a8fb-8eca-48a1-9f7a-b01a5d027206\" (UID: \"1159a8fb-8eca-48a1-9f7a-b01a5d027206\") " Feb 25 09:22:04 crc kubenswrapper[4978]: I0225 09:22:04.054617 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1159a8fb-8eca-48a1-9f7a-b01a5d027206-kube-api-access-qtgjd" (OuterVolumeSpecName: "kube-api-access-qtgjd") pod "1159a8fb-8eca-48a1-9f7a-b01a5d027206" (UID: "1159a8fb-8eca-48a1-9f7a-b01a5d027206"). InnerVolumeSpecName "kube-api-access-qtgjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:22:04 crc kubenswrapper[4978]: I0225 09:22:04.149103 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtgjd\" (UniqueName: \"kubernetes.io/projected/1159a8fb-8eca-48a1-9f7a-b01a5d027206-kube-api-access-qtgjd\") on node \"crc\" DevicePath \"\"" Feb 25 09:22:04 crc kubenswrapper[4978]: I0225 09:22:04.595109 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533522-zhsm2" event={"ID":"1159a8fb-8eca-48a1-9f7a-b01a5d027206","Type":"ContainerDied","Data":"21d4f561c55da6aa3b37d1a862c972ac57f85e7c567c5c31ad4661c6a5efb148"} Feb 25 09:22:04 crc kubenswrapper[4978]: I0225 09:22:04.595495 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21d4f561c55da6aa3b37d1a862c972ac57f85e7c567c5c31ad4661c6a5efb148" Feb 25 09:22:04 crc kubenswrapper[4978]: I0225 09:22:04.595181 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533522-zhsm2" Feb 25 09:22:05 crc kubenswrapper[4978]: I0225 09:22:05.070096 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533516-lnrdv"] Feb 25 09:22:05 crc kubenswrapper[4978]: I0225 09:22:05.082902 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533516-lnrdv"] Feb 25 09:22:05 crc kubenswrapper[4978]: I0225 09:22:05.343738 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd4afe50-4f1e-4522-bb21-9fb4208b350e" path="/var/lib/kubelet/pods/bd4afe50-4f1e-4522-bb21-9fb4208b350e/volumes" Feb 25 09:22:12 crc kubenswrapper[4978]: I0225 09:22:12.327560 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:22:12 crc kubenswrapper[4978]: E0225 09:22:12.329522 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:22:26 crc kubenswrapper[4978]: I0225 09:22:26.327939 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:22:26 crc kubenswrapper[4978]: E0225 09:22:26.329098 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:22:36 crc kubenswrapper[4978]: I0225 09:22:36.576673 4978 scope.go:117] "RemoveContainer" containerID="3818c5f3d610ca0d306ccdc828b4f4567c653ff242893ba9092078dac3838fff" Feb 25 09:22:37 crc kubenswrapper[4978]: I0225 09:22:37.337924 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:22:37 crc kubenswrapper[4978]: E0225 09:22:37.338824 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:22:48 crc kubenswrapper[4978]: I0225 09:22:48.328395 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:22:49 crc kubenswrapper[4978]: I0225 09:22:49.143883 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"ccf3e1f2fc79ae5a3b75d3365e4d16775ede1da95dce83d44c7c298fd7c0a322"} Feb 25 09:22:50 crc kubenswrapper[4978]: I0225 09:22:50.168707 4978 generic.go:334] "Generic (PLEG): container finished" podID="41f59382-722c-42ab-89f1-0e738135edcd" containerID="db3dc0fcf3282373845b13bc106f325899d230631191085bc6c0bc34d1733313" exitCode=0 Feb 25 09:22:50 crc kubenswrapper[4978]: I0225 09:22:50.168805 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2cksr" event={"ID":"41f59382-722c-42ab-89f1-0e738135edcd","Type":"ContainerDied","Data":"db3dc0fcf3282373845b13bc106f325899d230631191085bc6c0bc34d1733313"} Feb 25 09:22:51 crc kubenswrapper[4978]: I0225 09:22:51.640453 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:22:51 crc kubenswrapper[4978]: I0225 09:22:51.755039 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-combined-ca-bundle\") pod \"41f59382-722c-42ab-89f1-0e738135edcd\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " Feb 25 09:22:51 crc kubenswrapper[4978]: I0225 09:22:51.755117 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-inventory\") pod \"41f59382-722c-42ab-89f1-0e738135edcd\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " Feb 25 09:22:51 crc kubenswrapper[4978]: I0225 09:22:51.755287 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-ssh-key-openstack-cell1\") pod \"41f59382-722c-42ab-89f1-0e738135edcd\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " Feb 25 09:22:51 crc kubenswrapper[4978]: I0225 09:22:51.755425 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-secret-0\") pod \"41f59382-722c-42ab-89f1-0e738135edcd\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " Feb 25 09:22:51 crc kubenswrapper[4978]: I0225 09:22:51.755495 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2pc6\" (UniqueName: \"kubernetes.io/projected/41f59382-722c-42ab-89f1-0e738135edcd-kube-api-access-q2pc6\") pod \"41f59382-722c-42ab-89f1-0e738135edcd\" (UID: \"41f59382-722c-42ab-89f1-0e738135edcd\") " Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.193239 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2cksr" event={"ID":"41f59382-722c-42ab-89f1-0e738135edcd","Type":"ContainerDied","Data":"aa5a14c828aa28f757396831f12cc91c28106b8fb4072151f6d2aae1cd99ab6a"} Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.193294 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa5a14c828aa28f757396831f12cc91c28106b8fb4072151f6d2aae1cd99ab6a" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.193327 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2cksr" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.363007 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-qp6rn"] Feb 25 09:22:52 crc kubenswrapper[4978]: E0225 09:22:52.363470 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1159a8fb-8eca-48a1-9f7a-b01a5d027206" containerName="oc" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.363490 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1159a8fb-8eca-48a1-9f7a-b01a5d027206" containerName="oc" Feb 25 09:22:52 crc kubenswrapper[4978]: E0225 09:22:52.363521 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41f59382-722c-42ab-89f1-0e738135edcd" containerName="libvirt-openstack-openstack-cell1" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.363532 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="41f59382-722c-42ab-89f1-0e738135edcd" containerName="libvirt-openstack-openstack-cell1" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.363778 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="41f59382-722c-42ab-89f1-0e738135edcd" containerName="libvirt-openstack-openstack-cell1" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.363799 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1159a8fb-8eca-48a1-9f7a-b01a5d027206" containerName="oc" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.364627 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.367697 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.367869 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.368023 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.371035 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.371225 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.371298 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.371527 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.371674 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.371938 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.372039 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.372193 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-inventory\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.372249 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.372548 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.373476 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl6mg\" (UniqueName: \"kubernetes.io/projected/d39c9d7a-8206-4313-9915-aa3825c28444-kube-api-access-wl6mg\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.403965 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-qp6rn"] Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.475999 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.476090 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.476130 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-inventory\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.476157 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.477215 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.477273 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.477282 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl6mg\" (UniqueName: \"kubernetes.io/projected/d39c9d7a-8206-4313-9915-aa3825c28444-kube-api-access-wl6mg\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.477532 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.477604 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.477633 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.477736 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.477826 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.493907 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "41f59382-722c-42ab-89f1-0e738135edcd" (UID: "41f59382-722c-42ab-89f1-0e738135edcd"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.495300 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.495322 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41f59382-722c-42ab-89f1-0e738135edcd-kube-api-access-q2pc6" (OuterVolumeSpecName: "kube-api-access-q2pc6") pod "41f59382-722c-42ab-89f1-0e738135edcd" (UID: "41f59382-722c-42ab-89f1-0e738135edcd"). InnerVolumeSpecName "kube-api-access-q2pc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.499592 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.500364 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl6mg\" (UniqueName: \"kubernetes.io/projected/d39c9d7a-8206-4313-9915-aa3825c28444-kube-api-access-wl6mg\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.502912 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-inventory\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.503473 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.505036 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.505522 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.505886 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.508852 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.509067 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-qp6rn\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.550971 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-inventory" (OuterVolumeSpecName: "inventory") pod "41f59382-722c-42ab-89f1-0e738135edcd" (UID: "41f59382-722c-42ab-89f1-0e738135edcd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.552910 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "41f59382-722c-42ab-89f1-0e738135edcd" (UID: "41f59382-722c-42ab-89f1-0e738135edcd"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.557478 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "41f59382-722c-42ab-89f1-0e738135edcd" (UID: "41f59382-722c-42ab-89f1-0e738135edcd"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.579312 4978 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.579353 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.579383 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.579396 4978 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/41f59382-722c-42ab-89f1-0e738135edcd-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.579410 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2pc6\" (UniqueName: \"kubernetes.io/projected/41f59382-722c-42ab-89f1-0e738135edcd-kube-api-access-q2pc6\") on node \"crc\" DevicePath \"\"" Feb 25 09:22:52 crc kubenswrapper[4978]: I0225 09:22:52.683727 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:22:53 crc kubenswrapper[4978]: W0225 09:22:53.273868 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd39c9d7a_8206_4313_9915_aa3825c28444.slice/crio-a7ba051bd5bdbc219b25295b2ba8ef2a42969ff8d5bdbb90c73a4f85eeabdc6f WatchSource:0}: Error finding container a7ba051bd5bdbc219b25295b2ba8ef2a42969ff8d5bdbb90c73a4f85eeabdc6f: Status 404 returned error can't find the container with id a7ba051bd5bdbc219b25295b2ba8ef2a42969ff8d5bdbb90c73a4f85eeabdc6f Feb 25 09:22:53 crc kubenswrapper[4978]: I0225 09:22:53.275881 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-qp6rn"] Feb 25 09:22:54 crc kubenswrapper[4978]: I0225 09:22:54.220713 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" event={"ID":"d39c9d7a-8206-4313-9915-aa3825c28444","Type":"ContainerStarted","Data":"a7ba051bd5bdbc219b25295b2ba8ef2a42969ff8d5bdbb90c73a4f85eeabdc6f"} Feb 25 09:22:56 crc kubenswrapper[4978]: I0225 09:22:56.250842 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" event={"ID":"d39c9d7a-8206-4313-9915-aa3825c28444","Type":"ContainerStarted","Data":"4cdd0dc4ece5900888dfec5efbafbe1631859780f5b91138abffa38326748147"} Feb 25 09:22:56 crc kubenswrapper[4978]: I0225 09:22:56.284052 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" podStartSLOduration=3.108795867 podStartE2EDuration="4.284028678s" podCreationTimestamp="2026-02-25 09:22:52 +0000 UTC" firstStartedPulling="2026-02-25 09:22:53.276221561 +0000 UTC m=+9466.715478020" lastFinishedPulling="2026-02-25 09:22:54.451454372 +0000 UTC m=+9467.890710831" observedRunningTime="2026-02-25 09:22:56.278566098 +0000 UTC m=+9469.717822627" watchObservedRunningTime="2026-02-25 09:22:56.284028678 +0000 UTC m=+9469.723285147" Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.156978 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533524-qlzgg"] Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.158796 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533524-qlzgg" Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.160927 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.160974 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.167941 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.173855 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533524-qlzgg"] Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.187419 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcnbb\" (UniqueName: \"kubernetes.io/projected/7ff69947-557c-4d72-9179-1c9ea3a306cf-kube-api-access-lcnbb\") pod \"auto-csr-approver-29533524-qlzgg\" (UID: \"7ff69947-557c-4d72-9179-1c9ea3a306cf\") " pod="openshift-infra/auto-csr-approver-29533524-qlzgg" Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.294560 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcnbb\" (UniqueName: \"kubernetes.io/projected/7ff69947-557c-4d72-9179-1c9ea3a306cf-kube-api-access-lcnbb\") pod \"auto-csr-approver-29533524-qlzgg\" (UID: \"7ff69947-557c-4d72-9179-1c9ea3a306cf\") " pod="openshift-infra/auto-csr-approver-29533524-qlzgg" Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.320538 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcnbb\" (UniqueName: \"kubernetes.io/projected/7ff69947-557c-4d72-9179-1c9ea3a306cf-kube-api-access-lcnbb\") pod \"auto-csr-approver-29533524-qlzgg\" (UID: \"7ff69947-557c-4d72-9179-1c9ea3a306cf\") " pod="openshift-infra/auto-csr-approver-29533524-qlzgg" Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.481702 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533524-qlzgg" Feb 25 09:24:00 crc kubenswrapper[4978]: I0225 09:24:00.995569 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533524-qlzgg"] Feb 25 09:24:01 crc kubenswrapper[4978]: I0225 09:24:01.024601 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533524-qlzgg" event={"ID":"7ff69947-557c-4d72-9179-1c9ea3a306cf","Type":"ContainerStarted","Data":"345ce5dc33e148213fa33f9b2d0dcee065b9e75f7db3f9f0d92a2574049b6dd7"} Feb 25 09:24:03 crc kubenswrapper[4978]: I0225 09:24:03.048271 4978 generic.go:334] "Generic (PLEG): container finished" podID="7ff69947-557c-4d72-9179-1c9ea3a306cf" containerID="46453f0473c10c6c1b411d9e56d5ed7fb001be232100a085ae82d8f46ea4a024" exitCode=0 Feb 25 09:24:03 crc kubenswrapper[4978]: I0225 09:24:03.048405 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533524-qlzgg" event={"ID":"7ff69947-557c-4d72-9179-1c9ea3a306cf","Type":"ContainerDied","Data":"46453f0473c10c6c1b411d9e56d5ed7fb001be232100a085ae82d8f46ea4a024"} Feb 25 09:24:04 crc kubenswrapper[4978]: I0225 09:24:04.622424 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533524-qlzgg" Feb 25 09:24:04 crc kubenswrapper[4978]: I0225 09:24:04.695484 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcnbb\" (UniqueName: \"kubernetes.io/projected/7ff69947-557c-4d72-9179-1c9ea3a306cf-kube-api-access-lcnbb\") pod \"7ff69947-557c-4d72-9179-1c9ea3a306cf\" (UID: \"7ff69947-557c-4d72-9179-1c9ea3a306cf\") " Feb 25 09:24:04 crc kubenswrapper[4978]: I0225 09:24:04.703659 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ff69947-557c-4d72-9179-1c9ea3a306cf-kube-api-access-lcnbb" (OuterVolumeSpecName: "kube-api-access-lcnbb") pod "7ff69947-557c-4d72-9179-1c9ea3a306cf" (UID: "7ff69947-557c-4d72-9179-1c9ea3a306cf"). InnerVolumeSpecName "kube-api-access-lcnbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:24:04 crc kubenswrapper[4978]: I0225 09:24:04.798188 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcnbb\" (UniqueName: \"kubernetes.io/projected/7ff69947-557c-4d72-9179-1c9ea3a306cf-kube-api-access-lcnbb\") on node \"crc\" DevicePath \"\"" Feb 25 09:24:05 crc kubenswrapper[4978]: I0225 09:24:05.074073 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533524-qlzgg" event={"ID":"7ff69947-557c-4d72-9179-1c9ea3a306cf","Type":"ContainerDied","Data":"345ce5dc33e148213fa33f9b2d0dcee065b9e75f7db3f9f0d92a2574049b6dd7"} Feb 25 09:24:05 crc kubenswrapper[4978]: I0225 09:24:05.074113 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533524-qlzgg" Feb 25 09:24:05 crc kubenswrapper[4978]: I0225 09:24:05.074131 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="345ce5dc33e148213fa33f9b2d0dcee065b9e75f7db3f9f0d92a2574049b6dd7" Feb 25 09:24:05 crc kubenswrapper[4978]: I0225 09:24:05.704510 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533518-ph8zb"] Feb 25 09:24:05 crc kubenswrapper[4978]: I0225 09:24:05.714745 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533518-ph8zb"] Feb 25 09:24:07 crc kubenswrapper[4978]: I0225 09:24:07.347840 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecd6d99e-ee38-47dc-887b-d5ed8b2f4588" path="/var/lib/kubelet/pods/ecd6d99e-ee38-47dc-887b-d5ed8b2f4588/volumes" Feb 25 09:24:37 crc kubenswrapper[4978]: I0225 09:24:37.124960 4978 scope.go:117] "RemoveContainer" containerID="4ad66268b89d07953efd824b688873af6a7756a873a29b93e8c908a4dd839adc" Feb 25 09:25:16 crc kubenswrapper[4978]: I0225 09:25:16.540538 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:25:16 crc kubenswrapper[4978]: I0225 09:25:16.541291 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:25:46 crc kubenswrapper[4978]: I0225 09:25:46.540751 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:25:46 crc kubenswrapper[4978]: I0225 09:25:46.541359 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:25:58 crc kubenswrapper[4978]: I0225 09:25:58.464606 4978 generic.go:334] "Generic (PLEG): container finished" podID="d39c9d7a-8206-4313-9915-aa3825c28444" containerID="4cdd0dc4ece5900888dfec5efbafbe1631859780f5b91138abffa38326748147" exitCode=0 Feb 25 09:25:58 crc kubenswrapper[4978]: I0225 09:25:58.464710 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" event={"ID":"d39c9d7a-8206-4313-9915-aa3825c28444","Type":"ContainerDied","Data":"4cdd0dc4ece5900888dfec5efbafbe1631859780f5b91138abffa38326748147"} Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.208806 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533526-6k9dz"] Feb 25 09:26:00 crc kubenswrapper[4978]: E0225 09:26:00.210855 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff69947-557c-4d72-9179-1c9ea3a306cf" containerName="oc" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.210894 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff69947-557c-4d72-9179-1c9ea3a306cf" containerName="oc" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.211174 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ff69947-557c-4d72-9179-1c9ea3a306cf" containerName="oc" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.212080 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533526-6k9dz" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.215243 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.215432 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.215510 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.221200 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533526-6k9dz"] Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.390885 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxk4l\" (UniqueName: \"kubernetes.io/projected/c9d41622-d2f7-483a-90fa-00c2222fab68-kube-api-access-wxk4l\") pod \"auto-csr-approver-29533526-6k9dz\" (UID: \"c9d41622-d2f7-483a-90fa-00c2222fab68\") " pod="openshift-infra/auto-csr-approver-29533526-6k9dz" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.486906 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" event={"ID":"d39c9d7a-8206-4313-9915-aa3825c28444","Type":"ContainerDied","Data":"a7ba051bd5bdbc219b25295b2ba8ef2a42969ff8d5bdbb90c73a4f85eeabdc6f"} Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.486953 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7ba051bd5bdbc219b25295b2ba8ef2a42969ff8d5bdbb90c73a4f85eeabdc6f" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.494582 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxk4l\" (UniqueName: \"kubernetes.io/projected/c9d41622-d2f7-483a-90fa-00c2222fab68-kube-api-access-wxk4l\") pod \"auto-csr-approver-29533526-6k9dz\" (UID: \"c9d41622-d2f7-483a-90fa-00c2222fab68\") " pod="openshift-infra/auto-csr-approver-29533526-6k9dz" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.512471 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxk4l\" (UniqueName: \"kubernetes.io/projected/c9d41622-d2f7-483a-90fa-00c2222fab68-kube-api-access-wxk4l\") pod \"auto-csr-approver-29533526-6k9dz\" (UID: \"c9d41622-d2f7-483a-90fa-00c2222fab68\") " pod="openshift-infra/auto-csr-approver-29533526-6k9dz" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.535708 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533526-6k9dz" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.701094 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.800832 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cells-global-config-0\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.801095 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-combined-ca-bundle\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.801140 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl6mg\" (UniqueName: \"kubernetes.io/projected/d39c9d7a-8206-4313-9915-aa3825c28444-kube-api-access-wl6mg\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.801167 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-inventory\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.801198 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-0\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.801257 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-3\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.801318 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-1\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.801422 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-2\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.801450 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-ssh-key-openstack-cell1\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.801470 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-1\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.801514 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-0\") pod \"d39c9d7a-8206-4313-9915-aa3825c28444\" (UID: \"d39c9d7a-8206-4313-9915-aa3825c28444\") " Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.809813 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.826291 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d39c9d7a-8206-4313-9915-aa3825c28444-kube-api-access-wl6mg" (OuterVolumeSpecName: "kube-api-access-wl6mg") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "kube-api-access-wl6mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.841151 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.846464 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-inventory" (OuterVolumeSpecName: "inventory") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.872610 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.873841 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.877736 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.881638 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.891815 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.891891 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.899588 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "d39c9d7a-8206-4313-9915-aa3825c28444" (UID: "d39c9d7a-8206-4313-9915-aa3825c28444"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905871 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905899 4978 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905909 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905918 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905929 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905937 4978 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905946 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905954 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905962 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl6mg\" (UniqueName: \"kubernetes.io/projected/d39c9d7a-8206-4313-9915-aa3825c28444-kube-api-access-wl6mg\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905971 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:00 crc kubenswrapper[4978]: I0225 09:26:00.905979 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d39c9d7a-8206-4313-9915-aa3825c28444-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.049937 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533526-6k9dz"] Feb 25 09:26:01 crc kubenswrapper[4978]: W0225 09:26:01.059633 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc9d41622_d2f7_483a_90fa_00c2222fab68.slice/crio-37e0943a14167071440306dad25a149c7c3304dd4b722ef9928fbf2e3e5c90fd WatchSource:0}: Error finding container 37e0943a14167071440306dad25a149c7c3304dd4b722ef9928fbf2e3e5c90fd: Status 404 returned error can't find the container with id 37e0943a14167071440306dad25a149c7c3304dd4b722ef9928fbf2e3e5c90fd Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.496916 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533526-6k9dz" event={"ID":"c9d41622-d2f7-483a-90fa-00c2222fab68","Type":"ContainerStarted","Data":"37e0943a14167071440306dad25a149c7c3304dd4b722ef9928fbf2e3e5c90fd"} Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.496962 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-qp6rn" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.846981 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-z7lrc"] Feb 25 09:26:01 crc kubenswrapper[4978]: E0225 09:26:01.847560 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d39c9d7a-8206-4313-9915-aa3825c28444" containerName="nova-cell1-openstack-openstack-cell1" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.847582 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d39c9d7a-8206-4313-9915-aa3825c28444" containerName="nova-cell1-openstack-openstack-cell1" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.847994 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d39c9d7a-8206-4313-9915-aa3825c28444" containerName="nova-cell1-openstack-openstack-cell1" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.848913 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.852186 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.852195 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.852578 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.852794 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.855339 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 25 09:26:01 crc kubenswrapper[4978]: I0225 09:26:01.863101 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-z7lrc"] Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.029221 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.029325 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.029386 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-inventory\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.029495 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.029533 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.029586 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.029635 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs6nk\" (UniqueName: \"kubernetes.io/projected/bc76d2fb-7bba-48f7-9658-e55f0c225807-kube-api-access-gs6nk\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.131516 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.131580 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-inventory\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.131665 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.131706 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.131757 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.131808 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs6nk\" (UniqueName: \"kubernetes.io/projected/bc76d2fb-7bba-48f7-9658-e55f0c225807-kube-api-access-gs6nk\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.131915 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.687018 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.687105 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.687878 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.689524 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-inventory\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.689832 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.691025 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.692257 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs6nk\" (UniqueName: \"kubernetes.io/projected/bc76d2fb-7bba-48f7-9658-e55f0c225807-kube-api-access-gs6nk\") pod \"telemetry-openstack-openstack-cell1-z7lrc\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:02 crc kubenswrapper[4978]: I0225 09:26:02.780693 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:26:03 crc kubenswrapper[4978]: I0225 09:26:03.376075 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-z7lrc"] Feb 25 09:26:03 crc kubenswrapper[4978]: W0225 09:26:03.382561 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc76d2fb_7bba_48f7_9658_e55f0c225807.slice/crio-c5c8cd10ae13ddd500d43bc27801070ad6f0f0b2e617bc6ed9bae59229149a14 WatchSource:0}: Error finding container c5c8cd10ae13ddd500d43bc27801070ad6f0f0b2e617bc6ed9bae59229149a14: Status 404 returned error can't find the container with id c5c8cd10ae13ddd500d43bc27801070ad6f0f0b2e617bc6ed9bae59229149a14 Feb 25 09:26:03 crc kubenswrapper[4978]: I0225 09:26:03.520951 4978 generic.go:334] "Generic (PLEG): container finished" podID="c9d41622-d2f7-483a-90fa-00c2222fab68" containerID="ba392d1f7a7629d295f0d4a76332a2eea13bfcab82c8b9c704d81c5ed84041c0" exitCode=0 Feb 25 09:26:03 crc kubenswrapper[4978]: I0225 09:26:03.521055 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533526-6k9dz" event={"ID":"c9d41622-d2f7-483a-90fa-00c2222fab68","Type":"ContainerDied","Data":"ba392d1f7a7629d295f0d4a76332a2eea13bfcab82c8b9c704d81c5ed84041c0"} Feb 25 09:26:03 crc kubenswrapper[4978]: I0225 09:26:03.522949 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" event={"ID":"bc76d2fb-7bba-48f7-9658-e55f0c225807","Type":"ContainerStarted","Data":"c5c8cd10ae13ddd500d43bc27801070ad6f0f0b2e617bc6ed9bae59229149a14"} Feb 25 09:26:05 crc kubenswrapper[4978]: I0225 09:26:05.029399 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533526-6k9dz" Feb 25 09:26:05 crc kubenswrapper[4978]: I0225 09:26:05.194463 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxk4l\" (UniqueName: \"kubernetes.io/projected/c9d41622-d2f7-483a-90fa-00c2222fab68-kube-api-access-wxk4l\") pod \"c9d41622-d2f7-483a-90fa-00c2222fab68\" (UID: \"c9d41622-d2f7-483a-90fa-00c2222fab68\") " Feb 25 09:26:05 crc kubenswrapper[4978]: I0225 09:26:05.203760 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9d41622-d2f7-483a-90fa-00c2222fab68-kube-api-access-wxk4l" (OuterVolumeSpecName: "kube-api-access-wxk4l") pod "c9d41622-d2f7-483a-90fa-00c2222fab68" (UID: "c9d41622-d2f7-483a-90fa-00c2222fab68"). InnerVolumeSpecName "kube-api-access-wxk4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:26:05 crc kubenswrapper[4978]: I0225 09:26:05.301491 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxk4l\" (UniqueName: \"kubernetes.io/projected/c9d41622-d2f7-483a-90fa-00c2222fab68-kube-api-access-wxk4l\") on node \"crc\" DevicePath \"\"" Feb 25 09:26:05 crc kubenswrapper[4978]: I0225 09:26:05.549459 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533526-6k9dz" event={"ID":"c9d41622-d2f7-483a-90fa-00c2222fab68","Type":"ContainerDied","Data":"37e0943a14167071440306dad25a149c7c3304dd4b722ef9928fbf2e3e5c90fd"} Feb 25 09:26:05 crc kubenswrapper[4978]: I0225 09:26:05.549538 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37e0943a14167071440306dad25a149c7c3304dd4b722ef9928fbf2e3e5c90fd" Feb 25 09:26:05 crc kubenswrapper[4978]: I0225 09:26:05.549683 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533526-6k9dz" Feb 25 09:26:05 crc kubenswrapper[4978]: I0225 09:26:05.553992 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" event={"ID":"bc76d2fb-7bba-48f7-9658-e55f0c225807","Type":"ContainerStarted","Data":"8eef045326bcc5a995987c6d5d5aae11907578f026d7e687384e3fe7d3369184"} Feb 25 09:26:05 crc kubenswrapper[4978]: I0225 09:26:05.600065 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" podStartSLOduration=3.818916705 podStartE2EDuration="4.600034668s" podCreationTimestamp="2026-02-25 09:26:01 +0000 UTC" firstStartedPulling="2026-02-25 09:26:03.384727477 +0000 UTC m=+9656.823983936" lastFinishedPulling="2026-02-25 09:26:04.16584542 +0000 UTC m=+9657.605101899" observedRunningTime="2026-02-25 09:26:05.58011994 +0000 UTC m=+9659.019376459" watchObservedRunningTime="2026-02-25 09:26:05.600034668 +0000 UTC m=+9659.039291167" Feb 25 09:26:06 crc kubenswrapper[4978]: I0225 09:26:06.121001 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533520-ms2xm"] Feb 25 09:26:06 crc kubenswrapper[4978]: I0225 09:26:06.131570 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533520-ms2xm"] Feb 25 09:26:07 crc kubenswrapper[4978]: I0225 09:26:07.346595 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8b23b30-7e9c-474c-a663-ffdf14c61085" path="/var/lib/kubelet/pods/e8b23b30-7e9c-474c-a663-ffdf14c61085/volumes" Feb 25 09:26:16 crc kubenswrapper[4978]: I0225 09:26:16.540478 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:26:16 crc kubenswrapper[4978]: I0225 09:26:16.542625 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:26:16 crc kubenswrapper[4978]: I0225 09:26:16.542901 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 09:26:16 crc kubenswrapper[4978]: I0225 09:26:16.544192 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ccf3e1f2fc79ae5a3b75d3365e4d16775ede1da95dce83d44c7c298fd7c0a322"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 09:26:16 crc kubenswrapper[4978]: I0225 09:26:16.544524 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://ccf3e1f2fc79ae5a3b75d3365e4d16775ede1da95dce83d44c7c298fd7c0a322" gracePeriod=600 Feb 25 09:26:16 crc kubenswrapper[4978]: I0225 09:26:16.684221 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="ccf3e1f2fc79ae5a3b75d3365e4d16775ede1da95dce83d44c7c298fd7c0a322" exitCode=0 Feb 25 09:26:16 crc kubenswrapper[4978]: I0225 09:26:16.684258 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"ccf3e1f2fc79ae5a3b75d3365e4d16775ede1da95dce83d44c7c298fd7c0a322"} Feb 25 09:26:16 crc kubenswrapper[4978]: I0225 09:26:16.684892 4978 scope.go:117] "RemoveContainer" containerID="9182a657da785de05246b80f3e46a792de311ed3dbdfe2bbdb7f52feb2613cc4" Feb 25 09:26:17 crc kubenswrapper[4978]: I0225 09:26:17.702075 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec"} Feb 25 09:26:37 crc kubenswrapper[4978]: I0225 09:26:37.261050 4978 scope.go:117] "RemoveContainer" containerID="d37104d3ab4631454cffbeb4b63fb198b3ac792ebf9f978399d51c1fd526aee0" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.735152 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wshrn"] Feb 25 09:26:46 crc kubenswrapper[4978]: E0225 09:26:46.736089 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9d41622-d2f7-483a-90fa-00c2222fab68" containerName="oc" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.736102 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9d41622-d2f7-483a-90fa-00c2222fab68" containerName="oc" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.736323 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9d41622-d2f7-483a-90fa-00c2222fab68" containerName="oc" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.737837 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.749058 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wshrn"] Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.847876 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-utilities\") pod \"redhat-operators-wshrn\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.847969 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-catalog-content\") pod \"redhat-operators-wshrn\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.848118 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llq2n\" (UniqueName: \"kubernetes.io/projected/cefa364a-a80b-42bd-a354-7cb8bd53a90d-kube-api-access-llq2n\") pod \"redhat-operators-wshrn\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.949862 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-utilities\") pod \"redhat-operators-wshrn\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.949907 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-catalog-content\") pod \"redhat-operators-wshrn\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.950002 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llq2n\" (UniqueName: \"kubernetes.io/projected/cefa364a-a80b-42bd-a354-7cb8bd53a90d-kube-api-access-llq2n\") pod \"redhat-operators-wshrn\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.950320 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-utilities\") pod \"redhat-operators-wshrn\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.950423 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-catalog-content\") pod \"redhat-operators-wshrn\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:46 crc kubenswrapper[4978]: I0225 09:26:46.969045 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llq2n\" (UniqueName: \"kubernetes.io/projected/cefa364a-a80b-42bd-a354-7cb8bd53a90d-kube-api-access-llq2n\") pod \"redhat-operators-wshrn\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:47 crc kubenswrapper[4978]: I0225 09:26:47.091074 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:47 crc kubenswrapper[4978]: I0225 09:26:47.540747 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wshrn"] Feb 25 09:26:48 crc kubenswrapper[4978]: I0225 09:26:48.078182 4978 generic.go:334] "Generic (PLEG): container finished" podID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerID="738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36" exitCode=0 Feb 25 09:26:48 crc kubenswrapper[4978]: I0225 09:26:48.078226 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wshrn" event={"ID":"cefa364a-a80b-42bd-a354-7cb8bd53a90d","Type":"ContainerDied","Data":"738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36"} Feb 25 09:26:48 crc kubenswrapper[4978]: I0225 09:26:48.078252 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wshrn" event={"ID":"cefa364a-a80b-42bd-a354-7cb8bd53a90d","Type":"ContainerStarted","Data":"eb5f294f5a32f548725933def0f6984bc68b39531f401c1581879c1937fb48f7"} Feb 25 09:26:49 crc kubenswrapper[4978]: I0225 09:26:49.091299 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wshrn" event={"ID":"cefa364a-a80b-42bd-a354-7cb8bd53a90d","Type":"ContainerStarted","Data":"a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b"} Feb 25 09:26:52 crc kubenswrapper[4978]: I0225 09:26:52.137128 4978 generic.go:334] "Generic (PLEG): container finished" podID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerID="a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b" exitCode=0 Feb 25 09:26:52 crc kubenswrapper[4978]: I0225 09:26:52.137204 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wshrn" event={"ID":"cefa364a-a80b-42bd-a354-7cb8bd53a90d","Type":"ContainerDied","Data":"a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b"} Feb 25 09:26:53 crc kubenswrapper[4978]: I0225 09:26:53.149412 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wshrn" event={"ID":"cefa364a-a80b-42bd-a354-7cb8bd53a90d","Type":"ContainerStarted","Data":"2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827"} Feb 25 09:26:53 crc kubenswrapper[4978]: I0225 09:26:53.195161 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wshrn" podStartSLOduration=2.48798287 podStartE2EDuration="7.195137523s" podCreationTimestamp="2026-02-25 09:26:46 +0000 UTC" firstStartedPulling="2026-02-25 09:26:48.083302576 +0000 UTC m=+9701.522559035" lastFinishedPulling="2026-02-25 09:26:52.790457189 +0000 UTC m=+9706.229713688" observedRunningTime="2026-02-25 09:26:53.178822058 +0000 UTC m=+9706.618078547" watchObservedRunningTime="2026-02-25 09:26:53.195137523 +0000 UTC m=+9706.634393992" Feb 25 09:26:57 crc kubenswrapper[4978]: I0225 09:26:57.092118 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:57 crc kubenswrapper[4978]: I0225 09:26:57.092716 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:26:58 crc kubenswrapper[4978]: I0225 09:26:58.152971 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wshrn" podUID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerName="registry-server" probeResult="failure" output=< Feb 25 09:26:58 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 09:26:58 crc kubenswrapper[4978]: > Feb 25 09:27:07 crc kubenswrapper[4978]: I0225 09:27:07.183468 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:27:07 crc kubenswrapper[4978]: I0225 09:27:07.258600 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:27:07 crc kubenswrapper[4978]: I0225 09:27:07.435794 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wshrn"] Feb 25 09:27:08 crc kubenswrapper[4978]: I0225 09:27:08.325747 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wshrn" podUID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerName="registry-server" containerID="cri-o://2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827" gracePeriod=2 Feb 25 09:27:08 crc kubenswrapper[4978]: I0225 09:27:08.874972 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:27:08 crc kubenswrapper[4978]: I0225 09:27:08.999500 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llq2n\" (UniqueName: \"kubernetes.io/projected/cefa364a-a80b-42bd-a354-7cb8bd53a90d-kube-api-access-llq2n\") pod \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " Feb 25 09:27:08 crc kubenswrapper[4978]: I0225 09:27:08.999567 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-utilities\") pod \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:08.999669 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-catalog-content\") pod \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\" (UID: \"cefa364a-a80b-42bd-a354-7cb8bd53a90d\") " Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.001029 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-utilities" (OuterVolumeSpecName: "utilities") pod "cefa364a-a80b-42bd-a354-7cb8bd53a90d" (UID: "cefa364a-a80b-42bd-a354-7cb8bd53a90d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.010717 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cefa364a-a80b-42bd-a354-7cb8bd53a90d-kube-api-access-llq2n" (OuterVolumeSpecName: "kube-api-access-llq2n") pod "cefa364a-a80b-42bd-a354-7cb8bd53a90d" (UID: "cefa364a-a80b-42bd-a354-7cb8bd53a90d"). InnerVolumeSpecName "kube-api-access-llq2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.105825 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llq2n\" (UniqueName: \"kubernetes.io/projected/cefa364a-a80b-42bd-a354-7cb8bd53a90d-kube-api-access-llq2n\") on node \"crc\" DevicePath \"\"" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.106124 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.129237 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cefa364a-a80b-42bd-a354-7cb8bd53a90d" (UID: "cefa364a-a80b-42bd-a354-7cb8bd53a90d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.208135 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cefa364a-a80b-42bd-a354-7cb8bd53a90d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.358783 4978 generic.go:334] "Generic (PLEG): container finished" podID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerID="2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827" exitCode=0 Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.358859 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wshrn" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.358873 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wshrn" event={"ID":"cefa364a-a80b-42bd-a354-7cb8bd53a90d","Type":"ContainerDied","Data":"2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827"} Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.358903 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wshrn" event={"ID":"cefa364a-a80b-42bd-a354-7cb8bd53a90d","Type":"ContainerDied","Data":"eb5f294f5a32f548725933def0f6984bc68b39531f401c1581879c1937fb48f7"} Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.358922 4978 scope.go:117] "RemoveContainer" containerID="2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.382018 4978 scope.go:117] "RemoveContainer" containerID="a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.419953 4978 scope.go:117] "RemoveContainer" containerID="738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.421160 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wshrn"] Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.433078 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wshrn"] Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.465910 4978 scope.go:117] "RemoveContainer" containerID="2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827" Feb 25 09:27:09 crc kubenswrapper[4978]: E0225 09:27:09.466657 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827\": container with ID starting with 2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827 not found: ID does not exist" containerID="2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.466730 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827"} err="failed to get container status \"2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827\": rpc error: code = NotFound desc = could not find container \"2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827\": container with ID starting with 2cc160daa850d2232474cc64861d328c04ae3b6f6ff5bdcc588c49852951d827 not found: ID does not exist" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.466771 4978 scope.go:117] "RemoveContainer" containerID="a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b" Feb 25 09:27:09 crc kubenswrapper[4978]: E0225 09:27:09.469517 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b\": container with ID starting with a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b not found: ID does not exist" containerID="a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.469548 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b"} err="failed to get container status \"a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b\": rpc error: code = NotFound desc = could not find container \"a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b\": container with ID starting with a884141d3915f5c64e13a79449d11ea3ea17cd9c59eadf104198aba035d2d58b not found: ID does not exist" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.469566 4978 scope.go:117] "RemoveContainer" containerID="738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36" Feb 25 09:27:09 crc kubenswrapper[4978]: E0225 09:27:09.469926 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36\": container with ID starting with 738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36 not found: ID does not exist" containerID="738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36" Feb 25 09:27:09 crc kubenswrapper[4978]: I0225 09:27:09.469980 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36"} err="failed to get container status \"738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36\": rpc error: code = NotFound desc = could not find container \"738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36\": container with ID starting with 738c931b0e93e7b3114959173995fab071444ec91ecbd8c50c19209889df7a36 not found: ID does not exist" Feb 25 09:27:11 crc kubenswrapper[4978]: I0225 09:27:11.341285 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" path="/var/lib/kubelet/pods/cefa364a-a80b-42bd-a354-7cb8bd53a90d/volumes" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.169016 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533528-jc4rq"] Feb 25 09:28:00 crc kubenswrapper[4978]: E0225 09:28:00.170034 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerName="extract-content" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.170050 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerName="extract-content" Feb 25 09:28:00 crc kubenswrapper[4978]: E0225 09:28:00.170092 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerName="extract-utilities" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.170101 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerName="extract-utilities" Feb 25 09:28:00 crc kubenswrapper[4978]: E0225 09:28:00.170118 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerName="registry-server" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.170127 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerName="registry-server" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.170399 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="cefa364a-a80b-42bd-a354-7cb8bd53a90d" containerName="registry-server" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.171257 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533528-jc4rq" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.190388 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.190486 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.192801 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.211030 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533528-jc4rq"] Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.231069 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv6qf\" (UniqueName: \"kubernetes.io/projected/bc30ddfa-fe50-42b3-b97e-0967a0b69363-kube-api-access-rv6qf\") pod \"auto-csr-approver-29533528-jc4rq\" (UID: \"bc30ddfa-fe50-42b3-b97e-0967a0b69363\") " pod="openshift-infra/auto-csr-approver-29533528-jc4rq" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.333130 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv6qf\" (UniqueName: \"kubernetes.io/projected/bc30ddfa-fe50-42b3-b97e-0967a0b69363-kube-api-access-rv6qf\") pod \"auto-csr-approver-29533528-jc4rq\" (UID: \"bc30ddfa-fe50-42b3-b97e-0967a0b69363\") " pod="openshift-infra/auto-csr-approver-29533528-jc4rq" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.353837 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv6qf\" (UniqueName: \"kubernetes.io/projected/bc30ddfa-fe50-42b3-b97e-0967a0b69363-kube-api-access-rv6qf\") pod \"auto-csr-approver-29533528-jc4rq\" (UID: \"bc30ddfa-fe50-42b3-b97e-0967a0b69363\") " pod="openshift-infra/auto-csr-approver-29533528-jc4rq" Feb 25 09:28:00 crc kubenswrapper[4978]: I0225 09:28:00.511392 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533528-jc4rq" Feb 25 09:28:01 crc kubenswrapper[4978]: I0225 09:28:01.034496 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:28:01 crc kubenswrapper[4978]: I0225 09:28:01.040331 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533528-jc4rq"] Feb 25 09:28:01 crc kubenswrapper[4978]: I0225 09:28:01.568624 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533528-jc4rq" event={"ID":"bc30ddfa-fe50-42b3-b97e-0967a0b69363","Type":"ContainerStarted","Data":"bdea8e7f106c8d47a5cb1f9924def45b861dde2f0ca0ecfa9baa037cd77c42c6"} Feb 25 09:28:03 crc kubenswrapper[4978]: I0225 09:28:03.597944 4978 generic.go:334] "Generic (PLEG): container finished" podID="bc30ddfa-fe50-42b3-b97e-0967a0b69363" containerID="4aa507cfce53546fef515a839b98038defd2d4a02f451bc73c592d741c1e0417" exitCode=0 Feb 25 09:28:03 crc kubenswrapper[4978]: I0225 09:28:03.598053 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533528-jc4rq" event={"ID":"bc30ddfa-fe50-42b3-b97e-0967a0b69363","Type":"ContainerDied","Data":"4aa507cfce53546fef515a839b98038defd2d4a02f451bc73c592d741c1e0417"} Feb 25 09:28:05 crc kubenswrapper[4978]: I0225 09:28:05.028010 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533528-jc4rq" Feb 25 09:28:05 crc kubenswrapper[4978]: I0225 09:28:05.165901 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rv6qf\" (UniqueName: \"kubernetes.io/projected/bc30ddfa-fe50-42b3-b97e-0967a0b69363-kube-api-access-rv6qf\") pod \"bc30ddfa-fe50-42b3-b97e-0967a0b69363\" (UID: \"bc30ddfa-fe50-42b3-b97e-0967a0b69363\") " Feb 25 09:28:05 crc kubenswrapper[4978]: I0225 09:28:05.175905 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc30ddfa-fe50-42b3-b97e-0967a0b69363-kube-api-access-rv6qf" (OuterVolumeSpecName: "kube-api-access-rv6qf") pod "bc30ddfa-fe50-42b3-b97e-0967a0b69363" (UID: "bc30ddfa-fe50-42b3-b97e-0967a0b69363"). InnerVolumeSpecName "kube-api-access-rv6qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:28:05 crc kubenswrapper[4978]: I0225 09:28:05.268646 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rv6qf\" (UniqueName: \"kubernetes.io/projected/bc30ddfa-fe50-42b3-b97e-0967a0b69363-kube-api-access-rv6qf\") on node \"crc\" DevicePath \"\"" Feb 25 09:28:05 crc kubenswrapper[4978]: I0225 09:28:05.622704 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533528-jc4rq" event={"ID":"bc30ddfa-fe50-42b3-b97e-0967a0b69363","Type":"ContainerDied","Data":"bdea8e7f106c8d47a5cb1f9924def45b861dde2f0ca0ecfa9baa037cd77c42c6"} Feb 25 09:28:05 crc kubenswrapper[4978]: I0225 09:28:05.622768 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdea8e7f106c8d47a5cb1f9924def45b861dde2f0ca0ecfa9baa037cd77c42c6" Feb 25 09:28:05 crc kubenswrapper[4978]: I0225 09:28:05.622826 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533528-jc4rq" Feb 25 09:28:06 crc kubenswrapper[4978]: I0225 09:28:06.121357 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533522-zhsm2"] Feb 25 09:28:06 crc kubenswrapper[4978]: I0225 09:28:06.139099 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533522-zhsm2"] Feb 25 09:28:07 crc kubenswrapper[4978]: I0225 09:28:07.348929 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1159a8fb-8eca-48a1-9f7a-b01a5d027206" path="/var/lib/kubelet/pods/1159a8fb-8eca-48a1-9f7a-b01a5d027206/volumes" Feb 25 09:28:16 crc kubenswrapper[4978]: I0225 09:28:16.541498 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:28:16 crc kubenswrapper[4978]: I0225 09:28:16.543725 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:28:37 crc kubenswrapper[4978]: I0225 09:28:37.416883 4978 scope.go:117] "RemoveContainer" containerID="6bdce94aa874c75a28a7d62b34953625d11d56ad0107a56b72908133ec3803c0" Feb 25 09:28:46 crc kubenswrapper[4978]: I0225 09:28:46.540822 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:28:46 crc kubenswrapper[4978]: I0225 09:28:46.541791 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:29:16 crc kubenswrapper[4978]: I0225 09:29:16.541147 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:29:16 crc kubenswrapper[4978]: I0225 09:29:16.541864 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:29:16 crc kubenswrapper[4978]: I0225 09:29:16.541947 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 09:29:16 crc kubenswrapper[4978]: I0225 09:29:16.543326 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 09:29:16 crc kubenswrapper[4978]: I0225 09:29:16.543494 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" gracePeriod=600 Feb 25 09:29:16 crc kubenswrapper[4978]: E0225 09:29:16.700136 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:29:17 crc kubenswrapper[4978]: I0225 09:29:17.552109 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" exitCode=0 Feb 25 09:29:17 crc kubenswrapper[4978]: I0225 09:29:17.552193 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec"} Feb 25 09:29:17 crc kubenswrapper[4978]: I0225 09:29:17.552560 4978 scope.go:117] "RemoveContainer" containerID="ccf3e1f2fc79ae5a3b75d3365e4d16775ede1da95dce83d44c7c298fd7c0a322" Feb 25 09:29:17 crc kubenswrapper[4978]: I0225 09:29:17.553595 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:29:17 crc kubenswrapper[4978]: E0225 09:29:17.554109 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:29:31 crc kubenswrapper[4978]: I0225 09:29:31.327490 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:29:31 crc kubenswrapper[4978]: E0225 09:29:31.328210 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.381134 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f8xr9"] Feb 25 09:29:35 crc kubenswrapper[4978]: E0225 09:29:35.382452 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc30ddfa-fe50-42b3-b97e-0967a0b69363" containerName="oc" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.382472 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc30ddfa-fe50-42b3-b97e-0967a0b69363" containerName="oc" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.382755 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc30ddfa-fe50-42b3-b97e-0967a0b69363" containerName="oc" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.384806 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.408986 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f8xr9"] Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.485044 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-catalog-content\") pod \"certified-operators-f8xr9\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.485611 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-utilities\") pod \"certified-operators-f8xr9\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.485766 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl4ks\" (UniqueName: \"kubernetes.io/projected/5a216121-86b8-4454-9ce0-c1f8660c2842-kube-api-access-fl4ks\") pod \"certified-operators-f8xr9\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.573677 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zlpxk"] Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.575691 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.587750 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-utilities\") pod \"certified-operators-f8xr9\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.587810 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl4ks\" (UniqueName: \"kubernetes.io/projected/5a216121-86b8-4454-9ce0-c1f8660c2842-kube-api-access-fl4ks\") pod \"certified-operators-f8xr9\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.587879 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-catalog-content\") pod \"certified-operators-f8xr9\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.588480 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-catalog-content\") pod \"certified-operators-f8xr9\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.588495 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-utilities\") pod \"certified-operators-f8xr9\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.596049 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zlpxk"] Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.610735 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl4ks\" (UniqueName: \"kubernetes.io/projected/5a216121-86b8-4454-9ce0-c1f8660c2842-kube-api-access-fl4ks\") pod \"certified-operators-f8xr9\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.690181 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4g27\" (UniqueName: \"kubernetes.io/projected/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-kube-api-access-l4g27\") pod \"community-operators-zlpxk\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.690415 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-utilities\") pod \"community-operators-zlpxk\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.690853 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-catalog-content\") pod \"community-operators-zlpxk\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.708730 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.792131 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-catalog-content\") pod \"community-operators-zlpxk\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.792263 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4g27\" (UniqueName: \"kubernetes.io/projected/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-kube-api-access-l4g27\") pod \"community-operators-zlpxk\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.792654 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-utilities\") pod \"community-operators-zlpxk\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.792706 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-catalog-content\") pod \"community-operators-zlpxk\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.793061 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-utilities\") pod \"community-operators-zlpxk\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.812191 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4g27\" (UniqueName: \"kubernetes.io/projected/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-kube-api-access-l4g27\") pod \"community-operators-zlpxk\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:35 crc kubenswrapper[4978]: I0225 09:29:35.893547 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:36 crc kubenswrapper[4978]: I0225 09:29:36.491830 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f8xr9"] Feb 25 09:29:36 crc kubenswrapper[4978]: I0225 09:29:36.655296 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zlpxk"] Feb 25 09:29:36 crc kubenswrapper[4978]: I0225 09:29:36.765100 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlpxk" event={"ID":"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9","Type":"ContainerStarted","Data":"8803fc2e4ae0ee32a2d16f2a50c07ac4984aef91ba76ba6ffe08093c41a058c2"} Feb 25 09:29:36 crc kubenswrapper[4978]: I0225 09:29:36.766099 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8xr9" event={"ID":"5a216121-86b8-4454-9ce0-c1f8660c2842","Type":"ContainerStarted","Data":"d157d34c1214af5899815bf00482cac3baa6de8c5c92e1cbb9ff323a4238be1c"} Feb 25 09:29:37 crc kubenswrapper[4978]: I0225 09:29:37.778003 4978 generic.go:334] "Generic (PLEG): container finished" podID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerID="43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6" exitCode=0 Feb 25 09:29:37 crc kubenswrapper[4978]: I0225 09:29:37.778103 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlpxk" event={"ID":"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9","Type":"ContainerDied","Data":"43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6"} Feb 25 09:29:37 crc kubenswrapper[4978]: I0225 09:29:37.780359 4978 generic.go:334] "Generic (PLEG): container finished" podID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerID="4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b" exitCode=0 Feb 25 09:29:37 crc kubenswrapper[4978]: I0225 09:29:37.780443 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8xr9" event={"ID":"5a216121-86b8-4454-9ce0-c1f8660c2842","Type":"ContainerDied","Data":"4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b"} Feb 25 09:29:37 crc kubenswrapper[4978]: I0225 09:29:37.975601 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vgt8p"] Feb 25 09:29:37 crc kubenswrapper[4978]: I0225 09:29:37.978218 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:37 crc kubenswrapper[4978]: I0225 09:29:37.987288 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vgt8p"] Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.001542 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-utilities\") pod \"redhat-marketplace-vgt8p\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.001975 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-catalog-content\") pod \"redhat-marketplace-vgt8p\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.002035 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c62wx\" (UniqueName: \"kubernetes.io/projected/057276cb-c97b-4baa-941e-2397a8462a04-kube-api-access-c62wx\") pod \"redhat-marketplace-vgt8p\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.104040 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-catalog-content\") pod \"redhat-marketplace-vgt8p\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.104094 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c62wx\" (UniqueName: \"kubernetes.io/projected/057276cb-c97b-4baa-941e-2397a8462a04-kube-api-access-c62wx\") pod \"redhat-marketplace-vgt8p\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.104203 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-utilities\") pod \"redhat-marketplace-vgt8p\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.105787 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-utilities\") pod \"redhat-marketplace-vgt8p\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.105799 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-catalog-content\") pod \"redhat-marketplace-vgt8p\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.132139 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c62wx\" (UniqueName: \"kubernetes.io/projected/057276cb-c97b-4baa-941e-2397a8462a04-kube-api-access-c62wx\") pod \"redhat-marketplace-vgt8p\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.303143 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:38 crc kubenswrapper[4978]: I0225 09:29:38.788735 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vgt8p"] Feb 25 09:29:39 crc kubenswrapper[4978]: I0225 09:29:39.802176 4978 generic.go:334] "Generic (PLEG): container finished" podID="057276cb-c97b-4baa-941e-2397a8462a04" containerID="54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad" exitCode=0 Feb 25 09:29:39 crc kubenswrapper[4978]: I0225 09:29:39.802258 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgt8p" event={"ID":"057276cb-c97b-4baa-941e-2397a8462a04","Type":"ContainerDied","Data":"54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad"} Feb 25 09:29:39 crc kubenswrapper[4978]: I0225 09:29:39.803599 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgt8p" event={"ID":"057276cb-c97b-4baa-941e-2397a8462a04","Type":"ContainerStarted","Data":"62a38d1e5aa552efc941c138afb34c59423c197734c4e0f83b73d5f85b03449d"} Feb 25 09:29:39 crc kubenswrapper[4978]: I0225 09:29:39.807249 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlpxk" event={"ID":"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9","Type":"ContainerStarted","Data":"707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f"} Feb 25 09:29:39 crc kubenswrapper[4978]: I0225 09:29:39.809442 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8xr9" event={"ID":"5a216121-86b8-4454-9ce0-c1f8660c2842","Type":"ContainerStarted","Data":"3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c"} Feb 25 09:29:40 crc kubenswrapper[4978]: I0225 09:29:40.825172 4978 generic.go:334] "Generic (PLEG): container finished" podID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerID="707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f" exitCode=0 Feb 25 09:29:40 crc kubenswrapper[4978]: I0225 09:29:40.826556 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlpxk" event={"ID":"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9","Type":"ContainerDied","Data":"707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f"} Feb 25 09:29:40 crc kubenswrapper[4978]: I0225 09:29:40.832219 4978 generic.go:334] "Generic (PLEG): container finished" podID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerID="3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c" exitCode=0 Feb 25 09:29:40 crc kubenswrapper[4978]: I0225 09:29:40.832303 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8xr9" event={"ID":"5a216121-86b8-4454-9ce0-c1f8660c2842","Type":"ContainerDied","Data":"3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c"} Feb 25 09:29:40 crc kubenswrapper[4978]: I0225 09:29:40.836625 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgt8p" event={"ID":"057276cb-c97b-4baa-941e-2397a8462a04","Type":"ContainerStarted","Data":"e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178"} Feb 25 09:29:42 crc kubenswrapper[4978]: I0225 09:29:42.876083 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlpxk" event={"ID":"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9","Type":"ContainerStarted","Data":"779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318"} Feb 25 09:29:42 crc kubenswrapper[4978]: I0225 09:29:42.880636 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8xr9" event={"ID":"5a216121-86b8-4454-9ce0-c1f8660c2842","Type":"ContainerStarted","Data":"25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13"} Feb 25 09:29:42 crc kubenswrapper[4978]: I0225 09:29:42.884633 4978 generic.go:334] "Generic (PLEG): container finished" podID="057276cb-c97b-4baa-941e-2397a8462a04" containerID="e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178" exitCode=0 Feb 25 09:29:42 crc kubenswrapper[4978]: I0225 09:29:42.884688 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgt8p" event={"ID":"057276cb-c97b-4baa-941e-2397a8462a04","Type":"ContainerDied","Data":"e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178"} Feb 25 09:29:42 crc kubenswrapper[4978]: I0225 09:29:42.917309 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zlpxk" podStartSLOduration=4.296660588 podStartE2EDuration="7.917277311s" podCreationTimestamp="2026-02-25 09:29:35 +0000 UTC" firstStartedPulling="2026-02-25 09:29:37.781054597 +0000 UTC m=+9871.220311096" lastFinishedPulling="2026-02-25 09:29:41.40167136 +0000 UTC m=+9874.840927819" observedRunningTime="2026-02-25 09:29:42.903566576 +0000 UTC m=+9876.342823085" watchObservedRunningTime="2026-02-25 09:29:42.917277311 +0000 UTC m=+9876.356533770" Feb 25 09:29:42 crc kubenswrapper[4978]: I0225 09:29:42.954222 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f8xr9" podStartSLOduration=3.986985749 podStartE2EDuration="7.954205975s" podCreationTimestamp="2026-02-25 09:29:35 +0000 UTC" firstStartedPulling="2026-02-25 09:29:37.781990877 +0000 UTC m=+9871.221247376" lastFinishedPulling="2026-02-25 09:29:41.749211133 +0000 UTC m=+9875.188467602" observedRunningTime="2026-02-25 09:29:42.948849179 +0000 UTC m=+9876.388105648" watchObservedRunningTime="2026-02-25 09:29:42.954205975 +0000 UTC m=+9876.393462434" Feb 25 09:29:43 crc kubenswrapper[4978]: I0225 09:29:43.332250 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:29:43 crc kubenswrapper[4978]: E0225 09:29:43.332583 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:29:43 crc kubenswrapper[4978]: I0225 09:29:43.896821 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgt8p" event={"ID":"057276cb-c97b-4baa-941e-2397a8462a04","Type":"ContainerStarted","Data":"faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b"} Feb 25 09:29:43 crc kubenswrapper[4978]: I0225 09:29:43.919317 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vgt8p" podStartSLOduration=3.4247388069999998 podStartE2EDuration="6.919302392s" podCreationTimestamp="2026-02-25 09:29:37 +0000 UTC" firstStartedPulling="2026-02-25 09:29:39.803715326 +0000 UTC m=+9873.242971795" lastFinishedPulling="2026-02-25 09:29:43.298278911 +0000 UTC m=+9876.737535380" observedRunningTime="2026-02-25 09:29:43.916216496 +0000 UTC m=+9877.355472985" watchObservedRunningTime="2026-02-25 09:29:43.919302392 +0000 UTC m=+9877.358558851" Feb 25 09:29:45 crc kubenswrapper[4978]: I0225 09:29:45.709894 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:45 crc kubenswrapper[4978]: I0225 09:29:45.710300 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:45 crc kubenswrapper[4978]: I0225 09:29:45.784846 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:45 crc kubenswrapper[4978]: I0225 09:29:45.894472 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:45 crc kubenswrapper[4978]: I0225 09:29:45.894607 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:45 crc kubenswrapper[4978]: I0225 09:29:45.949834 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:48 crc kubenswrapper[4978]: I0225 09:29:48.310514 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:48 crc kubenswrapper[4978]: I0225 09:29:48.310625 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:48 crc kubenswrapper[4978]: I0225 09:29:48.394685 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:49 crc kubenswrapper[4978]: I0225 09:29:49.024570 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:49 crc kubenswrapper[4978]: I0225 09:29:49.769996 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vgt8p"] Feb 25 09:29:50 crc kubenswrapper[4978]: I0225 09:29:50.983146 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vgt8p" podUID="057276cb-c97b-4baa-941e-2397a8462a04" containerName="registry-server" containerID="cri-o://faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b" gracePeriod=2 Feb 25 09:29:51 crc kubenswrapper[4978]: I0225 09:29:51.917603 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:51 crc kubenswrapper[4978]: I0225 09:29:51.994251 4978 generic.go:334] "Generic (PLEG): container finished" podID="057276cb-c97b-4baa-941e-2397a8462a04" containerID="faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b" exitCode=0 Feb 25 09:29:51 crc kubenswrapper[4978]: I0225 09:29:51.994295 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgt8p" event={"ID":"057276cb-c97b-4baa-941e-2397a8462a04","Type":"ContainerDied","Data":"faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b"} Feb 25 09:29:51 crc kubenswrapper[4978]: I0225 09:29:51.994307 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vgt8p" Feb 25 09:29:51 crc kubenswrapper[4978]: I0225 09:29:51.994322 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vgt8p" event={"ID":"057276cb-c97b-4baa-941e-2397a8462a04","Type":"ContainerDied","Data":"62a38d1e5aa552efc941c138afb34c59423c197734c4e0f83b73d5f85b03449d"} Feb 25 09:29:51 crc kubenswrapper[4978]: I0225 09:29:51.994342 4978 scope.go:117] "RemoveContainer" containerID="faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.033655 4978 scope.go:117] "RemoveContainer" containerID="e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.043425 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c62wx\" (UniqueName: \"kubernetes.io/projected/057276cb-c97b-4baa-941e-2397a8462a04-kube-api-access-c62wx\") pod \"057276cb-c97b-4baa-941e-2397a8462a04\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.043657 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-utilities\") pod \"057276cb-c97b-4baa-941e-2397a8462a04\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.043863 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-catalog-content\") pod \"057276cb-c97b-4baa-941e-2397a8462a04\" (UID: \"057276cb-c97b-4baa-941e-2397a8462a04\") " Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.045306 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-utilities" (OuterVolumeSpecName: "utilities") pod "057276cb-c97b-4baa-941e-2397a8462a04" (UID: "057276cb-c97b-4baa-941e-2397a8462a04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.054720 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057276cb-c97b-4baa-941e-2397a8462a04-kube-api-access-c62wx" (OuterVolumeSpecName: "kube-api-access-c62wx") pod "057276cb-c97b-4baa-941e-2397a8462a04" (UID: "057276cb-c97b-4baa-941e-2397a8462a04"). InnerVolumeSpecName "kube-api-access-c62wx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.064483 4978 scope.go:117] "RemoveContainer" containerID="54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.088669 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "057276cb-c97b-4baa-941e-2397a8462a04" (UID: "057276cb-c97b-4baa-941e-2397a8462a04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.146036 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.146069 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c62wx\" (UniqueName: \"kubernetes.io/projected/057276cb-c97b-4baa-941e-2397a8462a04-kube-api-access-c62wx\") on node \"crc\" DevicePath \"\"" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.146085 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/057276cb-c97b-4baa-941e-2397a8462a04-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.150750 4978 scope.go:117] "RemoveContainer" containerID="faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b" Feb 25 09:29:52 crc kubenswrapper[4978]: E0225 09:29:52.151420 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b\": container with ID starting with faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b not found: ID does not exist" containerID="faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.151468 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b"} err="failed to get container status \"faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b\": rpc error: code = NotFound desc = could not find container \"faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b\": container with ID starting with faf72f963577bd4892dd88f8879d9739482b69ab03e45aff26d78fe94954002b not found: ID does not exist" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.151501 4978 scope.go:117] "RemoveContainer" containerID="e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178" Feb 25 09:29:52 crc kubenswrapper[4978]: E0225 09:29:52.152048 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178\": container with ID starting with e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178 not found: ID does not exist" containerID="e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.152078 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178"} err="failed to get container status \"e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178\": rpc error: code = NotFound desc = could not find container \"e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178\": container with ID starting with e63b0b10b3e9c2edbfe2fb41261e9451a585313bbb606954e1992a5551c2d178 not found: ID does not exist" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.152099 4978 scope.go:117] "RemoveContainer" containerID="54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad" Feb 25 09:29:52 crc kubenswrapper[4978]: E0225 09:29:52.152488 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad\": container with ID starting with 54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad not found: ID does not exist" containerID="54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.152524 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad"} err="failed to get container status \"54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad\": rpc error: code = NotFound desc = could not find container \"54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad\": container with ID starting with 54d146ccab047a41c8a1f2fa1aaa9519ea99ccd21d1e5b6744c4c694c42c6aad not found: ID does not exist" Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.340278 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vgt8p"] Feb 25 09:29:52 crc kubenswrapper[4978]: I0225 09:29:52.351915 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vgt8p"] Feb 25 09:29:53 crc kubenswrapper[4978]: I0225 09:29:53.353222 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="057276cb-c97b-4baa-941e-2397a8462a04" path="/var/lib/kubelet/pods/057276cb-c97b-4baa-941e-2397a8462a04/volumes" Feb 25 09:29:55 crc kubenswrapper[4978]: I0225 09:29:55.789200 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:55 crc kubenswrapper[4978]: I0225 09:29:55.950871 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:57 crc kubenswrapper[4978]: I0225 09:29:57.339579 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:29:57 crc kubenswrapper[4978]: E0225 09:29:57.340117 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:29:58 crc kubenswrapper[4978]: I0225 09:29:58.165520 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f8xr9"] Feb 25 09:29:58 crc kubenswrapper[4978]: I0225 09:29:58.166167 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f8xr9" podUID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerName="registry-server" containerID="cri-o://25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13" gracePeriod=2 Feb 25 09:29:58 crc kubenswrapper[4978]: I0225 09:29:58.736723 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:58 crc kubenswrapper[4978]: I0225 09:29:58.915027 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-catalog-content\") pod \"5a216121-86b8-4454-9ce0-c1f8660c2842\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " Feb 25 09:29:58 crc kubenswrapper[4978]: I0225 09:29:58.915518 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl4ks\" (UniqueName: \"kubernetes.io/projected/5a216121-86b8-4454-9ce0-c1f8660c2842-kube-api-access-fl4ks\") pod \"5a216121-86b8-4454-9ce0-c1f8660c2842\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " Feb 25 09:29:58 crc kubenswrapper[4978]: I0225 09:29:58.915574 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-utilities\") pod \"5a216121-86b8-4454-9ce0-c1f8660c2842\" (UID: \"5a216121-86b8-4454-9ce0-c1f8660c2842\") " Feb 25 09:29:58 crc kubenswrapper[4978]: I0225 09:29:58.916947 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-utilities" (OuterVolumeSpecName: "utilities") pod "5a216121-86b8-4454-9ce0-c1f8660c2842" (UID: "5a216121-86b8-4454-9ce0-c1f8660c2842"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:29:58 crc kubenswrapper[4978]: I0225 09:29:58.927787 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a216121-86b8-4454-9ce0-c1f8660c2842-kube-api-access-fl4ks" (OuterVolumeSpecName: "kube-api-access-fl4ks") pod "5a216121-86b8-4454-9ce0-c1f8660c2842" (UID: "5a216121-86b8-4454-9ce0-c1f8660c2842"). InnerVolumeSpecName "kube-api-access-fl4ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:29:58 crc kubenswrapper[4978]: I0225 09:29:58.962414 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5a216121-86b8-4454-9ce0-c1f8660c2842" (UID: "5a216121-86b8-4454-9ce0-c1f8660c2842"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.017755 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl4ks\" (UniqueName: \"kubernetes.io/projected/5a216121-86b8-4454-9ce0-c1f8660c2842-kube-api-access-fl4ks\") on node \"crc\" DevicePath \"\"" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.018032 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.018094 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a216121-86b8-4454-9ce0-c1f8660c2842-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.071397 4978 generic.go:334] "Generic (PLEG): container finished" podID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerID="25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13" exitCode=0 Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.071467 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8xr9" event={"ID":"5a216121-86b8-4454-9ce0-c1f8660c2842","Type":"ContainerDied","Data":"25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13"} Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.071544 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f8xr9" event={"ID":"5a216121-86b8-4454-9ce0-c1f8660c2842","Type":"ContainerDied","Data":"d157d34c1214af5899815bf00482cac3baa6de8c5c92e1cbb9ff323a4238be1c"} Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.071585 4978 scope.go:117] "RemoveContainer" containerID="25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.071874 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f8xr9" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.122781 4978 scope.go:117] "RemoveContainer" containerID="3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.146540 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f8xr9"] Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.171003 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f8xr9"] Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.182582 4978 scope.go:117] "RemoveContainer" containerID="4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.196678 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zlpxk"] Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.196968 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zlpxk" podUID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerName="registry-server" containerID="cri-o://779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318" gracePeriod=2 Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.211396 4978 scope.go:117] "RemoveContainer" containerID="25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13" Feb 25 09:29:59 crc kubenswrapper[4978]: E0225 09:29:59.211899 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13\": container with ID starting with 25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13 not found: ID does not exist" containerID="25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.211939 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13"} err="failed to get container status \"25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13\": rpc error: code = NotFound desc = could not find container \"25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13\": container with ID starting with 25303151f7593be26510cc9462547d00af2be967414bb5005edba71a4e582d13 not found: ID does not exist" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.211966 4978 scope.go:117] "RemoveContainer" containerID="3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c" Feb 25 09:29:59 crc kubenswrapper[4978]: E0225 09:29:59.212215 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c\": container with ID starting with 3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c not found: ID does not exist" containerID="3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.212241 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c"} err="failed to get container status \"3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c\": rpc error: code = NotFound desc = could not find container \"3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c\": container with ID starting with 3901555ab93982528809e7532f77e3ae0b97ba5a664a442c6c9e344287d9da2c not found: ID does not exist" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.212256 4978 scope.go:117] "RemoveContainer" containerID="4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b" Feb 25 09:29:59 crc kubenswrapper[4978]: E0225 09:29:59.212455 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b\": container with ID starting with 4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b not found: ID does not exist" containerID="4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.212484 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b"} err="failed to get container status \"4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b\": rpc error: code = NotFound desc = could not find container \"4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b\": container with ID starting with 4ee0b73a6452ccd38868e64d4fefeb998947cb2902d46234a62e59528af5dc8b not found: ID does not exist" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.340346 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a216121-86b8-4454-9ce0-c1f8660c2842" path="/var/lib/kubelet/pods/5a216121-86b8-4454-9ce0-c1f8660c2842/volumes" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.675641 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.759566 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-utilities\") pod \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.759794 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-catalog-content\") pod \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.759942 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4g27\" (UniqueName: \"kubernetes.io/projected/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-kube-api-access-l4g27\") pod \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\" (UID: \"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9\") " Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.760766 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-utilities" (OuterVolumeSpecName: "utilities") pod "a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" (UID: "a7f9cc97-c95a-4de5-a563-b2a73d5db1c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.764637 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-kube-api-access-l4g27" (OuterVolumeSpecName: "kube-api-access-l4g27") pod "a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" (UID: "a7f9cc97-c95a-4de5-a563-b2a73d5db1c9"). InnerVolumeSpecName "kube-api-access-l4g27". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.809777 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" (UID: "a7f9cc97-c95a-4de5-a563-b2a73d5db1c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.863415 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.863452 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4g27\" (UniqueName: \"kubernetes.io/projected/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-kube-api-access-l4g27\") on node \"crc\" DevicePath \"\"" Feb 25 09:29:59 crc kubenswrapper[4978]: I0225 09:29:59.863465 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.099575 4978 generic.go:334] "Generic (PLEG): container finished" podID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerID="779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318" exitCode=0 Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.099634 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlpxk" event={"ID":"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9","Type":"ContainerDied","Data":"779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318"} Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.099661 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlpxk" event={"ID":"a7f9cc97-c95a-4de5-a563-b2a73d5db1c9","Type":"ContainerDied","Data":"8803fc2e4ae0ee32a2d16f2a50c07ac4984aef91ba76ba6ffe08093c41a058c2"} Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.099693 4978 scope.go:117] "RemoveContainer" containerID="779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.099909 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlpxk" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.137459 4978 scope.go:117] "RemoveContainer" containerID="707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.173591 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr"] Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.174350 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerName="extract-utilities" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.174394 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerName="extract-utilities" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.174415 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerName="registry-server" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.174424 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerName="registry-server" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.174441 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerName="extract-content" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.174449 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerName="extract-content" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.174483 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerName="extract-utilities" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.174494 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerName="extract-utilities" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.174508 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057276cb-c97b-4baa-941e-2397a8462a04" containerName="extract-content" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.174519 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="057276cb-c97b-4baa-941e-2397a8462a04" containerName="extract-content" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.174534 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057276cb-c97b-4baa-941e-2397a8462a04" containerName="extract-utilities" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.174561 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="057276cb-c97b-4baa-941e-2397a8462a04" containerName="extract-utilities" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.174578 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057276cb-c97b-4baa-941e-2397a8462a04" containerName="registry-server" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.174587 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="057276cb-c97b-4baa-941e-2397a8462a04" containerName="registry-server" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.174607 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerName="extract-content" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.174615 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerName="extract-content" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.174657 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerName="registry-server" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.174668 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerName="registry-server" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.175966 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" containerName="registry-server" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.176036 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="057276cb-c97b-4baa-941e-2397a8462a04" containerName="registry-server" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.176053 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a216121-86b8-4454-9ce0-c1f8660c2842" containerName="registry-server" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.177421 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.188467 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.188778 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.194055 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533530-xqkpb"] Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.195679 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533530-xqkpb" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.198535 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.199459 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.199644 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.199958 4978 scope.go:117] "RemoveContainer" containerID="43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.208537 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zlpxk"] Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.219902 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zlpxk"] Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.231170 4978 scope.go:117] "RemoveContainer" containerID="779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.231585 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318\": container with ID starting with 779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318 not found: ID does not exist" containerID="779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.231618 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318"} err="failed to get container status \"779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318\": rpc error: code = NotFound desc = could not find container \"779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318\": container with ID starting with 779426ea4873c2306a47c12257a216fe17d5fd28821d5618e01db31d34c2d318 not found: ID does not exist" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.231639 4978 scope.go:117] "RemoveContainer" containerID="707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.231974 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f\": container with ID starting with 707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f not found: ID does not exist" containerID="707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.232001 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f"} err="failed to get container status \"707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f\": rpc error: code = NotFound desc = could not find container \"707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f\": container with ID starting with 707c7c6703c368c4ca9e4e4adf3012edd643ee7bf527572d982c84a9b61e834f not found: ID does not exist" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.232016 4978 scope.go:117] "RemoveContainer" containerID="43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6" Feb 25 09:30:00 crc kubenswrapper[4978]: E0225 09:30:00.232229 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6\": container with ID starting with 43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6 not found: ID does not exist" containerID="43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.232249 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6"} err="failed to get container status \"43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6\": rpc error: code = NotFound desc = could not find container \"43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6\": container with ID starting with 43823d28924f45bf6e2a4b10d6258372dd11d0e4268f451adda98516e1ee66d6 not found: ID does not exist" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.232513 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533530-xqkpb"] Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.244073 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr"] Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.269514 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjdwr\" (UniqueName: \"kubernetes.io/projected/bbd51857-44ff-4aa8-b388-5848273ba5eb-kube-api-access-qjdwr\") pod \"collect-profiles-29533530-nqgpr\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.269585 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbd51857-44ff-4aa8-b388-5848273ba5eb-secret-volume\") pod \"collect-profiles-29533530-nqgpr\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.269637 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57qdq\" (UniqueName: \"kubernetes.io/projected/8218e0f8-4605-401c-aeb1-a62888c01b58-kube-api-access-57qdq\") pod \"auto-csr-approver-29533530-xqkpb\" (UID: \"8218e0f8-4605-401c-aeb1-a62888c01b58\") " pod="openshift-infra/auto-csr-approver-29533530-xqkpb" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.269725 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbd51857-44ff-4aa8-b388-5848273ba5eb-config-volume\") pod \"collect-profiles-29533530-nqgpr\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.376728 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjdwr\" (UniqueName: \"kubernetes.io/projected/bbd51857-44ff-4aa8-b388-5848273ba5eb-kube-api-access-qjdwr\") pod \"collect-profiles-29533530-nqgpr\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.376814 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbd51857-44ff-4aa8-b388-5848273ba5eb-secret-volume\") pod \"collect-profiles-29533530-nqgpr\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.376867 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57qdq\" (UniqueName: \"kubernetes.io/projected/8218e0f8-4605-401c-aeb1-a62888c01b58-kube-api-access-57qdq\") pod \"auto-csr-approver-29533530-xqkpb\" (UID: \"8218e0f8-4605-401c-aeb1-a62888c01b58\") " pod="openshift-infra/auto-csr-approver-29533530-xqkpb" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.376921 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbd51857-44ff-4aa8-b388-5848273ba5eb-config-volume\") pod \"collect-profiles-29533530-nqgpr\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.378129 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbd51857-44ff-4aa8-b388-5848273ba5eb-config-volume\") pod \"collect-profiles-29533530-nqgpr\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.383103 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbd51857-44ff-4aa8-b388-5848273ba5eb-secret-volume\") pod \"collect-profiles-29533530-nqgpr\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.398527 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjdwr\" (UniqueName: \"kubernetes.io/projected/bbd51857-44ff-4aa8-b388-5848273ba5eb-kube-api-access-qjdwr\") pod \"collect-profiles-29533530-nqgpr\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.398604 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57qdq\" (UniqueName: \"kubernetes.io/projected/8218e0f8-4605-401c-aeb1-a62888c01b58-kube-api-access-57qdq\") pod \"auto-csr-approver-29533530-xqkpb\" (UID: \"8218e0f8-4605-401c-aeb1-a62888c01b58\") " pod="openshift-infra/auto-csr-approver-29533530-xqkpb" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.517022 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533530-xqkpb" Feb 25 09:30:00 crc kubenswrapper[4978]: I0225 09:30:00.521058 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:01 crc kubenswrapper[4978]: I0225 09:30:01.053398 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533530-xqkpb"] Feb 25 09:30:01 crc kubenswrapper[4978]: W0225 09:30:01.054003 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8218e0f8_4605_401c_aeb1_a62888c01b58.slice/crio-183c7e0f18d0972d6289b2315df2d90b7edde840fd657a6e49525a6dd86018cc WatchSource:0}: Error finding container 183c7e0f18d0972d6289b2315df2d90b7edde840fd657a6e49525a6dd86018cc: Status 404 returned error can't find the container with id 183c7e0f18d0972d6289b2315df2d90b7edde840fd657a6e49525a6dd86018cc Feb 25 09:30:01 crc kubenswrapper[4978]: I0225 09:30:01.075061 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr"] Feb 25 09:30:01 crc kubenswrapper[4978]: I0225 09:30:01.112199 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533530-xqkpb" event={"ID":"8218e0f8-4605-401c-aeb1-a62888c01b58","Type":"ContainerStarted","Data":"183c7e0f18d0972d6289b2315df2d90b7edde840fd657a6e49525a6dd86018cc"} Feb 25 09:30:01 crc kubenswrapper[4978]: I0225 09:30:01.115933 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" event={"ID":"bbd51857-44ff-4aa8-b388-5848273ba5eb","Type":"ContainerStarted","Data":"863071203b28c7fc761574bea9c83e1a5f98ba5417c9bdfd276c0c5fcb0b35a9"} Feb 25 09:30:01 crc kubenswrapper[4978]: I0225 09:30:01.339681 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7f9cc97-c95a-4de5-a563-b2a73d5db1c9" path="/var/lib/kubelet/pods/a7f9cc97-c95a-4de5-a563-b2a73d5db1c9/volumes" Feb 25 09:30:02 crc kubenswrapper[4978]: I0225 09:30:02.128793 4978 generic.go:334] "Generic (PLEG): container finished" podID="bbd51857-44ff-4aa8-b388-5848273ba5eb" containerID="4496ac9d54ffb77e8d4eeea4f3beb20acbc36ea662c186f151130fdefae29eb8" exitCode=0 Feb 25 09:30:02 crc kubenswrapper[4978]: I0225 09:30:02.128856 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" event={"ID":"bbd51857-44ff-4aa8-b388-5848273ba5eb","Type":"ContainerDied","Data":"4496ac9d54ffb77e8d4eeea4f3beb20acbc36ea662c186f151130fdefae29eb8"} Feb 25 09:30:03 crc kubenswrapper[4978]: I0225 09:30:03.139818 4978 generic.go:334] "Generic (PLEG): container finished" podID="8218e0f8-4605-401c-aeb1-a62888c01b58" containerID="a2d30affde97a5f34e759bfe290c97b1625ea333a18a0c2ee4e523a8478455bc" exitCode=0 Feb 25 09:30:03 crc kubenswrapper[4978]: I0225 09:30:03.139957 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533530-xqkpb" event={"ID":"8218e0f8-4605-401c-aeb1-a62888c01b58","Type":"ContainerDied","Data":"a2d30affde97a5f34e759bfe290c97b1625ea333a18a0c2ee4e523a8478455bc"} Feb 25 09:30:03 crc kubenswrapper[4978]: I0225 09:30:03.495050 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:03 crc kubenswrapper[4978]: I0225 09:30:03.544937 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjdwr\" (UniqueName: \"kubernetes.io/projected/bbd51857-44ff-4aa8-b388-5848273ba5eb-kube-api-access-qjdwr\") pod \"bbd51857-44ff-4aa8-b388-5848273ba5eb\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " Feb 25 09:30:03 crc kubenswrapper[4978]: I0225 09:30:03.545054 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbd51857-44ff-4aa8-b388-5848273ba5eb-secret-volume\") pod \"bbd51857-44ff-4aa8-b388-5848273ba5eb\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " Feb 25 09:30:03 crc kubenswrapper[4978]: I0225 09:30:03.545261 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbd51857-44ff-4aa8-b388-5848273ba5eb-config-volume\") pod \"bbd51857-44ff-4aa8-b388-5848273ba5eb\" (UID: \"bbd51857-44ff-4aa8-b388-5848273ba5eb\") " Feb 25 09:30:03 crc kubenswrapper[4978]: I0225 09:30:03.545954 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbd51857-44ff-4aa8-b388-5848273ba5eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "bbd51857-44ff-4aa8-b388-5848273ba5eb" (UID: "bbd51857-44ff-4aa8-b388-5848273ba5eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:30:03 crc kubenswrapper[4978]: I0225 09:30:03.647301 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bbd51857-44ff-4aa8-b388-5848273ba5eb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 09:30:03 crc kubenswrapper[4978]: I0225 09:30:03.986301 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbd51857-44ff-4aa8-b388-5848273ba5eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bbd51857-44ff-4aa8-b388-5848273ba5eb" (UID: "bbd51857-44ff-4aa8-b388-5848273ba5eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:30:03 crc kubenswrapper[4978]: I0225 09:30:03.989483 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbd51857-44ff-4aa8-b388-5848273ba5eb-kube-api-access-qjdwr" (OuterVolumeSpecName: "kube-api-access-qjdwr") pod "bbd51857-44ff-4aa8-b388-5848273ba5eb" (UID: "bbd51857-44ff-4aa8-b388-5848273ba5eb"). InnerVolumeSpecName "kube-api-access-qjdwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.057081 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjdwr\" (UniqueName: \"kubernetes.io/projected/bbd51857-44ff-4aa8-b388-5848273ba5eb-kube-api-access-qjdwr\") on node \"crc\" DevicePath \"\"" Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.057781 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bbd51857-44ff-4aa8-b388-5848273ba5eb-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.149850 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.149847 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr" event={"ID":"bbd51857-44ff-4aa8-b388-5848273ba5eb","Type":"ContainerDied","Data":"863071203b28c7fc761574bea9c83e1a5f98ba5417c9bdfd276c0c5fcb0b35a9"} Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.149984 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="863071203b28c7fc761574bea9c83e1a5f98ba5417c9bdfd276c0c5fcb0b35a9" Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.503060 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533530-xqkpb" Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.568039 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57qdq\" (UniqueName: \"kubernetes.io/projected/8218e0f8-4605-401c-aeb1-a62888c01b58-kube-api-access-57qdq\") pod \"8218e0f8-4605-401c-aeb1-a62888c01b58\" (UID: \"8218e0f8-4605-401c-aeb1-a62888c01b58\") " Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.581426 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc"] Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.585539 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8218e0f8-4605-401c-aeb1-a62888c01b58-kube-api-access-57qdq" (OuterVolumeSpecName: "kube-api-access-57qdq") pod "8218e0f8-4605-401c-aeb1-a62888c01b58" (UID: "8218e0f8-4605-401c-aeb1-a62888c01b58"). InnerVolumeSpecName "kube-api-access-57qdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.592208 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533485-s78gc"] Feb 25 09:30:04 crc kubenswrapper[4978]: I0225 09:30:04.670012 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57qdq\" (UniqueName: \"kubernetes.io/projected/8218e0f8-4605-401c-aeb1-a62888c01b58-kube-api-access-57qdq\") on node \"crc\" DevicePath \"\"" Feb 25 09:30:05 crc kubenswrapper[4978]: I0225 09:30:05.171252 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533530-xqkpb" event={"ID":"8218e0f8-4605-401c-aeb1-a62888c01b58","Type":"ContainerDied","Data":"183c7e0f18d0972d6289b2315df2d90b7edde840fd657a6e49525a6dd86018cc"} Feb 25 09:30:05 crc kubenswrapper[4978]: I0225 09:30:05.171287 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="183c7e0f18d0972d6289b2315df2d90b7edde840fd657a6e49525a6dd86018cc" Feb 25 09:30:05 crc kubenswrapper[4978]: I0225 09:30:05.171324 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533530-xqkpb" Feb 25 09:30:05 crc kubenswrapper[4978]: I0225 09:30:05.341890 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a15832c-a661-450a-911f-8d6294511806" path="/var/lib/kubelet/pods/7a15832c-a661-450a-911f-8d6294511806/volumes" Feb 25 09:30:05 crc kubenswrapper[4978]: I0225 09:30:05.607677 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533524-qlzgg"] Feb 25 09:30:05 crc kubenswrapper[4978]: I0225 09:30:05.625831 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533524-qlzgg"] Feb 25 09:30:07 crc kubenswrapper[4978]: I0225 09:30:07.363258 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ff69947-557c-4d72-9179-1c9ea3a306cf" path="/var/lib/kubelet/pods/7ff69947-557c-4d72-9179-1c9ea3a306cf/volumes" Feb 25 09:30:09 crc kubenswrapper[4978]: I0225 09:30:09.328223 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:30:09 crc kubenswrapper[4978]: E0225 09:30:09.329221 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:30:21 crc kubenswrapper[4978]: I0225 09:30:21.327882 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:30:21 crc kubenswrapper[4978]: E0225 09:30:21.328590 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:30:33 crc kubenswrapper[4978]: I0225 09:30:33.329157 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:30:33 crc kubenswrapper[4978]: E0225 09:30:33.330436 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:30:37 crc kubenswrapper[4978]: I0225 09:30:37.555627 4978 scope.go:117] "RemoveContainer" containerID="246de34d3d0c2174bd838d7222114a0ea1d253bd9c2e26f8f73b99f1000ad6af" Feb 25 09:30:37 crc kubenswrapper[4978]: I0225 09:30:37.611029 4978 scope.go:117] "RemoveContainer" containerID="46453f0473c10c6c1b411d9e56d5ed7fb001be232100a085ae82d8f46ea4a024" Feb 25 09:30:48 crc kubenswrapper[4978]: I0225 09:30:48.328752 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:30:48 crc kubenswrapper[4978]: E0225 09:30:48.329902 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:31:03 crc kubenswrapper[4978]: I0225 09:31:03.327699 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:31:03 crc kubenswrapper[4978]: E0225 09:31:03.328437 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:31:14 crc kubenswrapper[4978]: I0225 09:31:14.328034 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:31:14 crc kubenswrapper[4978]: E0225 09:31:14.328887 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:31:26 crc kubenswrapper[4978]: I0225 09:31:26.328877 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:31:26 crc kubenswrapper[4978]: E0225 09:31:26.331846 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:31:38 crc kubenswrapper[4978]: I0225 09:31:38.329047 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:31:38 crc kubenswrapper[4978]: E0225 09:31:38.330132 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:31:51 crc kubenswrapper[4978]: I0225 09:31:51.327429 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:31:51 crc kubenswrapper[4978]: E0225 09:31:51.329196 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:31:58 crc kubenswrapper[4978]: I0225 09:31:58.583715 4978 generic.go:334] "Generic (PLEG): container finished" podID="bc76d2fb-7bba-48f7-9658-e55f0c225807" containerID="8eef045326bcc5a995987c6d5d5aae11907578f026d7e687384e3fe7d3369184" exitCode=0 Feb 25 09:31:58 crc kubenswrapper[4978]: I0225 09:31:58.583852 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" event={"ID":"bc76d2fb-7bba-48f7-9658-e55f0c225807","Type":"ContainerDied","Data":"8eef045326bcc5a995987c6d5d5aae11907578f026d7e687384e3fe7d3369184"} Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.143482 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533532-vs9kk"] Feb 25 09:32:00 crc kubenswrapper[4978]: E0225 09:32:00.145543 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8218e0f8-4605-401c-aeb1-a62888c01b58" containerName="oc" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.145560 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="8218e0f8-4605-401c-aeb1-a62888c01b58" containerName="oc" Feb 25 09:32:00 crc kubenswrapper[4978]: E0225 09:32:00.145592 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd51857-44ff-4aa8-b388-5848273ba5eb" containerName="collect-profiles" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.145598 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd51857-44ff-4aa8-b388-5848273ba5eb" containerName="collect-profiles" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.145810 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbd51857-44ff-4aa8-b388-5848273ba5eb" containerName="collect-profiles" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.145835 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="8218e0f8-4605-401c-aeb1-a62888c01b58" containerName="oc" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.146547 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533532-vs9kk" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.153339 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.153637 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.153768 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.157860 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533532-vs9kk"] Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.192004 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.271354 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swrjr\" (UniqueName: \"kubernetes.io/projected/acaac0e4-7ca1-4d72-b1a2-1282a35cb609-kube-api-access-swrjr\") pod \"auto-csr-approver-29533532-vs9kk\" (UID: \"acaac0e4-7ca1-4d72-b1a2-1282a35cb609\") " pod="openshift-infra/auto-csr-approver-29533532-vs9kk" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.373501 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-2\") pod \"bc76d2fb-7bba-48f7-9658-e55f0c225807\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.373654 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-inventory\") pod \"bc76d2fb-7bba-48f7-9658-e55f0c225807\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.373700 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-1\") pod \"bc76d2fb-7bba-48f7-9658-e55f0c225807\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.373856 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-0\") pod \"bc76d2fb-7bba-48f7-9658-e55f0c225807\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.373910 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs6nk\" (UniqueName: \"kubernetes.io/projected/bc76d2fb-7bba-48f7-9658-e55f0c225807-kube-api-access-gs6nk\") pod \"bc76d2fb-7bba-48f7-9658-e55f0c225807\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.374049 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-telemetry-combined-ca-bundle\") pod \"bc76d2fb-7bba-48f7-9658-e55f0c225807\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.374176 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ssh-key-openstack-cell1\") pod \"bc76d2fb-7bba-48f7-9658-e55f0c225807\" (UID: \"bc76d2fb-7bba-48f7-9658-e55f0c225807\") " Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.374701 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swrjr\" (UniqueName: \"kubernetes.io/projected/acaac0e4-7ca1-4d72-b1a2-1282a35cb609-kube-api-access-swrjr\") pod \"auto-csr-approver-29533532-vs9kk\" (UID: \"acaac0e4-7ca1-4d72-b1a2-1282a35cb609\") " pod="openshift-infra/auto-csr-approver-29533532-vs9kk" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.382752 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "bc76d2fb-7bba-48f7-9658-e55f0c225807" (UID: "bc76d2fb-7bba-48f7-9658-e55f0c225807"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.383244 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc76d2fb-7bba-48f7-9658-e55f0c225807-kube-api-access-gs6nk" (OuterVolumeSpecName: "kube-api-access-gs6nk") pod "bc76d2fb-7bba-48f7-9658-e55f0c225807" (UID: "bc76d2fb-7bba-48f7-9658-e55f0c225807"). InnerVolumeSpecName "kube-api-access-gs6nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.397602 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swrjr\" (UniqueName: \"kubernetes.io/projected/acaac0e4-7ca1-4d72-b1a2-1282a35cb609-kube-api-access-swrjr\") pod \"auto-csr-approver-29533532-vs9kk\" (UID: \"acaac0e4-7ca1-4d72-b1a2-1282a35cb609\") " pod="openshift-infra/auto-csr-approver-29533532-vs9kk" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.411170 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "bc76d2fb-7bba-48f7-9658-e55f0c225807" (UID: "bc76d2fb-7bba-48f7-9658-e55f0c225807"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.436664 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "bc76d2fb-7bba-48f7-9658-e55f0c225807" (UID: "bc76d2fb-7bba-48f7-9658-e55f0c225807"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.438572 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-inventory" (OuterVolumeSpecName: "inventory") pod "bc76d2fb-7bba-48f7-9658-e55f0c225807" (UID: "bc76d2fb-7bba-48f7-9658-e55f0c225807"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.444042 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "bc76d2fb-7bba-48f7-9658-e55f0c225807" (UID: "bc76d2fb-7bba-48f7-9658-e55f0c225807"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.453646 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "bc76d2fb-7bba-48f7-9658-e55f0c225807" (UID: "bc76d2fb-7bba-48f7-9658-e55f0c225807"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.477235 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.477289 4978 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.477311 4978 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.477331 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs6nk\" (UniqueName: \"kubernetes.io/projected/bc76d2fb-7bba-48f7-9658-e55f0c225807-kube-api-access-gs6nk\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.477352 4978 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.477454 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.477474 4978 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/bc76d2fb-7bba-48f7-9658-e55f0c225807-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.504834 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533532-vs9kk" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.615061 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" event={"ID":"bc76d2fb-7bba-48f7-9658-e55f0c225807","Type":"ContainerDied","Data":"c5c8cd10ae13ddd500d43bc27801070ad6f0f0b2e617bc6ed9bae59229149a14"} Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.615309 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5c8cd10ae13ddd500d43bc27801070ad6f0f0b2e617bc6ed9bae59229149a14" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.615161 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-z7lrc" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.750534 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-rd5lr"] Feb 25 09:32:00 crc kubenswrapper[4978]: E0225 09:32:00.751275 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc76d2fb-7bba-48f7-9658-e55f0c225807" containerName="telemetry-openstack-openstack-cell1" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.751296 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc76d2fb-7bba-48f7-9658-e55f0c225807" containerName="telemetry-openstack-openstack-cell1" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.751509 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc76d2fb-7bba-48f7-9658-e55f0c225807" containerName="telemetry-openstack-openstack-cell1" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.752205 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.754909 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.755005 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.755155 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.755312 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.760018 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Feb 25 09:32:00 crc kubenswrapper[4978]: I0225 09:32:00.772943 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-rd5lr"] Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.887524 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.887924 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrpkt\" (UniqueName: \"kubernetes.io/projected/01398882-0eec-4662-b7fd-5d7c97308a4d-kube-api-access-qrpkt\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.888292 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.888490 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.888644 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.990356 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrpkt\" (UniqueName: \"kubernetes.io/projected/01398882-0eec-4662-b7fd-5d7c97308a4d-kube-api-access-qrpkt\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.990488 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.990553 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.990603 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.990644 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.996866 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.996982 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:00.997970 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:01.015353 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrpkt\" (UniqueName: \"kubernetes.io/projected/01398882-0eec-4662-b7fd-5d7c97308a4d-kube-api-access-qrpkt\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:01.035044 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-rd5lr\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:01.077142 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:01.088699 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533532-vs9kk"] Feb 25 09:32:01 crc kubenswrapper[4978]: I0225 09:32:01.632100 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533532-vs9kk" event={"ID":"acaac0e4-7ca1-4d72-b1a2-1282a35cb609","Type":"ContainerStarted","Data":"e5a315d5623bc4f641dba65b3769cbe776c08a32558c021e80023b5d35ec3d9c"} Feb 25 09:32:02 crc kubenswrapper[4978]: I0225 09:32:02.094064 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-rd5lr"] Feb 25 09:32:02 crc kubenswrapper[4978]: I0225 09:32:02.646205 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" event={"ID":"01398882-0eec-4662-b7fd-5d7c97308a4d","Type":"ContainerStarted","Data":"f53c86c731182f7a13414c20f6d5a079d5adf2db294fc12e3072b4be75bf8f1a"} Feb 25 09:32:02 crc kubenswrapper[4978]: I0225 09:32:02.649029 4978 generic.go:334] "Generic (PLEG): container finished" podID="acaac0e4-7ca1-4d72-b1a2-1282a35cb609" containerID="16c71a6225eb0b9f063723a31a848fab68e785aaa67a6585b86725c3957697ad" exitCode=0 Feb 25 09:32:02 crc kubenswrapper[4978]: I0225 09:32:02.649075 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533532-vs9kk" event={"ID":"acaac0e4-7ca1-4d72-b1a2-1282a35cb609","Type":"ContainerDied","Data":"16c71a6225eb0b9f063723a31a848fab68e785aaa67a6585b86725c3957697ad"} Feb 25 09:32:04 crc kubenswrapper[4978]: I0225 09:32:04.080423 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533532-vs9kk" Feb 25 09:32:04 crc kubenswrapper[4978]: I0225 09:32:04.184267 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swrjr\" (UniqueName: \"kubernetes.io/projected/acaac0e4-7ca1-4d72-b1a2-1282a35cb609-kube-api-access-swrjr\") pod \"acaac0e4-7ca1-4d72-b1a2-1282a35cb609\" (UID: \"acaac0e4-7ca1-4d72-b1a2-1282a35cb609\") " Feb 25 09:32:04 crc kubenswrapper[4978]: I0225 09:32:04.190867 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acaac0e4-7ca1-4d72-b1a2-1282a35cb609-kube-api-access-swrjr" (OuterVolumeSpecName: "kube-api-access-swrjr") pod "acaac0e4-7ca1-4d72-b1a2-1282a35cb609" (UID: "acaac0e4-7ca1-4d72-b1a2-1282a35cb609"). InnerVolumeSpecName "kube-api-access-swrjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:32:04 crc kubenswrapper[4978]: I0225 09:32:04.286628 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swrjr\" (UniqueName: \"kubernetes.io/projected/acaac0e4-7ca1-4d72-b1a2-1282a35cb609-kube-api-access-swrjr\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:04 crc kubenswrapper[4978]: I0225 09:32:04.329180 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:32:04 crc kubenswrapper[4978]: E0225 09:32:04.329923 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:32:04 crc kubenswrapper[4978]: I0225 09:32:04.677011 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" event={"ID":"01398882-0eec-4662-b7fd-5d7c97308a4d","Type":"ContainerStarted","Data":"481a35c408da3789302b80aa2a916a996f29ff005191911cdcf223d3690e495c"} Feb 25 09:32:04 crc kubenswrapper[4978]: I0225 09:32:04.680782 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533532-vs9kk" event={"ID":"acaac0e4-7ca1-4d72-b1a2-1282a35cb609","Type":"ContainerDied","Data":"e5a315d5623bc4f641dba65b3769cbe776c08a32558c021e80023b5d35ec3d9c"} Feb 25 09:32:04 crc kubenswrapper[4978]: I0225 09:32:04.680840 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5a315d5623bc4f641dba65b3769cbe776c08a32558c021e80023b5d35ec3d9c" Feb 25 09:32:04 crc kubenswrapper[4978]: I0225 09:32:04.680906 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533532-vs9kk" Feb 25 09:32:04 crc kubenswrapper[4978]: I0225 09:32:04.737611 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" podStartSLOduration=4.314653374 podStartE2EDuration="4.737579654s" podCreationTimestamp="2026-02-25 09:32:00 +0000 UTC" firstStartedPulling="2026-02-25 09:32:02.100703446 +0000 UTC m=+10015.539959915" lastFinishedPulling="2026-02-25 09:32:02.523629726 +0000 UTC m=+10015.962886195" observedRunningTime="2026-02-25 09:32:04.718263795 +0000 UTC m=+10018.157520334" watchObservedRunningTime="2026-02-25 09:32:04.737579654 +0000 UTC m=+10018.176836163" Feb 25 09:32:05 crc kubenswrapper[4978]: I0225 09:32:05.192550 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533526-6k9dz"] Feb 25 09:32:05 crc kubenswrapper[4978]: I0225 09:32:05.204943 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533526-6k9dz"] Feb 25 09:32:05 crc kubenswrapper[4978]: I0225 09:32:05.343300 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9d41622-d2f7-483a-90fa-00c2222fab68" path="/var/lib/kubelet/pods/c9d41622-d2f7-483a-90fa-00c2222fab68/volumes" Feb 25 09:32:19 crc kubenswrapper[4978]: I0225 09:32:19.328310 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:32:19 crc kubenswrapper[4978]: E0225 09:32:19.329411 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:32:31 crc kubenswrapper[4978]: I0225 09:32:31.327309 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:32:31 crc kubenswrapper[4978]: E0225 09:32:31.328081 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:32:37 crc kubenswrapper[4978]: I0225 09:32:37.897783 4978 scope.go:117] "RemoveContainer" containerID="ba392d1f7a7629d295f0d4a76332a2eea13bfcab82c8b9c704d81c5ed84041c0" Feb 25 09:32:43 crc kubenswrapper[4978]: I0225 09:32:43.328552 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:32:43 crc kubenswrapper[4978]: E0225 09:32:43.329504 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:32:48 crc kubenswrapper[4978]: I0225 09:32:48.207868 4978 generic.go:334] "Generic (PLEG): container finished" podID="01398882-0eec-4662-b7fd-5d7c97308a4d" containerID="481a35c408da3789302b80aa2a916a996f29ff005191911cdcf223d3690e495c" exitCode=0 Feb 25 09:32:48 crc kubenswrapper[4978]: I0225 09:32:48.207941 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" event={"ID":"01398882-0eec-4662-b7fd-5d7c97308a4d","Type":"ContainerDied","Data":"481a35c408da3789302b80aa2a916a996f29ff005191911cdcf223d3690e495c"} Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.786477 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.869060 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-combined-ca-bundle\") pod \"01398882-0eec-4662-b7fd-5d7c97308a4d\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.869138 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-inventory\") pod \"01398882-0eec-4662-b7fd-5d7c97308a4d\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.869326 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-agent-neutron-config-0\") pod \"01398882-0eec-4662-b7fd-5d7c97308a4d\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.869385 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrpkt\" (UniqueName: \"kubernetes.io/projected/01398882-0eec-4662-b7fd-5d7c97308a4d-kube-api-access-qrpkt\") pod \"01398882-0eec-4662-b7fd-5d7c97308a4d\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.869471 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-ssh-key-openstack-cell1\") pod \"01398882-0eec-4662-b7fd-5d7c97308a4d\" (UID: \"01398882-0eec-4662-b7fd-5d7c97308a4d\") " Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.874553 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "01398882-0eec-4662-b7fd-5d7c97308a4d" (UID: "01398882-0eec-4662-b7fd-5d7c97308a4d"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.874633 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01398882-0eec-4662-b7fd-5d7c97308a4d-kube-api-access-qrpkt" (OuterVolumeSpecName: "kube-api-access-qrpkt") pod "01398882-0eec-4662-b7fd-5d7c97308a4d" (UID: "01398882-0eec-4662-b7fd-5d7c97308a4d"). InnerVolumeSpecName "kube-api-access-qrpkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.895387 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "01398882-0eec-4662-b7fd-5d7c97308a4d" (UID: "01398882-0eec-4662-b7fd-5d7c97308a4d"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.895628 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "01398882-0eec-4662-b7fd-5d7c97308a4d" (UID: "01398882-0eec-4662-b7fd-5d7c97308a4d"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.895761 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-inventory" (OuterVolumeSpecName: "inventory") pod "01398882-0eec-4662-b7fd-5d7c97308a4d" (UID: "01398882-0eec-4662-b7fd-5d7c97308a4d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.971887 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.971925 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.971938 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrpkt\" (UniqueName: \"kubernetes.io/projected/01398882-0eec-4662-b7fd-5d7c97308a4d-kube-api-access-qrpkt\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.971950 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:49 crc kubenswrapper[4978]: I0225 09:32:49.971959 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01398882-0eec-4662-b7fd-5d7c97308a4d-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.235898 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" event={"ID":"01398882-0eec-4662-b7fd-5d7c97308a4d","Type":"ContainerDied","Data":"f53c86c731182f7a13414c20f6d5a079d5adf2db294fc12e3072b4be75bf8f1a"} Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.236187 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f53c86c731182f7a13414c20f6d5a079d5adf2db294fc12e3072b4be75bf8f1a" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.236244 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-rd5lr" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.362713 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm"] Feb 25 09:32:50 crc kubenswrapper[4978]: E0225 09:32:50.363250 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01398882-0eec-4662-b7fd-5d7c97308a4d" containerName="neutron-sriov-openstack-openstack-cell1" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.363270 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="01398882-0eec-4662-b7fd-5d7c97308a4d" containerName="neutron-sriov-openstack-openstack-cell1" Feb 25 09:32:50 crc kubenswrapper[4978]: E0225 09:32:50.363314 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acaac0e4-7ca1-4d72-b1a2-1282a35cb609" containerName="oc" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.363323 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="acaac0e4-7ca1-4d72-b1a2-1282a35cb609" containerName="oc" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.363589 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="acaac0e4-7ca1-4d72-b1a2-1282a35cb609" containerName="oc" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.363622 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="01398882-0eec-4662-b7fd-5d7c97308a4d" containerName="neutron-sriov-openstack-openstack-cell1" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.364567 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.366737 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.366807 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.366894 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.367377 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.369052 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.374555 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm"] Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.502766 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.502824 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.503240 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.503306 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.503442 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87rvv\" (UniqueName: \"kubernetes.io/projected/a5f51a06-f15a-4967-a8e6-25c422c09c32-kube-api-access-87rvv\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.605234 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.605313 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.605448 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87rvv\" (UniqueName: \"kubernetes.io/projected/a5f51a06-f15a-4967-a8e6-25c422c09c32-kube-api-access-87rvv\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.605547 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.605606 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.611541 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.611832 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.612370 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.616774 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.635099 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87rvv\" (UniqueName: \"kubernetes.io/projected/a5f51a06-f15a-4967-a8e6-25c422c09c32-kube-api-access-87rvv\") pod \"neutron-dhcp-openstack-openstack-cell1-qnnvm\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:50 crc kubenswrapper[4978]: I0225 09:32:50.686951 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:32:51 crc kubenswrapper[4978]: I0225 09:32:51.269995 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm"] Feb 25 09:32:52 crc kubenswrapper[4978]: I0225 09:32:52.261944 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" event={"ID":"a5f51a06-f15a-4967-a8e6-25c422c09c32","Type":"ContainerStarted","Data":"80148c017c143b50f90ea96d3f431a3070357562b2868d703dba7f8939781e70"} Feb 25 09:32:52 crc kubenswrapper[4978]: I0225 09:32:52.262457 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" event={"ID":"a5f51a06-f15a-4967-a8e6-25c422c09c32","Type":"ContainerStarted","Data":"c708f5c73a69b283a51f61d406e9a5e1e74f2654307367f6d070aa99a4fb1d8d"} Feb 25 09:32:58 crc kubenswrapper[4978]: I0225 09:32:58.328532 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:32:58 crc kubenswrapper[4978]: E0225 09:32:58.329996 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:33:12 crc kubenswrapper[4978]: I0225 09:33:12.328458 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:33:12 crc kubenswrapper[4978]: E0225 09:33:12.329805 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:33:24 crc kubenswrapper[4978]: I0225 09:33:24.328077 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:33:24 crc kubenswrapper[4978]: E0225 09:33:24.329276 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:33:38 crc kubenswrapper[4978]: I0225 09:33:38.329623 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:33:38 crc kubenswrapper[4978]: E0225 09:33:38.330917 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:33:50 crc kubenswrapper[4978]: I0225 09:33:50.327845 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:33:50 crc kubenswrapper[4978]: E0225 09:33:50.328700 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:33:56 crc kubenswrapper[4978]: I0225 09:33:56.062490 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f51a06-f15a-4967-a8e6-25c422c09c32" containerID="80148c017c143b50f90ea96d3f431a3070357562b2868d703dba7f8939781e70" exitCode=0 Feb 25 09:33:56 crc kubenswrapper[4978]: I0225 09:33:56.062615 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" event={"ID":"a5f51a06-f15a-4967-a8e6-25c422c09c32","Type":"ContainerDied","Data":"80148c017c143b50f90ea96d3f431a3070357562b2868d703dba7f8939781e70"} Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.579989 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.668301 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-inventory\") pod \"a5f51a06-f15a-4967-a8e6-25c422c09c32\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.668435 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-agent-neutron-config-0\") pod \"a5f51a06-f15a-4967-a8e6-25c422c09c32\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.668490 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-ssh-key-openstack-cell1\") pod \"a5f51a06-f15a-4967-a8e6-25c422c09c32\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.668556 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-combined-ca-bundle\") pod \"a5f51a06-f15a-4967-a8e6-25c422c09c32\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.668733 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87rvv\" (UniqueName: \"kubernetes.io/projected/a5f51a06-f15a-4967-a8e6-25c422c09c32-kube-api-access-87rvv\") pod \"a5f51a06-f15a-4967-a8e6-25c422c09c32\" (UID: \"a5f51a06-f15a-4967-a8e6-25c422c09c32\") " Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.674402 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "a5f51a06-f15a-4967-a8e6-25c422c09c32" (UID: "a5f51a06-f15a-4967-a8e6-25c422c09c32"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.674786 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f51a06-f15a-4967-a8e6-25c422c09c32-kube-api-access-87rvv" (OuterVolumeSpecName: "kube-api-access-87rvv") pod "a5f51a06-f15a-4967-a8e6-25c422c09c32" (UID: "a5f51a06-f15a-4967-a8e6-25c422c09c32"). InnerVolumeSpecName "kube-api-access-87rvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.706255 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "a5f51a06-f15a-4967-a8e6-25c422c09c32" (UID: "a5f51a06-f15a-4967-a8e6-25c422c09c32"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.718017 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "a5f51a06-f15a-4967-a8e6-25c422c09c32" (UID: "a5f51a06-f15a-4967-a8e6-25c422c09c32"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.731353 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-inventory" (OuterVolumeSpecName: "inventory") pod "a5f51a06-f15a-4967-a8e6-25c422c09c32" (UID: "a5f51a06-f15a-4967-a8e6-25c422c09c32"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.790133 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.795455 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.795684 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.795765 4978 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f51a06-f15a-4967-a8e6-25c422c09c32-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:33:57 crc kubenswrapper[4978]: I0225 09:33:57.795820 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87rvv\" (UniqueName: \"kubernetes.io/projected/a5f51a06-f15a-4967-a8e6-25c422c09c32-kube-api-access-87rvv\") on node \"crc\" DevicePath \"\"" Feb 25 09:33:58 crc kubenswrapper[4978]: I0225 09:33:58.092035 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" event={"ID":"a5f51a06-f15a-4967-a8e6-25c422c09c32","Type":"ContainerDied","Data":"c708f5c73a69b283a51f61d406e9a5e1e74f2654307367f6d070aa99a4fb1d8d"} Feb 25 09:33:58 crc kubenswrapper[4978]: I0225 09:33:58.092505 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c708f5c73a69b283a51f61d406e9a5e1e74f2654307367f6d070aa99a4fb1d8d" Feb 25 09:33:58 crc kubenswrapper[4978]: I0225 09:33:58.092170 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-qnnvm" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.175080 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533534-wpxld"] Feb 25 09:34:00 crc kubenswrapper[4978]: E0225 09:34:00.175985 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f51a06-f15a-4967-a8e6-25c422c09c32" containerName="neutron-dhcp-openstack-openstack-cell1" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.176010 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f51a06-f15a-4967-a8e6-25c422c09c32" containerName="neutron-dhcp-openstack-openstack-cell1" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.176425 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f51a06-f15a-4967-a8e6-25c422c09c32" containerName="neutron-dhcp-openstack-openstack-cell1" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.177580 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533534-wpxld" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.180545 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.181182 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.185563 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.188063 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533534-wpxld"] Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.255660 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d5d2\" (UniqueName: \"kubernetes.io/projected/fed10669-c439-41b5-bab7-081176ff23b5-kube-api-access-6d5d2\") pod \"auto-csr-approver-29533534-wpxld\" (UID: \"fed10669-c439-41b5-bab7-081176ff23b5\") " pod="openshift-infra/auto-csr-approver-29533534-wpxld" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.358961 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d5d2\" (UniqueName: \"kubernetes.io/projected/fed10669-c439-41b5-bab7-081176ff23b5-kube-api-access-6d5d2\") pod \"auto-csr-approver-29533534-wpxld\" (UID: \"fed10669-c439-41b5-bab7-081176ff23b5\") " pod="openshift-infra/auto-csr-approver-29533534-wpxld" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.393303 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d5d2\" (UniqueName: \"kubernetes.io/projected/fed10669-c439-41b5-bab7-081176ff23b5-kube-api-access-6d5d2\") pod \"auto-csr-approver-29533534-wpxld\" (UID: \"fed10669-c439-41b5-bab7-081176ff23b5\") " pod="openshift-infra/auto-csr-approver-29533534-wpxld" Feb 25 09:34:00 crc kubenswrapper[4978]: I0225 09:34:00.501557 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533534-wpxld" Feb 25 09:34:01 crc kubenswrapper[4978]: I0225 09:34:01.054830 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533534-wpxld"] Feb 25 09:34:01 crc kubenswrapper[4978]: W0225 09:34:01.059566 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfed10669_c439_41b5_bab7_081176ff23b5.slice/crio-c74ec32e743419c1b2d204f0f0b8aef5ac88340bb8424435cdcf6e9e579be433 WatchSource:0}: Error finding container c74ec32e743419c1b2d204f0f0b8aef5ac88340bb8424435cdcf6e9e579be433: Status 404 returned error can't find the container with id c74ec32e743419c1b2d204f0f0b8aef5ac88340bb8424435cdcf6e9e579be433 Feb 25 09:34:01 crc kubenswrapper[4978]: I0225 09:34:01.063291 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:34:01 crc kubenswrapper[4978]: I0225 09:34:01.131290 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533534-wpxld" event={"ID":"fed10669-c439-41b5-bab7-081176ff23b5","Type":"ContainerStarted","Data":"c74ec32e743419c1b2d204f0f0b8aef5ac88340bb8424435cdcf6e9e579be433"} Feb 25 09:34:02 crc kubenswrapper[4978]: I0225 09:34:02.328147 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:34:02 crc kubenswrapper[4978]: E0225 09:34:02.328675 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:34:02 crc kubenswrapper[4978]: I0225 09:34:02.521966 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 09:34:02 crc kubenswrapper[4978]: I0225 09:34:02.522541 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="1282752c-57e3-4e24-aa3f-cecdea868016" containerName="nova-cell0-conductor-conductor" containerID="cri-o://a96209f29de2fef3610427b4fa5713572b82f84289fa56e7de27488ba09e7301" gracePeriod=30 Feb 25 09:34:02 crc kubenswrapper[4978]: I0225 09:34:02.560467 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 09:34:02 crc kubenswrapper[4978]: I0225 09:34:02.560770 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="57345093-7635-4852-bf51-5f0610196ab3" containerName="nova-cell1-conductor-conductor" containerID="cri-o://660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2" gracePeriod=30 Feb 25 09:34:03 crc kubenswrapper[4978]: I0225 09:34:03.151331 4978 generic.go:334] "Generic (PLEG): container finished" podID="fed10669-c439-41b5-bab7-081176ff23b5" containerID="4705e77197a4af0d64e72956803bbc8fd40c56a0dea45bc7278fa28a769d0b7e" exitCode=0 Feb 25 09:34:03 crc kubenswrapper[4978]: I0225 09:34:03.151428 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533534-wpxld" event={"ID":"fed10669-c439-41b5-bab7-081176ff23b5","Type":"ContainerDied","Data":"4705e77197a4af0d64e72956803bbc8fd40c56a0dea45bc7278fa28a769d0b7e"} Feb 25 09:34:03 crc kubenswrapper[4978]: I0225 09:34:03.561004 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 09:34:03 crc kubenswrapper[4978]: I0225 09:34:03.561298 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerName="nova-api-log" containerID="cri-o://6f3953c76379bf4d3f3b4386500608827daf936585aaadf1474621514f22f5f4" gracePeriod=30 Feb 25 09:34:03 crc kubenswrapper[4978]: I0225 09:34:03.561756 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerName="nova-api-api" containerID="cri-o://9a241ef1e2aa2ef3236b4dbcd65d86341b67ad51d21eae2baaada557ac2a0e09" gracePeriod=30 Feb 25 09:34:03 crc kubenswrapper[4978]: I0225 09:34:03.572024 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 09:34:03 crc kubenswrapper[4978]: I0225 09:34:03.572241 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c14894d9-d8c4-49df-a6e4-d9823efe8f50" containerName="nova-scheduler-scheduler" containerID="cri-o://5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9" gracePeriod=30 Feb 25 09:34:03 crc kubenswrapper[4978]: I0225 09:34:03.606510 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 09:34:03 crc kubenswrapper[4978]: I0225 09:34:03.606812 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerName="nova-metadata-log" containerID="cri-o://dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2" gracePeriod=30 Feb 25 09:34:03 crc kubenswrapper[4978]: I0225 09:34:03.606885 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerName="nova-metadata-metadata" containerID="cri-o://2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19" gracePeriod=30 Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.148276 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.159852 4978 generic.go:334] "Generic (PLEG): container finished" podID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerID="dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2" exitCode=143 Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.160015 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bb14a83b-160a-473c-bbe7-d4720ed8bcb4","Type":"ContainerDied","Data":"dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2"} Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.161793 4978 generic.go:334] "Generic (PLEG): container finished" podID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerID="6f3953c76379bf4d3f3b4386500608827daf936585aaadf1474621514f22f5f4" exitCode=143 Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.161878 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22a39e3e-3d98-4240-b845-75e2d5456ac6","Type":"ContainerDied","Data":"6f3953c76379bf4d3f3b4386500608827daf936585aaadf1474621514f22f5f4"} Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.165754 4978 generic.go:334] "Generic (PLEG): container finished" podID="57345093-7635-4852-bf51-5f0610196ab3" containerID="660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2" exitCode=0 Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.165815 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"57345093-7635-4852-bf51-5f0610196ab3","Type":"ContainerDied","Data":"660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2"} Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.165831 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.165854 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"57345093-7635-4852-bf51-5f0610196ab3","Type":"ContainerDied","Data":"4b0a930d88c1f1a5e447360c5190cb1c46abb0fc034bb84801375adf1bdd5d42"} Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.165873 4978 scope.go:117] "RemoveContainer" containerID="660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.202790 4978 scope.go:117] "RemoveContainer" containerID="660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2" Feb 25 09:34:04 crc kubenswrapper[4978]: E0225 09:34:04.204018 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2\": container with ID starting with 660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2 not found: ID does not exist" containerID="660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.204059 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2"} err="failed to get container status \"660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2\": rpc error: code = NotFound desc = could not find container \"660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2\": container with ID starting with 660d4a7e5be374017e70a685b8233b12a2bc8c3cb705fedaec1e4ef9e76ae9d2 not found: ID does not exist" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.239114 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-config-data\") pod \"57345093-7635-4852-bf51-5f0610196ab3\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.239394 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-combined-ca-bundle\") pod \"57345093-7635-4852-bf51-5f0610196ab3\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.239520 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh85w\" (UniqueName: \"kubernetes.io/projected/57345093-7635-4852-bf51-5f0610196ab3-kube-api-access-qh85w\") pod \"57345093-7635-4852-bf51-5f0610196ab3\" (UID: \"57345093-7635-4852-bf51-5f0610196ab3\") " Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.245809 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57345093-7635-4852-bf51-5f0610196ab3-kube-api-access-qh85w" (OuterVolumeSpecName: "kube-api-access-qh85w") pod "57345093-7635-4852-bf51-5f0610196ab3" (UID: "57345093-7635-4852-bf51-5f0610196ab3"). InnerVolumeSpecName "kube-api-access-qh85w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.283154 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-config-data" (OuterVolumeSpecName: "config-data") pod "57345093-7635-4852-bf51-5f0610196ab3" (UID: "57345093-7635-4852-bf51-5f0610196ab3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.303507 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57345093-7635-4852-bf51-5f0610196ab3" (UID: "57345093-7635-4852-bf51-5f0610196ab3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.341688 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.341728 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh85w\" (UniqueName: \"kubernetes.io/projected/57345093-7635-4852-bf51-5f0610196ab3-kube-api-access-qh85w\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.341739 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57345093-7635-4852-bf51-5f0610196ab3-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:04 crc kubenswrapper[4978]: E0225 09:34:04.423781 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 09:34:04 crc kubenswrapper[4978]: E0225 09:34:04.425413 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 09:34:04 crc kubenswrapper[4978]: E0225 09:34:04.428449 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 25 09:34:04 crc kubenswrapper[4978]: E0225 09:34:04.428481 4978 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c14894d9-d8c4-49df-a6e4-d9823efe8f50" containerName="nova-scheduler-scheduler" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.528991 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533534-wpxld" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.541524 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.552711 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.617741 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 09:34:04 crc kubenswrapper[4978]: E0225 09:34:04.618133 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57345093-7635-4852-bf51-5f0610196ab3" containerName="nova-cell1-conductor-conductor" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.618147 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="57345093-7635-4852-bf51-5f0610196ab3" containerName="nova-cell1-conductor-conductor" Feb 25 09:34:04 crc kubenswrapper[4978]: E0225 09:34:04.618178 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed10669-c439-41b5-bab7-081176ff23b5" containerName="oc" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.618186 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed10669-c439-41b5-bab7-081176ff23b5" containerName="oc" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.618586 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="57345093-7635-4852-bf51-5f0610196ab3" containerName="nova-cell1-conductor-conductor" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.618605 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="fed10669-c439-41b5-bab7-081176ff23b5" containerName="oc" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.619228 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.622641 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.637204 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.650842 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d5d2\" (UniqueName: \"kubernetes.io/projected/fed10669-c439-41b5-bab7-081176ff23b5-kube-api-access-6d5d2\") pod \"fed10669-c439-41b5-bab7-081176ff23b5\" (UID: \"fed10669-c439-41b5-bab7-081176ff23b5\") " Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.655790 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fed10669-c439-41b5-bab7-081176ff23b5-kube-api-access-6d5d2" (OuterVolumeSpecName: "kube-api-access-6d5d2") pod "fed10669-c439-41b5-bab7-081176ff23b5" (UID: "fed10669-c439-41b5-bab7-081176ff23b5"). InnerVolumeSpecName "kube-api-access-6d5d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.753157 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4\") " pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.753272 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4\") " pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.753313 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn75p\" (UniqueName: \"kubernetes.io/projected/f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4-kube-api-access-hn75p\") pod \"nova-cell1-conductor-0\" (UID: \"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4\") " pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.753408 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d5d2\" (UniqueName: \"kubernetes.io/projected/fed10669-c439-41b5-bab7-081176ff23b5-kube-api-access-6d5d2\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.855282 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn75p\" (UniqueName: \"kubernetes.io/projected/f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4-kube-api-access-hn75p\") pod \"nova-cell1-conductor-0\" (UID: \"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4\") " pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.855754 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4\") " pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.855850 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4\") " pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.861109 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4\") " pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.864115 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4\") " pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.873955 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn75p\" (UniqueName: \"kubernetes.io/projected/f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4-kube-api-access-hn75p\") pod \"nova-cell1-conductor-0\" (UID: \"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4\") " pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:04 crc kubenswrapper[4978]: I0225 09:34:04.934491 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.181678 4978 generic.go:334] "Generic (PLEG): container finished" podID="1282752c-57e3-4e24-aa3f-cecdea868016" containerID="a96209f29de2fef3610427b4fa5713572b82f84289fa56e7de27488ba09e7301" exitCode=0 Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.181768 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1282752c-57e3-4e24-aa3f-cecdea868016","Type":"ContainerDied","Data":"a96209f29de2fef3610427b4fa5713572b82f84289fa56e7de27488ba09e7301"} Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.185633 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533534-wpxld" event={"ID":"fed10669-c439-41b5-bab7-081176ff23b5","Type":"ContainerDied","Data":"c74ec32e743419c1b2d204f0f0b8aef5ac88340bb8424435cdcf6e9e579be433"} Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.185664 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c74ec32e743419c1b2d204f0f0b8aef5ac88340bb8424435cdcf6e9e579be433" Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.185719 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533534-wpxld" Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.348395 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57345093-7635-4852-bf51-5f0610196ab3" path="/var/lib/kubelet/pods/57345093-7635-4852-bf51-5f0610196ab3/volumes" Feb 25 09:34:05 crc kubenswrapper[4978]: W0225 09:34:05.447628 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2fb3fa7_1a7e_41e4_a11f_9ae6bd9380c4.slice/crio-98d3ab737d0dac305c17ec929cb388f0eeebe5b97b43d0ff040c4bd620f22149 WatchSource:0}: Error finding container 98d3ab737d0dac305c17ec929cb388f0eeebe5b97b43d0ff040c4bd620f22149: Status 404 returned error can't find the container with id 98d3ab737d0dac305c17ec929cb388f0eeebe5b97b43d0ff040c4bd620f22149 Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.447846 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.618572 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533528-jc4rq"] Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.627940 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533528-jc4rq"] Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.697441 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.782909 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-combined-ca-bundle\") pod \"1282752c-57e3-4e24-aa3f-cecdea868016\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.783195 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-config-data\") pod \"1282752c-57e3-4e24-aa3f-cecdea868016\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.783483 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn5zp\" (UniqueName: \"kubernetes.io/projected/1282752c-57e3-4e24-aa3f-cecdea868016-kube-api-access-vn5zp\") pod \"1282752c-57e3-4e24-aa3f-cecdea868016\" (UID: \"1282752c-57e3-4e24-aa3f-cecdea868016\") " Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.788687 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1282752c-57e3-4e24-aa3f-cecdea868016-kube-api-access-vn5zp" (OuterVolumeSpecName: "kube-api-access-vn5zp") pod "1282752c-57e3-4e24-aa3f-cecdea868016" (UID: "1282752c-57e3-4e24-aa3f-cecdea868016"). InnerVolumeSpecName "kube-api-access-vn5zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.823617 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-config-data" (OuterVolumeSpecName: "config-data") pod "1282752c-57e3-4e24-aa3f-cecdea868016" (UID: "1282752c-57e3-4e24-aa3f-cecdea868016"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.825147 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1282752c-57e3-4e24-aa3f-cecdea868016" (UID: "1282752c-57e3-4e24-aa3f-cecdea868016"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.885406 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn5zp\" (UniqueName: \"kubernetes.io/projected/1282752c-57e3-4e24-aa3f-cecdea868016-kube-api-access-vn5zp\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.885443 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:05 crc kubenswrapper[4978]: I0225 09:34:05.885456 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1282752c-57e3-4e24-aa3f-cecdea868016-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.198497 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1282752c-57e3-4e24-aa3f-cecdea868016","Type":"ContainerDied","Data":"178d8e1065554bc7c5ad34b946b36eba3b039a2726d6341870a42d43c3a5e5b8"} Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.198531 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.198543 4978 scope.go:117] "RemoveContainer" containerID="a96209f29de2fef3610427b4fa5713572b82f84289fa56e7de27488ba09e7301" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.200783 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4","Type":"ContainerStarted","Data":"0c2f43f7295020418adeba5a407b395c816ad3672c7f410b1402bce82e86d0b5"} Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.200822 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4","Type":"ContainerStarted","Data":"98d3ab737d0dac305c17ec929cb388f0eeebe5b97b43d0ff040c4bd620f22149"} Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.200927 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.229708 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.2296865439999998 podStartE2EDuration="2.229686544s" podCreationTimestamp="2026-02-25 09:34:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:34:06.215855835 +0000 UTC m=+10139.655112354" watchObservedRunningTime="2026-02-25 09:34:06.229686544 +0000 UTC m=+10139.668943013" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.259841 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.279585 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.294903 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 09:34:06 crc kubenswrapper[4978]: E0225 09:34:06.295282 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1282752c-57e3-4e24-aa3f-cecdea868016" containerName="nova-cell0-conductor-conductor" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.295298 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1282752c-57e3-4e24-aa3f-cecdea868016" containerName="nova-cell0-conductor-conductor" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.295530 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1282752c-57e3-4e24-aa3f-cecdea868016" containerName="nova-cell0-conductor-conductor" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.296243 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.298635 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.303330 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.395193 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/621cc593-53a8-4319-91c7-7ec49790758a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"621cc593-53a8-4319-91c7-7ec49790758a\") " pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.395297 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/621cc593-53a8-4319-91c7-7ec49790758a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"621cc593-53a8-4319-91c7-7ec49790758a\") " pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.395349 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r29v\" (UniqueName: \"kubernetes.io/projected/621cc593-53a8-4319-91c7-7ec49790758a-kube-api-access-2r29v\") pod \"nova-cell0-conductor-0\" (UID: \"621cc593-53a8-4319-91c7-7ec49790758a\") " pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.496885 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/621cc593-53a8-4319-91c7-7ec49790758a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"621cc593-53a8-4319-91c7-7ec49790758a\") " pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.497005 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/621cc593-53a8-4319-91c7-7ec49790758a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"621cc593-53a8-4319-91c7-7ec49790758a\") " pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.497044 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r29v\" (UniqueName: \"kubernetes.io/projected/621cc593-53a8-4319-91c7-7ec49790758a-kube-api-access-2r29v\") pod \"nova-cell0-conductor-0\" (UID: \"621cc593-53a8-4319-91c7-7ec49790758a\") " pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.504250 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/621cc593-53a8-4319-91c7-7ec49790758a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"621cc593-53a8-4319-91c7-7ec49790758a\") " pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.518425 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/621cc593-53a8-4319-91c7-7ec49790758a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"621cc593-53a8-4319-91c7-7ec49790758a\") " pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.521453 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r29v\" (UniqueName: \"kubernetes.io/projected/621cc593-53a8-4319-91c7-7ec49790758a-kube-api-access-2r29v\") pod \"nova-cell0-conductor-0\" (UID: \"621cc593-53a8-4319-91c7-7ec49790758a\") " pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:06 crc kubenswrapper[4978]: I0225 09:34:06.647772 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.121798 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.136835 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 25 09:34:07 crc kubenswrapper[4978]: W0225 09:34:07.147209 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod621cc593_53a8_4319_91c7_7ec49790758a.slice/crio-02dfa9b75777367dcd26d8f9ee60b1afedc11d23d4d3eeabc1de3c444b75850b WatchSource:0}: Error finding container 02dfa9b75777367dcd26d8f9ee60b1afedc11d23d4d3eeabc1de3c444b75850b: Status 404 returned error can't find the container with id 02dfa9b75777367dcd26d8f9ee60b1afedc11d23d4d3eeabc1de3c444b75850b Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.213421 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-config-data\") pod \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.213500 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-combined-ca-bundle\") pod \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.213522 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5bwt\" (UniqueName: \"kubernetes.io/projected/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-kube-api-access-f5bwt\") pod \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.213710 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-nova-metadata-tls-certs\") pod \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.213771 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-logs\") pod \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\" (UID: \"bb14a83b-160a-473c-bbe7-d4720ed8bcb4\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.214466 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-logs" (OuterVolumeSpecName: "logs") pod "bb14a83b-160a-473c-bbe7-d4720ed8bcb4" (UID: "bb14a83b-160a-473c-bbe7-d4720ed8bcb4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.229011 4978 generic.go:334] "Generic (PLEG): container finished" podID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerID="2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19" exitCode=0 Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.229112 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bb14a83b-160a-473c-bbe7-d4720ed8bcb4","Type":"ContainerDied","Data":"2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19"} Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.229141 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bb14a83b-160a-473c-bbe7-d4720ed8bcb4","Type":"ContainerDied","Data":"e312c4b0b988910ebedf21b46037fce4cf3817540f99e05ea2c2c9fac23a21a9"} Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.229159 4978 scope.go:117] "RemoveContainer" containerID="2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.229277 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.230657 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-kube-api-access-f5bwt" (OuterVolumeSpecName: "kube-api-access-f5bwt") pod "bb14a83b-160a-473c-bbe7-d4720ed8bcb4" (UID: "bb14a83b-160a-473c-bbe7-d4720ed8bcb4"). InnerVolumeSpecName "kube-api-access-f5bwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.238907 4978 generic.go:334] "Generic (PLEG): container finished" podID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerID="9a241ef1e2aa2ef3236b4dbcd65d86341b67ad51d21eae2baaada557ac2a0e09" exitCode=0 Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.238999 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22a39e3e-3d98-4240-b845-75e2d5456ac6","Type":"ContainerDied","Data":"9a241ef1e2aa2ef3236b4dbcd65d86341b67ad51d21eae2baaada557ac2a0e09"} Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.259517 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"621cc593-53a8-4319-91c7-7ec49790758a","Type":"ContainerStarted","Data":"02dfa9b75777367dcd26d8f9ee60b1afedc11d23d4d3eeabc1de3c444b75850b"} Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.267724 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb14a83b-160a-473c-bbe7-d4720ed8bcb4" (UID: "bb14a83b-160a-473c-bbe7-d4720ed8bcb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.274947 4978 scope.go:117] "RemoveContainer" containerID="dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.287274 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "bb14a83b-160a-473c-bbe7-d4720ed8bcb4" (UID: "bb14a83b-160a-473c-bbe7-d4720ed8bcb4"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.312464 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-config-data" (OuterVolumeSpecName: "config-data") pod "bb14a83b-160a-473c-bbe7-d4720ed8bcb4" (UID: "bb14a83b-160a-473c-bbe7-d4720ed8bcb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.315901 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-logs\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.315928 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.315937 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.315950 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5bwt\" (UniqueName: \"kubernetes.io/projected/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-kube-api-access-f5bwt\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.315958 4978 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb14a83b-160a-473c-bbe7-d4720ed8bcb4-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.352056 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1282752c-57e3-4e24-aa3f-cecdea868016" path="/var/lib/kubelet/pods/1282752c-57e3-4e24-aa3f-cecdea868016/volumes" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.366273 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc30ddfa-fe50-42b3-b97e-0967a0b69363" path="/var/lib/kubelet/pods/bc30ddfa-fe50-42b3-b97e-0967a0b69363/volumes" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.426170 4978 scope.go:117] "RemoveContainer" containerID="2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19" Feb 25 09:34:07 crc kubenswrapper[4978]: E0225 09:34:07.426669 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19\": container with ID starting with 2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19 not found: ID does not exist" containerID="2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.426711 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19"} err="failed to get container status \"2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19\": rpc error: code = NotFound desc = could not find container \"2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19\": container with ID starting with 2f5208212f75711cdbe57053d1abb471ff24cebef1d96259bebbc21d84d84d19 not found: ID does not exist" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.426776 4978 scope.go:117] "RemoveContainer" containerID="dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2" Feb 25 09:34:07 crc kubenswrapper[4978]: E0225 09:34:07.427049 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2\": container with ID starting with dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2 not found: ID does not exist" containerID="dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.427081 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2"} err="failed to get container status \"dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2\": rpc error: code = NotFound desc = could not find container \"dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2\": container with ID starting with dc9794f371f84510c134888cef592ad83990fee0a9f3d710bc1d3a6cf50a62f2 not found: ID does not exist" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.573561 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.590864 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.614514 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.626604 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 25 09:34:07 crc kubenswrapper[4978]: E0225 09:34:07.626962 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerName="nova-metadata-metadata" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.626975 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerName="nova-metadata-metadata" Feb 25 09:34:07 crc kubenswrapper[4978]: E0225 09:34:07.626990 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerName="nova-metadata-log" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.626997 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerName="nova-metadata-log" Feb 25 09:34:07 crc kubenswrapper[4978]: E0225 09:34:07.627032 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerName="nova-api-log" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.627038 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerName="nova-api-log" Feb 25 09:34:07 crc kubenswrapper[4978]: E0225 09:34:07.627050 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerName="nova-api-api" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.627055 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerName="nova-api-api" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.627229 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerName="nova-metadata-metadata" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.627245 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerName="nova-api-api" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.627338 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" containerName="nova-metadata-log" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.627353 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" containerName="nova-api-log" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.628278 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.634831 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.635094 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.654995 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.737464 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-public-tls-certs\") pod \"22a39e3e-3d98-4240-b845-75e2d5456ac6\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.737523 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57wmn\" (UniqueName: \"kubernetes.io/projected/22a39e3e-3d98-4240-b845-75e2d5456ac6-kube-api-access-57wmn\") pod \"22a39e3e-3d98-4240-b845-75e2d5456ac6\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.737594 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a39e3e-3d98-4240-b845-75e2d5456ac6-logs\") pod \"22a39e3e-3d98-4240-b845-75e2d5456ac6\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.737667 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-internal-tls-certs\") pod \"22a39e3e-3d98-4240-b845-75e2d5456ac6\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.737705 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-combined-ca-bundle\") pod \"22a39e3e-3d98-4240-b845-75e2d5456ac6\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.737820 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-config-data\") pod \"22a39e3e-3d98-4240-b845-75e2d5456ac6\" (UID: \"22a39e3e-3d98-4240-b845-75e2d5456ac6\") " Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.738056 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtt9t\" (UniqueName: \"kubernetes.io/projected/d842555c-bc9c-492f-8a1a-db131c37f5ae-kube-api-access-gtt9t\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.738113 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d842555c-bc9c-492f-8a1a-db131c37f5ae-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.738173 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d842555c-bc9c-492f-8a1a-db131c37f5ae-config-data\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.738243 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d842555c-bc9c-492f-8a1a-db131c37f5ae-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.738272 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d842555c-bc9c-492f-8a1a-db131c37f5ae-logs\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.738582 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22a39e3e-3d98-4240-b845-75e2d5456ac6-logs" (OuterVolumeSpecName: "logs") pod "22a39e3e-3d98-4240-b845-75e2d5456ac6" (UID: "22a39e3e-3d98-4240-b845-75e2d5456ac6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.744550 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22a39e3e-3d98-4240-b845-75e2d5456ac6-kube-api-access-57wmn" (OuterVolumeSpecName: "kube-api-access-57wmn") pod "22a39e3e-3d98-4240-b845-75e2d5456ac6" (UID: "22a39e3e-3d98-4240-b845-75e2d5456ac6"). InnerVolumeSpecName "kube-api-access-57wmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.773954 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-config-data" (OuterVolumeSpecName: "config-data") pod "22a39e3e-3d98-4240-b845-75e2d5456ac6" (UID: "22a39e3e-3d98-4240-b845-75e2d5456ac6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.777553 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22a39e3e-3d98-4240-b845-75e2d5456ac6" (UID: "22a39e3e-3d98-4240-b845-75e2d5456ac6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.810775 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "22a39e3e-3d98-4240-b845-75e2d5456ac6" (UID: "22a39e3e-3d98-4240-b845-75e2d5456ac6"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.822645 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "22a39e3e-3d98-4240-b845-75e2d5456ac6" (UID: "22a39e3e-3d98-4240-b845-75e2d5456ac6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.839751 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d842555c-bc9c-492f-8a1a-db131c37f5ae-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.839807 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d842555c-bc9c-492f-8a1a-db131c37f5ae-logs\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.839859 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtt9t\" (UniqueName: \"kubernetes.io/projected/d842555c-bc9c-492f-8a1a-db131c37f5ae-kube-api-access-gtt9t\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.839907 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d842555c-bc9c-492f-8a1a-db131c37f5ae-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.839968 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d842555c-bc9c-492f-8a1a-db131c37f5ae-config-data\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.840071 4978 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.840086 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57wmn\" (UniqueName: \"kubernetes.io/projected/22a39e3e-3d98-4240-b845-75e2d5456ac6-kube-api-access-57wmn\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.840102 4978 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a39e3e-3d98-4240-b845-75e2d5456ac6-logs\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.840111 4978 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.840120 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.840128 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a39e3e-3d98-4240-b845-75e2d5456ac6-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.840825 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d842555c-bc9c-492f-8a1a-db131c37f5ae-logs\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.843479 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d842555c-bc9c-492f-8a1a-db131c37f5ae-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.844073 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d842555c-bc9c-492f-8a1a-db131c37f5ae-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.851822 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d842555c-bc9c-492f-8a1a-db131c37f5ae-config-data\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.854155 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtt9t\" (UniqueName: \"kubernetes.io/projected/d842555c-bc9c-492f-8a1a-db131c37f5ae-kube-api-access-gtt9t\") pod \"nova-metadata-0\" (UID: \"d842555c-bc9c-492f-8a1a-db131c37f5ae\") " pod="openstack/nova-metadata-0" Feb 25 09:34:07 crc kubenswrapper[4978]: I0225 09:34:07.957465 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.284650 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22a39e3e-3d98-4240-b845-75e2d5456ac6","Type":"ContainerDied","Data":"4f87c102a46a718d446be361acd2387a2e51c003b5f39634a40c6746dd3650b8"} Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.284699 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.286854 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.286887 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"621cc593-53a8-4319-91c7-7ec49790758a","Type":"ContainerStarted","Data":"d694c31fd6a82b7723709677b28af0398ee98eb87c9e91b21060dabe2abfe6c9"} Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.286936 4978 scope.go:117] "RemoveContainer" containerID="9a241ef1e2aa2ef3236b4dbcd65d86341b67ad51d21eae2baaada557ac2a0e09" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.312349 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.312332052 podStartE2EDuration="2.312332052s" podCreationTimestamp="2026-02-25 09:34:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:34:08.309175184 +0000 UTC m=+10141.748431663" watchObservedRunningTime="2026-02-25 09:34:08.312332052 +0000 UTC m=+10141.751588511" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.407702 4978 scope.go:117] "RemoveContainer" containerID="6f3953c76379bf4d3f3b4386500608827daf936585aaadf1474621514f22f5f4" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.424414 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.444359 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.452039 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.453642 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.459743 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.460338 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.460486 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.468151 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.546433 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.557830 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.557868 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2ad5e9e-1c73-4316-940f-5b58e5c39076-logs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.557902 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-public-tls-certs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.557955 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-config-data\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.557980 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wndqs\" (UniqueName: \"kubernetes.io/projected/d2ad5e9e-1c73-4316-940f-5b58e5c39076-kube-api-access-wndqs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.558034 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.659978 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.660117 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.660145 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2ad5e9e-1c73-4316-940f-5b58e5c39076-logs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.660720 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-public-tls-certs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.660812 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-config-data\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.660832 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2ad5e9e-1c73-4316-940f-5b58e5c39076-logs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.660838 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wndqs\" (UniqueName: \"kubernetes.io/projected/d2ad5e9e-1c73-4316-940f-5b58e5c39076-kube-api-access-wndqs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.664470 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.673082 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-config-data\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.673609 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.674750 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2ad5e9e-1c73-4316-940f-5b58e5c39076-public-tls-certs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.676882 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wndqs\" (UniqueName: \"kubernetes.io/projected/d2ad5e9e-1c73-4316-940f-5b58e5c39076-kube-api-access-wndqs\") pod \"nova-api-0\" (UID: \"d2ad5e9e-1c73-4316-940f-5b58e5c39076\") " pod="openstack/nova-api-0" Feb 25 09:34:08 crc kubenswrapper[4978]: I0225 09:34:08.776834 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.089537 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.169271 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-combined-ca-bundle\") pod \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.169347 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvgxc\" (UniqueName: \"kubernetes.io/projected/c14894d9-d8c4-49df-a6e4-d9823efe8f50-kube-api-access-tvgxc\") pod \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.169433 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-config-data\") pod \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\" (UID: \"c14894d9-d8c4-49df-a6e4-d9823efe8f50\") " Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.173107 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14894d9-d8c4-49df-a6e4-d9823efe8f50-kube-api-access-tvgxc" (OuterVolumeSpecName: "kube-api-access-tvgxc") pod "c14894d9-d8c4-49df-a6e4-d9823efe8f50" (UID: "c14894d9-d8c4-49df-a6e4-d9823efe8f50"). InnerVolumeSpecName "kube-api-access-tvgxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.194852 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c14894d9-d8c4-49df-a6e4-d9823efe8f50" (UID: "c14894d9-d8c4-49df-a6e4-d9823efe8f50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.202670 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-config-data" (OuterVolumeSpecName: "config-data") pod "c14894d9-d8c4-49df-a6e4-d9823efe8f50" (UID: "c14894d9-d8c4-49df-a6e4-d9823efe8f50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.271914 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.271949 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvgxc\" (UniqueName: \"kubernetes.io/projected/c14894d9-d8c4-49df-a6e4-d9823efe8f50-kube-api-access-tvgxc\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.271960 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c14894d9-d8c4-49df-a6e4-d9823efe8f50-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.298735 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d842555c-bc9c-492f-8a1a-db131c37f5ae","Type":"ContainerStarted","Data":"218590123a0f7da6ed188ace30ac3fc1c1b0325cc46ffb0d27e75c1f19e7d957"} Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.299085 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d842555c-bc9c-492f-8a1a-db131c37f5ae","Type":"ContainerStarted","Data":"126d0f767863ef21610aa699d6e3ee4bc667468030573845779b7b1796812898"} Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.299129 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d842555c-bc9c-492f-8a1a-db131c37f5ae","Type":"ContainerStarted","Data":"ebbdfb28ffcecda4e020f5d511a3ffce5c7dee2c9b4bc7f30787fa5c231f7926"} Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.302333 4978 generic.go:334] "Generic (PLEG): container finished" podID="c14894d9-d8c4-49df-a6e4-d9823efe8f50" containerID="5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9" exitCode=0 Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.302414 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.302443 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c14894d9-d8c4-49df-a6e4-d9823efe8f50","Type":"ContainerDied","Data":"5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9"} Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.302544 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c14894d9-d8c4-49df-a6e4-d9823efe8f50","Type":"ContainerDied","Data":"6155a77c902297b891038fe236796ea8ede8b14eafd0460abaed99c9b196ee58"} Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.302575 4978 scope.go:117] "RemoveContainer" containerID="5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9" Feb 25 09:34:09 crc kubenswrapper[4978]: W0225 09:34:09.340926 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2ad5e9e_1c73_4316_940f_5b58e5c39076.slice/crio-446f47477bcbe386e9c1e5f46aa5788556dd3717461870dd7f818fdeb473a4e7 WatchSource:0}: Error finding container 446f47477bcbe386e9c1e5f46aa5788556dd3717461870dd7f818fdeb473a4e7: Status 404 returned error can't find the container with id 446f47477bcbe386e9c1e5f46aa5788556dd3717461870dd7f818fdeb473a4e7 Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.346000 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.345979213 podStartE2EDuration="2.345979213s" podCreationTimestamp="2026-02-25 09:34:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:34:09.327843891 +0000 UTC m=+10142.767100370" watchObservedRunningTime="2026-02-25 09:34:09.345979213 +0000 UTC m=+10142.785235672" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.360466 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22a39e3e-3d98-4240-b845-75e2d5456ac6" path="/var/lib/kubelet/pods/22a39e3e-3d98-4240-b845-75e2d5456ac6/volumes" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.361335 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb14a83b-160a-473c-bbe7-d4720ed8bcb4" path="/var/lib/kubelet/pods/bb14a83b-160a-473c-bbe7-d4720ed8bcb4/volumes" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.362319 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.463099 4978 scope.go:117] "RemoveContainer" containerID="5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9" Feb 25 09:34:09 crc kubenswrapper[4978]: E0225 09:34:09.466769 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9\": container with ID starting with 5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9 not found: ID does not exist" containerID="5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9" Feb 25 09:34:09 crc kubenswrapper[4978]: I0225 09:34:09.466819 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9"} err="failed to get container status \"5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9\": rpc error: code = NotFound desc = could not find container \"5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9\": container with ID starting with 5e522e974a498ac599213e01cc1da41ba2a869adfa9ef0ce0eadd25b49522bc9 not found: ID does not exist" Feb 25 09:34:10 crc kubenswrapper[4978]: I0225 09:34:10.337857 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d2ad5e9e-1c73-4316-940f-5b58e5c39076","Type":"ContainerStarted","Data":"247d05ea5f2667547a038c73ca8af16ffc1feb49e748d1cd11344b6d1e14d220"} Feb 25 09:34:10 crc kubenswrapper[4978]: I0225 09:34:10.338441 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d2ad5e9e-1c73-4316-940f-5b58e5c39076","Type":"ContainerStarted","Data":"a3716dcfbe0f78ac87d428911de6cb0fe8a4b899e42c0848dcd71144974d9e0e"} Feb 25 09:34:10 crc kubenswrapper[4978]: I0225 09:34:10.338457 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d2ad5e9e-1c73-4316-940f-5b58e5c39076","Type":"ContainerStarted","Data":"446f47477bcbe386e9c1e5f46aa5788556dd3717461870dd7f818fdeb473a4e7"} Feb 25 09:34:10 crc kubenswrapper[4978]: I0225 09:34:10.380181 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.380157581 podStartE2EDuration="2.380157581s" podCreationTimestamp="2026-02-25 09:34:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:34:10.36561138 +0000 UTC m=+10143.804867859" watchObservedRunningTime="2026-02-25 09:34:10.380157581 +0000 UTC m=+10143.819414060" Feb 25 09:34:12 crc kubenswrapper[4978]: I0225 09:34:12.958075 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 09:34:12 crc kubenswrapper[4978]: I0225 09:34:12.958695 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 25 09:34:15 crc kubenswrapper[4978]: I0225 09:34:15.329046 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:34:15 crc kubenswrapper[4978]: E0225 09:34:15.329824 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:34:15 crc kubenswrapper[4978]: I0225 09:34:15.342785 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 25 09:34:16 crc kubenswrapper[4978]: I0225 09:34:16.696582 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 25 09:34:17 crc kubenswrapper[4978]: I0225 09:34:17.958655 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 09:34:17 crc kubenswrapper[4978]: I0225 09:34:17.960573 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 25 09:34:18 crc kubenswrapper[4978]: I0225 09:34:18.777577 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 09:34:18 crc kubenswrapper[4978]: I0225 09:34:18.777647 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 25 09:34:19 crc kubenswrapper[4978]: I0225 09:34:19.008546 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d842555c-bc9c-492f-8a1a-db131c37f5ae" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.115:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 09:34:19 crc kubenswrapper[4978]: I0225 09:34:19.008545 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d842555c-bc9c-492f-8a1a-db131c37f5ae" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.115:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 09:34:19 crc kubenswrapper[4978]: I0225 09:34:19.790596 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d2ad5e9e-1c73-4316-940f-5b58e5c39076" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.116:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 25 09:34:19 crc kubenswrapper[4978]: I0225 09:34:19.790607 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d2ad5e9e-1c73-4316-940f-5b58e5c39076" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.116:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 25 09:34:26 crc kubenswrapper[4978]: I0225 09:34:26.328588 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:34:27 crc kubenswrapper[4978]: I0225 09:34:27.613873 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"859a4fb26d2b478e0589aefabf4ce096e14e697af774d19395e2b4f3b29e5be9"} Feb 25 09:34:27 crc kubenswrapper[4978]: I0225 09:34:27.969269 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 09:34:27 crc kubenswrapper[4978]: I0225 09:34:27.969940 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 25 09:34:27 crc kubenswrapper[4978]: I0225 09:34:27.979459 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 09:34:27 crc kubenswrapper[4978]: I0225 09:34:27.982193 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 25 09:34:28 crc kubenswrapper[4978]: I0225 09:34:28.787987 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 09:34:28 crc kubenswrapper[4978]: I0225 09:34:28.788563 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 09:34:28 crc kubenswrapper[4978]: I0225 09:34:28.796283 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 25 09:34:28 crc kubenswrapper[4978]: I0225 09:34:28.797119 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 09:34:29 crc kubenswrapper[4978]: I0225 09:34:29.631180 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 25 09:34:29 crc kubenswrapper[4978]: I0225 09:34:29.639781 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 25 09:34:38 crc kubenswrapper[4978]: I0225 09:34:38.017820 4978 scope.go:117] "RemoveContainer" containerID="4aa507cfce53546fef515a839b98038defd2d4a02f451bc73c592d741c1e0417" Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.433971 4978 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podc14894d9-d8c4-49df-a6e4-d9823efe8f50"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podc14894d9-d8c4-49df-a6e4-d9823efe8f50] : Timed out while waiting for systemd to remove kubepods-besteffort-podc14894d9_d8c4_49df_a6e4_d9823efe8f50.slice" Feb 25 09:34:39 crc kubenswrapper[4978]: E0225 09:34:39.434478 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podc14894d9-d8c4-49df-a6e4-d9823efe8f50] : unable to destroy cgroup paths for cgroup [kubepods besteffort podc14894d9-d8c4-49df-a6e4-d9823efe8f50] : Timed out while waiting for systemd to remove kubepods-besteffort-podc14894d9_d8c4_49df_a6e4_d9823efe8f50.slice" pod="openstack/nova-scheduler-0" podUID="c14894d9-d8c4-49df-a6e4-d9823efe8f50" Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.770147 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.839149 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.850265 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.865495 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 09:34:39 crc kubenswrapper[4978]: E0225 09:34:39.866172 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14894d9-d8c4-49df-a6e4-d9823efe8f50" containerName="nova-scheduler-scheduler" Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.866198 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14894d9-d8c4-49df-a6e4-d9823efe8f50" containerName="nova-scheduler-scheduler" Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.866590 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="c14894d9-d8c4-49df-a6e4-d9823efe8f50" containerName="nova-scheduler-scheduler" Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.867743 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.870659 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.882231 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.974663 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd7e65e5-0901-4fa6-95ba-d4649dd9a62e-config-data\") pod \"nova-scheduler-0\" (UID: \"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e\") " pod="openstack/nova-scheduler-0" Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.974742 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl5xz\" (UniqueName: \"kubernetes.io/projected/fd7e65e5-0901-4fa6-95ba-d4649dd9a62e-kube-api-access-hl5xz\") pod \"nova-scheduler-0\" (UID: \"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e\") " pod="openstack/nova-scheduler-0" Feb 25 09:34:39 crc kubenswrapper[4978]: I0225 09:34:39.974781 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd7e65e5-0901-4fa6-95ba-d4649dd9a62e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e\") " pod="openstack/nova-scheduler-0" Feb 25 09:34:40 crc kubenswrapper[4978]: I0225 09:34:40.076586 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd7e65e5-0901-4fa6-95ba-d4649dd9a62e-config-data\") pod \"nova-scheduler-0\" (UID: \"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e\") " pod="openstack/nova-scheduler-0" Feb 25 09:34:40 crc kubenswrapper[4978]: I0225 09:34:40.076641 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl5xz\" (UniqueName: \"kubernetes.io/projected/fd7e65e5-0901-4fa6-95ba-d4649dd9a62e-kube-api-access-hl5xz\") pod \"nova-scheduler-0\" (UID: \"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e\") " pod="openstack/nova-scheduler-0" Feb 25 09:34:40 crc kubenswrapper[4978]: I0225 09:34:40.076671 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd7e65e5-0901-4fa6-95ba-d4649dd9a62e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e\") " pod="openstack/nova-scheduler-0" Feb 25 09:34:40 crc kubenswrapper[4978]: I0225 09:34:40.688620 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd7e65e5-0901-4fa6-95ba-d4649dd9a62e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e\") " pod="openstack/nova-scheduler-0" Feb 25 09:34:40 crc kubenswrapper[4978]: I0225 09:34:40.689104 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl5xz\" (UniqueName: \"kubernetes.io/projected/fd7e65e5-0901-4fa6-95ba-d4649dd9a62e-kube-api-access-hl5xz\") pod \"nova-scheduler-0\" (UID: \"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e\") " pod="openstack/nova-scheduler-0" Feb 25 09:34:40 crc kubenswrapper[4978]: I0225 09:34:40.691649 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd7e65e5-0901-4fa6-95ba-d4649dd9a62e-config-data\") pod \"nova-scheduler-0\" (UID: \"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e\") " pod="openstack/nova-scheduler-0" Feb 25 09:34:40 crc kubenswrapper[4978]: I0225 09:34:40.790883 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 25 09:34:41 crc kubenswrapper[4978]: I0225 09:34:41.241314 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 25 09:34:41 crc kubenswrapper[4978]: I0225 09:34:41.366963 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c14894d9-d8c4-49df-a6e4-d9823efe8f50" path="/var/lib/kubelet/pods/c14894d9-d8c4-49df-a6e4-d9823efe8f50/volumes" Feb 25 09:34:41 crc kubenswrapper[4978]: I0225 09:34:41.787629 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e","Type":"ContainerStarted","Data":"ed36ca0d554f86163b58308380f06864748d5d88018c3f7e8606667b5208e9d0"} Feb 25 09:34:41 crc kubenswrapper[4978]: I0225 09:34:41.787670 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"fd7e65e5-0901-4fa6-95ba-d4649dd9a62e","Type":"ContainerStarted","Data":"3b6eda196ec49c03f5e7488dd7678a46a166deea6c9d542f215ab7484fa014c1"} Feb 25 09:34:41 crc kubenswrapper[4978]: I0225 09:34:41.806529 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.806499365 podStartE2EDuration="2.806499365s" podCreationTimestamp="2026-02-25 09:34:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:34:41.804364219 +0000 UTC m=+10175.243620718" watchObservedRunningTime="2026-02-25 09:34:41.806499365 +0000 UTC m=+10175.245755814" Feb 25 09:34:45 crc kubenswrapper[4978]: I0225 09:34:45.791308 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 25 09:34:50 crc kubenswrapper[4978]: I0225 09:34:50.791106 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 25 09:34:50 crc kubenswrapper[4978]: I0225 09:34:50.866950 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 25 09:34:50 crc kubenswrapper[4978]: I0225 09:34:50.939547 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.194892 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht"] Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.197124 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.200572 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.201872 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.201876 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.202056 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.202066 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.202086 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.210485 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-6svrz" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.217869 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht"] Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.278802 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.278862 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.278924 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.279012 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.279040 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.279056 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.279073 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.279101 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57gdj\" (UniqueName: \"kubernetes.io/projected/03d45c16-5d3e-4342-b59f-453eeba6b33e-kube-api-access-57gdj\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.279134 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.279183 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.279210 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.381460 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.381683 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.382745 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.382837 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.383493 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.383578 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.383667 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57gdj\" (UniqueName: \"kubernetes.io/projected/03d45c16-5d3e-4342-b59f-453eeba6b33e-kube-api-access-57gdj\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.383775 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.383940 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.384579 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.384698 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.386610 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.986808 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.987986 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.988062 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.988453 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.988590 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.988661 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.989064 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:53 crc kubenswrapper[4978]: I0225 09:34:53.999912 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:54 crc kubenswrapper[4978]: I0225 09:34:54.000225 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:54 crc kubenswrapper[4978]: I0225 09:34:54.000590 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57gdj\" (UniqueName: \"kubernetes.io/projected/03d45c16-5d3e-4342-b59f-453eeba6b33e-kube-api-access-57gdj\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:54 crc kubenswrapper[4978]: I0225 09:34:54.120646 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:34:54 crc kubenswrapper[4978]: I0225 09:34:54.781208 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht"] Feb 25 09:34:54 crc kubenswrapper[4978]: I0225 09:34:54.936914 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" event={"ID":"03d45c16-5d3e-4342-b59f-453eeba6b33e","Type":"ContainerStarted","Data":"00f38e8ed0a2ba19cf753f022992bcaec029c9a40381f76e226b953a29235b8f"} Feb 25 09:34:55 crc kubenswrapper[4978]: I0225 09:34:55.962866 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" event={"ID":"03d45c16-5d3e-4342-b59f-453eeba6b33e","Type":"ContainerStarted","Data":"9e232b62243093681b48bac676fc55379435ceccbd61358654f24a99590e4b74"} Feb 25 09:34:55 crc kubenswrapper[4978]: I0225 09:34:55.984525 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" podStartSLOduration=2.528581863 podStartE2EDuration="2.984503156s" podCreationTimestamp="2026-02-25 09:34:53 +0000 UTC" firstStartedPulling="2026-02-25 09:34:54.780622438 +0000 UTC m=+10188.219878917" lastFinishedPulling="2026-02-25 09:34:55.236543711 +0000 UTC m=+10188.675800210" observedRunningTime="2026-02-25 09:34:55.981697799 +0000 UTC m=+10189.420954278" watchObservedRunningTime="2026-02-25 09:34:55.984503156 +0000 UTC m=+10189.423759635" Feb 25 09:36:00 crc kubenswrapper[4978]: I0225 09:36:00.162259 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533536-jdbn9"] Feb 25 09:36:00 crc kubenswrapper[4978]: I0225 09:36:00.164697 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533536-jdbn9" Feb 25 09:36:00 crc kubenswrapper[4978]: I0225 09:36:00.167845 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:36:00 crc kubenswrapper[4978]: I0225 09:36:00.169293 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:36:00 crc kubenswrapper[4978]: I0225 09:36:00.169346 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:36:00 crc kubenswrapper[4978]: I0225 09:36:00.181984 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533536-jdbn9"] Feb 25 09:36:00 crc kubenswrapper[4978]: I0225 09:36:00.362206 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlht7\" (UniqueName: \"kubernetes.io/projected/de6a92a4-917e-49a6-ad0d-e4455c8a0eb8-kube-api-access-zlht7\") pod \"auto-csr-approver-29533536-jdbn9\" (UID: \"de6a92a4-917e-49a6-ad0d-e4455c8a0eb8\") " pod="openshift-infra/auto-csr-approver-29533536-jdbn9" Feb 25 09:36:00 crc kubenswrapper[4978]: I0225 09:36:00.465144 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlht7\" (UniqueName: \"kubernetes.io/projected/de6a92a4-917e-49a6-ad0d-e4455c8a0eb8-kube-api-access-zlht7\") pod \"auto-csr-approver-29533536-jdbn9\" (UID: \"de6a92a4-917e-49a6-ad0d-e4455c8a0eb8\") " pod="openshift-infra/auto-csr-approver-29533536-jdbn9" Feb 25 09:36:00 crc kubenswrapper[4978]: I0225 09:36:00.485075 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlht7\" (UniqueName: \"kubernetes.io/projected/de6a92a4-917e-49a6-ad0d-e4455c8a0eb8-kube-api-access-zlht7\") pod \"auto-csr-approver-29533536-jdbn9\" (UID: \"de6a92a4-917e-49a6-ad0d-e4455c8a0eb8\") " pod="openshift-infra/auto-csr-approver-29533536-jdbn9" Feb 25 09:36:00 crc kubenswrapper[4978]: I0225 09:36:00.494135 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533536-jdbn9" Feb 25 09:36:01 crc kubenswrapper[4978]: I0225 09:36:01.028399 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533536-jdbn9"] Feb 25 09:36:01 crc kubenswrapper[4978]: I0225 09:36:01.825424 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533536-jdbn9" event={"ID":"de6a92a4-917e-49a6-ad0d-e4455c8a0eb8","Type":"ContainerStarted","Data":"0c4e4a4d9a34b21576f85f6266ee4483552ae3d486ce031d2bb89168e878dca0"} Feb 25 09:36:02 crc kubenswrapper[4978]: I0225 09:36:02.838259 4978 generic.go:334] "Generic (PLEG): container finished" podID="de6a92a4-917e-49a6-ad0d-e4455c8a0eb8" containerID="b25b3cb54f308bc7cf51fa83778276997329ef0e190f6c54b92979e345caa0bf" exitCode=0 Feb 25 09:36:02 crc kubenswrapper[4978]: I0225 09:36:02.838522 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533536-jdbn9" event={"ID":"de6a92a4-917e-49a6-ad0d-e4455c8a0eb8","Type":"ContainerDied","Data":"b25b3cb54f308bc7cf51fa83778276997329ef0e190f6c54b92979e345caa0bf"} Feb 25 09:36:04 crc kubenswrapper[4978]: I0225 09:36:04.305010 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533536-jdbn9" Feb 25 09:36:04 crc kubenswrapper[4978]: I0225 09:36:04.362066 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlht7\" (UniqueName: \"kubernetes.io/projected/de6a92a4-917e-49a6-ad0d-e4455c8a0eb8-kube-api-access-zlht7\") pod \"de6a92a4-917e-49a6-ad0d-e4455c8a0eb8\" (UID: \"de6a92a4-917e-49a6-ad0d-e4455c8a0eb8\") " Feb 25 09:36:04 crc kubenswrapper[4978]: I0225 09:36:04.367927 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de6a92a4-917e-49a6-ad0d-e4455c8a0eb8-kube-api-access-zlht7" (OuterVolumeSpecName: "kube-api-access-zlht7") pod "de6a92a4-917e-49a6-ad0d-e4455c8a0eb8" (UID: "de6a92a4-917e-49a6-ad0d-e4455c8a0eb8"). InnerVolumeSpecName "kube-api-access-zlht7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:36:04 crc kubenswrapper[4978]: I0225 09:36:04.465333 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlht7\" (UniqueName: \"kubernetes.io/projected/de6a92a4-917e-49a6-ad0d-e4455c8a0eb8-kube-api-access-zlht7\") on node \"crc\" DevicePath \"\"" Feb 25 09:36:04 crc kubenswrapper[4978]: I0225 09:36:04.864662 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533536-jdbn9" event={"ID":"de6a92a4-917e-49a6-ad0d-e4455c8a0eb8","Type":"ContainerDied","Data":"0c4e4a4d9a34b21576f85f6266ee4483552ae3d486ce031d2bb89168e878dca0"} Feb 25 09:36:04 crc kubenswrapper[4978]: I0225 09:36:04.864742 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c4e4a4d9a34b21576f85f6266ee4483552ae3d486ce031d2bb89168e878dca0" Feb 25 09:36:04 crc kubenswrapper[4978]: I0225 09:36:04.864772 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533536-jdbn9" Feb 25 09:36:05 crc kubenswrapper[4978]: I0225 09:36:05.403714 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533530-xqkpb"] Feb 25 09:36:05 crc kubenswrapper[4978]: I0225 09:36:05.417677 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533530-xqkpb"] Feb 25 09:36:07 crc kubenswrapper[4978]: I0225 09:36:07.353067 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8218e0f8-4605-401c-aeb1-a62888c01b58" path="/var/lib/kubelet/pods/8218e0f8-4605-401c-aeb1-a62888c01b58/volumes" Feb 25 09:36:38 crc kubenswrapper[4978]: I0225 09:36:38.275104 4978 scope.go:117] "RemoveContainer" containerID="a2d30affde97a5f34e759bfe290c97b1625ea333a18a0c2ee4e523a8478455bc" Feb 25 09:36:46 crc kubenswrapper[4978]: I0225 09:36:46.540738 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:36:46 crc kubenswrapper[4978]: I0225 09:36:46.542895 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:37:16 crc kubenswrapper[4978]: I0225 09:37:16.541142 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:37:16 crc kubenswrapper[4978]: I0225 09:37:16.543299 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:37:41 crc kubenswrapper[4978]: I0225 09:37:41.022068 4978 generic.go:334] "Generic (PLEG): container finished" podID="03d45c16-5d3e-4342-b59f-453eeba6b33e" containerID="9e232b62243093681b48bac676fc55379435ceccbd61358654f24a99590e4b74" exitCode=0 Feb 25 09:37:41 crc kubenswrapper[4978]: I0225 09:37:41.022166 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" event={"ID":"03d45c16-5d3e-4342-b59f-453eeba6b33e","Type":"ContainerDied","Data":"9e232b62243093681b48bac676fc55379435ceccbd61358654f24a99590e4b74"} Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.542432 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712116 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-ssh-key-openstack-cell1\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712215 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-0\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712288 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-3\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712422 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57gdj\" (UniqueName: \"kubernetes.io/projected/03d45c16-5d3e-4342-b59f-453eeba6b33e-kube-api-access-57gdj\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712456 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cells-global-config-0\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712485 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-2\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712519 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-1\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712535 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-inventory\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712564 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-combined-ca-bundle\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712583 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-1\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.712600 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-0\") pod \"03d45c16-5d3e-4342-b59f-453eeba6b33e\" (UID: \"03d45c16-5d3e-4342-b59f-453eeba6b33e\") " Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.718322 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.721104 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d45c16-5d3e-4342-b59f-453eeba6b33e-kube-api-access-57gdj" (OuterVolumeSpecName: "kube-api-access-57gdj") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "kube-api-access-57gdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.741592 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.743693 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.750624 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.756270 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.759908 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.766772 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.767235 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.768486 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-inventory" (OuterVolumeSpecName: "inventory") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.786484 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "03d45c16-5d3e-4342-b59f-453eeba6b33e" (UID: "03d45c16-5d3e-4342-b59f-453eeba6b33e"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.815507 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57gdj\" (UniqueName: \"kubernetes.io/projected/03d45c16-5d3e-4342-b59f-453eeba6b33e-kube-api-access-57gdj\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.815927 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.815947 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.815959 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.815972 4978 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-inventory\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.815985 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.815998 4978 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.816011 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.816023 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.816034 4978 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:42 crc kubenswrapper[4978]: I0225 09:37:42.816046 4978 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/03d45c16-5d3e-4342-b59f-453eeba6b33e-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 25 09:37:43 crc kubenswrapper[4978]: I0225 09:37:43.048708 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" event={"ID":"03d45c16-5d3e-4342-b59f-453eeba6b33e","Type":"ContainerDied","Data":"00f38e8ed0a2ba19cf753f022992bcaec029c9a40381f76e226b953a29235b8f"} Feb 25 09:37:43 crc kubenswrapper[4978]: I0225 09:37:43.048744 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00f38e8ed0a2ba19cf753f022992bcaec029c9a40381f76e226b953a29235b8f" Feb 25 09:37:43 crc kubenswrapper[4978]: I0225 09:37:43.048825 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht" Feb 25 09:37:46 crc kubenswrapper[4978]: I0225 09:37:46.540322 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:37:46 crc kubenswrapper[4978]: I0225 09:37:46.540895 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:37:46 crc kubenswrapper[4978]: I0225 09:37:46.540952 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 09:37:46 crc kubenswrapper[4978]: I0225 09:37:46.541691 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"859a4fb26d2b478e0589aefabf4ce096e14e697af774d19395e2b4f3b29e5be9"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 09:37:46 crc kubenswrapper[4978]: I0225 09:37:46.541759 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://859a4fb26d2b478e0589aefabf4ce096e14e697af774d19395e2b4f3b29e5be9" gracePeriod=600 Feb 25 09:37:47 crc kubenswrapper[4978]: I0225 09:37:47.097871 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="859a4fb26d2b478e0589aefabf4ce096e14e697af774d19395e2b4f3b29e5be9" exitCode=0 Feb 25 09:37:47 crc kubenswrapper[4978]: I0225 09:37:47.097957 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"859a4fb26d2b478e0589aefabf4ce096e14e697af774d19395e2b4f3b29e5be9"} Feb 25 09:37:47 crc kubenswrapper[4978]: I0225 09:37:47.098217 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136"} Feb 25 09:37:47 crc kubenswrapper[4978]: I0225 09:37:47.098240 4978 scope.go:117] "RemoveContainer" containerID="17cb4f1d1f99d5f5bb2ddc2d534129158e6df9551997eb9a1a595bfc5a34efec" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.159131 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533538-7swrx"] Feb 25 09:38:00 crc kubenswrapper[4978]: E0225 09:38:00.160091 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d45c16-5d3e-4342-b59f-453eeba6b33e" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.160104 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d45c16-5d3e-4342-b59f-453eeba6b33e" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Feb 25 09:38:00 crc kubenswrapper[4978]: E0225 09:38:00.160134 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de6a92a4-917e-49a6-ad0d-e4455c8a0eb8" containerName="oc" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.160140 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="de6a92a4-917e-49a6-ad0d-e4455c8a0eb8" containerName="oc" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.160397 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="03d45c16-5d3e-4342-b59f-453eeba6b33e" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.160413 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="de6a92a4-917e-49a6-ad0d-e4455c8a0eb8" containerName="oc" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.161124 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533538-7swrx" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.163319 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.164258 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.167619 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.182284 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533538-7swrx"] Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.273313 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h4vx\" (UniqueName: \"kubernetes.io/projected/57853581-8f9a-497b-9d05-ab705c0fdfab-kube-api-access-6h4vx\") pod \"auto-csr-approver-29533538-7swrx\" (UID: \"57853581-8f9a-497b-9d05-ab705c0fdfab\") " pod="openshift-infra/auto-csr-approver-29533538-7swrx" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.374242 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x7xpl"] Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.374911 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h4vx\" (UniqueName: \"kubernetes.io/projected/57853581-8f9a-497b-9d05-ab705c0fdfab-kube-api-access-6h4vx\") pod \"auto-csr-approver-29533538-7swrx\" (UID: \"57853581-8f9a-497b-9d05-ab705c0fdfab\") " pod="openshift-infra/auto-csr-approver-29533538-7swrx" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.379984 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.402763 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h4vx\" (UniqueName: \"kubernetes.io/projected/57853581-8f9a-497b-9d05-ab705c0fdfab-kube-api-access-6h4vx\") pod \"auto-csr-approver-29533538-7swrx\" (UID: \"57853581-8f9a-497b-9d05-ab705c0fdfab\") " pod="openshift-infra/auto-csr-approver-29533538-7swrx" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.406455 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7xpl"] Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.476396 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-utilities\") pod \"redhat-operators-x7xpl\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.476681 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2djrn\" (UniqueName: \"kubernetes.io/projected/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-kube-api-access-2djrn\") pod \"redhat-operators-x7xpl\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.476801 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-catalog-content\") pod \"redhat-operators-x7xpl\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.494185 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533538-7swrx" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.579241 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-utilities\") pod \"redhat-operators-x7xpl\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.578772 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-utilities\") pod \"redhat-operators-x7xpl\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.586543 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2djrn\" (UniqueName: \"kubernetes.io/projected/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-kube-api-access-2djrn\") pod \"redhat-operators-x7xpl\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.586704 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-catalog-content\") pod \"redhat-operators-x7xpl\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.587410 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-catalog-content\") pod \"redhat-operators-x7xpl\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.619481 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2djrn\" (UniqueName: \"kubernetes.io/projected/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-kube-api-access-2djrn\") pod \"redhat-operators-x7xpl\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.740657 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:00 crc kubenswrapper[4978]: I0225 09:38:00.974336 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533538-7swrx"] Feb 25 09:38:01 crc kubenswrapper[4978]: I0225 09:38:01.238933 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7xpl"] Feb 25 09:38:01 crc kubenswrapper[4978]: W0225 09:38:01.240398 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf22d48dc_aa09_46f7_b2a7_e94e66bd6757.slice/crio-8a0ac7a0228c5b50bb66f43aa8e288fe75a7307e47605b1568ef6c8533c3e997 WatchSource:0}: Error finding container 8a0ac7a0228c5b50bb66f43aa8e288fe75a7307e47605b1568ef6c8533c3e997: Status 404 returned error can't find the container with id 8a0ac7a0228c5b50bb66f43aa8e288fe75a7307e47605b1568ef6c8533c3e997 Feb 25 09:38:01 crc kubenswrapper[4978]: I0225 09:38:01.282073 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533538-7swrx" event={"ID":"57853581-8f9a-497b-9d05-ab705c0fdfab","Type":"ContainerStarted","Data":"1e6b09a07ff436ca63c4754f2c5aa70797e55423da66280a656db96a2d72c7a1"} Feb 25 09:38:01 crc kubenswrapper[4978]: I0225 09:38:01.296108 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xpl" event={"ID":"f22d48dc-aa09-46f7-b2a7-e94e66bd6757","Type":"ContainerStarted","Data":"8a0ac7a0228c5b50bb66f43aa8e288fe75a7307e47605b1568ef6c8533c3e997"} Feb 25 09:38:02 crc kubenswrapper[4978]: I0225 09:38:02.310690 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533538-7swrx" event={"ID":"57853581-8f9a-497b-9d05-ab705c0fdfab","Type":"ContainerStarted","Data":"4bc469cac38f40815aa233bedb9c6696570cd77ac80bd2387af5171deabb9161"} Feb 25 09:38:02 crc kubenswrapper[4978]: I0225 09:38:02.315286 4978 generic.go:334] "Generic (PLEG): container finished" podID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerID="b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a" exitCode=0 Feb 25 09:38:02 crc kubenswrapper[4978]: I0225 09:38:02.315356 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xpl" event={"ID":"f22d48dc-aa09-46f7-b2a7-e94e66bd6757","Type":"ContainerDied","Data":"b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a"} Feb 25 09:38:02 crc kubenswrapper[4978]: I0225 09:38:02.339700 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533538-7swrx" podStartSLOduration=1.389839673 podStartE2EDuration="2.339675501s" podCreationTimestamp="2026-02-25 09:38:00 +0000 UTC" firstStartedPulling="2026-02-25 09:38:00.978022138 +0000 UTC m=+10374.417278597" lastFinishedPulling="2026-02-25 09:38:01.927857966 +0000 UTC m=+10375.367114425" observedRunningTime="2026-02-25 09:38:02.335896603 +0000 UTC m=+10375.775153102" watchObservedRunningTime="2026-02-25 09:38:02.339675501 +0000 UTC m=+10375.778931980" Feb 25 09:38:03 crc kubenswrapper[4978]: I0225 09:38:03.329584 4978 generic.go:334] "Generic (PLEG): container finished" podID="57853581-8f9a-497b-9d05-ab705c0fdfab" containerID="4bc469cac38f40815aa233bedb9c6696570cd77ac80bd2387af5171deabb9161" exitCode=0 Feb 25 09:38:03 crc kubenswrapper[4978]: I0225 09:38:03.345640 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533538-7swrx" event={"ID":"57853581-8f9a-497b-9d05-ab705c0fdfab","Type":"ContainerDied","Data":"4bc469cac38f40815aa233bedb9c6696570cd77ac80bd2387af5171deabb9161"} Feb 25 09:38:03 crc kubenswrapper[4978]: I0225 09:38:03.345746 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xpl" event={"ID":"f22d48dc-aa09-46f7-b2a7-e94e66bd6757","Type":"ContainerStarted","Data":"e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487"} Feb 25 09:38:04 crc kubenswrapper[4978]: I0225 09:38:04.834706 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533538-7swrx" Feb 25 09:38:04 crc kubenswrapper[4978]: I0225 09:38:04.984825 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h4vx\" (UniqueName: \"kubernetes.io/projected/57853581-8f9a-497b-9d05-ab705c0fdfab-kube-api-access-6h4vx\") pod \"57853581-8f9a-497b-9d05-ab705c0fdfab\" (UID: \"57853581-8f9a-497b-9d05-ab705c0fdfab\") " Feb 25 09:38:04 crc kubenswrapper[4978]: I0225 09:38:04.991429 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57853581-8f9a-497b-9d05-ab705c0fdfab-kube-api-access-6h4vx" (OuterVolumeSpecName: "kube-api-access-6h4vx") pod "57853581-8f9a-497b-9d05-ab705c0fdfab" (UID: "57853581-8f9a-497b-9d05-ab705c0fdfab"). InnerVolumeSpecName "kube-api-access-6h4vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:38:05 crc kubenswrapper[4978]: I0225 09:38:05.088902 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h4vx\" (UniqueName: \"kubernetes.io/projected/57853581-8f9a-497b-9d05-ab705c0fdfab-kube-api-access-6h4vx\") on node \"crc\" DevicePath \"\"" Feb 25 09:38:05 crc kubenswrapper[4978]: I0225 09:38:05.365013 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533538-7swrx" event={"ID":"57853581-8f9a-497b-9d05-ab705c0fdfab","Type":"ContainerDied","Data":"1e6b09a07ff436ca63c4754f2c5aa70797e55423da66280a656db96a2d72c7a1"} Feb 25 09:38:05 crc kubenswrapper[4978]: I0225 09:38:05.365068 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e6b09a07ff436ca63c4754f2c5aa70797e55423da66280a656db96a2d72c7a1" Feb 25 09:38:05 crc kubenswrapper[4978]: I0225 09:38:05.365074 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533538-7swrx" Feb 25 09:38:05 crc kubenswrapper[4978]: I0225 09:38:05.422037 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533532-vs9kk"] Feb 25 09:38:05 crc kubenswrapper[4978]: I0225 09:38:05.434543 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533532-vs9kk"] Feb 25 09:38:06 crc kubenswrapper[4978]: I0225 09:38:06.378951 4978 generic.go:334] "Generic (PLEG): container finished" podID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerID="e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487" exitCode=0 Feb 25 09:38:06 crc kubenswrapper[4978]: I0225 09:38:06.378992 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xpl" event={"ID":"f22d48dc-aa09-46f7-b2a7-e94e66bd6757","Type":"ContainerDied","Data":"e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487"} Feb 25 09:38:07 crc kubenswrapper[4978]: I0225 09:38:07.349357 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acaac0e4-7ca1-4d72-b1a2-1282a35cb609" path="/var/lib/kubelet/pods/acaac0e4-7ca1-4d72-b1a2-1282a35cb609/volumes" Feb 25 09:38:07 crc kubenswrapper[4978]: I0225 09:38:07.394492 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xpl" event={"ID":"f22d48dc-aa09-46f7-b2a7-e94e66bd6757","Type":"ContainerStarted","Data":"3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536"} Feb 25 09:38:07 crc kubenswrapper[4978]: I0225 09:38:07.437917 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x7xpl" podStartSLOduration=2.804539926 podStartE2EDuration="7.43789352s" podCreationTimestamp="2026-02-25 09:38:00 +0000 UTC" firstStartedPulling="2026-02-25 09:38:02.318250744 +0000 UTC m=+10375.757507223" lastFinishedPulling="2026-02-25 09:38:06.951604358 +0000 UTC m=+10380.390860817" observedRunningTime="2026-02-25 09:38:07.418130565 +0000 UTC m=+10380.857387044" watchObservedRunningTime="2026-02-25 09:38:07.43789352 +0000 UTC m=+10380.877149989" Feb 25 09:38:10 crc kubenswrapper[4978]: I0225 09:38:10.740868 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:10 crc kubenswrapper[4978]: I0225 09:38:10.741488 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:11 crc kubenswrapper[4978]: I0225 09:38:11.787247 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x7xpl" podUID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerName="registry-server" probeResult="failure" output=< Feb 25 09:38:11 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 09:38:11 crc kubenswrapper[4978]: > Feb 25 09:38:20 crc kubenswrapper[4978]: I0225 09:38:20.803069 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:20 crc kubenswrapper[4978]: I0225 09:38:20.863453 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:21 crc kubenswrapper[4978]: I0225 09:38:21.054422 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7xpl"] Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.082445 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x7xpl" podUID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerName="registry-server" containerID="cri-o://3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536" gracePeriod=2 Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.733571 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.788973 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-catalog-content\") pod \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.789183 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2djrn\" (UniqueName: \"kubernetes.io/projected/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-kube-api-access-2djrn\") pod \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.789283 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-utilities\") pod \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\" (UID: \"f22d48dc-aa09-46f7-b2a7-e94e66bd6757\") " Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.791338 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-utilities" (OuterVolumeSpecName: "utilities") pod "f22d48dc-aa09-46f7-b2a7-e94e66bd6757" (UID: "f22d48dc-aa09-46f7-b2a7-e94e66bd6757"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.802690 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-kube-api-access-2djrn" (OuterVolumeSpecName: "kube-api-access-2djrn") pod "f22d48dc-aa09-46f7-b2a7-e94e66bd6757" (UID: "f22d48dc-aa09-46f7-b2a7-e94e66bd6757"). InnerVolumeSpecName "kube-api-access-2djrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.892266 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2djrn\" (UniqueName: \"kubernetes.io/projected/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-kube-api-access-2djrn\") on node \"crc\" DevicePath \"\"" Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.892318 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.939290 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f22d48dc-aa09-46f7-b2a7-e94e66bd6757" (UID: "f22d48dc-aa09-46f7-b2a7-e94e66bd6757"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:38:22 crc kubenswrapper[4978]: I0225 09:38:22.994344 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f22d48dc-aa09-46f7-b2a7-e94e66bd6757-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.100816 4978 generic.go:334] "Generic (PLEG): container finished" podID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerID="3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536" exitCode=0 Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.100927 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7xpl" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.100915 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xpl" event={"ID":"f22d48dc-aa09-46f7-b2a7-e94e66bd6757","Type":"ContainerDied","Data":"3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536"} Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.103598 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7xpl" event={"ID":"f22d48dc-aa09-46f7-b2a7-e94e66bd6757","Type":"ContainerDied","Data":"8a0ac7a0228c5b50bb66f43aa8e288fe75a7307e47605b1568ef6c8533c3e997"} Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.103664 4978 scope.go:117] "RemoveContainer" containerID="3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.166186 4978 scope.go:117] "RemoveContainer" containerID="e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.176249 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7xpl"] Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.196598 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x7xpl"] Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.217210 4978 scope.go:117] "RemoveContainer" containerID="b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.245659 4978 scope.go:117] "RemoveContainer" containerID="3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536" Feb 25 09:38:23 crc kubenswrapper[4978]: E0225 09:38:23.246041 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536\": container with ID starting with 3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536 not found: ID does not exist" containerID="3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.246098 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536"} err="failed to get container status \"3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536\": rpc error: code = NotFound desc = could not find container \"3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536\": container with ID starting with 3fad4c1d639650fc54e7e32d50086130fe9bbe741b3a309e46ca541c4b5ea536 not found: ID does not exist" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.246142 4978 scope.go:117] "RemoveContainer" containerID="e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487" Feb 25 09:38:23 crc kubenswrapper[4978]: E0225 09:38:23.246441 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487\": container with ID starting with e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487 not found: ID does not exist" containerID="e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.246498 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487"} err="failed to get container status \"e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487\": rpc error: code = NotFound desc = could not find container \"e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487\": container with ID starting with e7d68711b963ba8120f94d758e72a20ba79ebd64f58551b5bae4bb6c27ccc487 not found: ID does not exist" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.246530 4978 scope.go:117] "RemoveContainer" containerID="b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a" Feb 25 09:38:23 crc kubenswrapper[4978]: E0225 09:38:23.246755 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a\": container with ID starting with b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a not found: ID does not exist" containerID="b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.246782 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a"} err="failed to get container status \"b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a\": rpc error: code = NotFound desc = could not find container \"b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a\": container with ID starting with b1c27008d0b80a3fddcdc3d849964ee1c4e4bb939e0a247cbb5df6dbbf31653a not found: ID does not exist" Feb 25 09:38:23 crc kubenswrapper[4978]: I0225 09:38:23.345000 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" path="/var/lib/kubelet/pods/f22d48dc-aa09-46f7-b2a7-e94e66bd6757/volumes" Feb 25 09:38:38 crc kubenswrapper[4978]: I0225 09:38:38.438404 4978 scope.go:117] "RemoveContainer" containerID="16c71a6225eb0b9f063723a31a848fab68e785aaa67a6585b86725c3957697ad" Feb 25 09:39:27 crc kubenswrapper[4978]: I0225 09:39:27.611579 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Feb 25 09:39:27 crc kubenswrapper[4978]: I0225 09:39:27.612285 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="4c2aab8b-8456-4e4c-b282-6b7074a116cc" containerName="adoption" containerID="cri-o://2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a" gracePeriod=30 Feb 25 09:39:35 crc kubenswrapper[4978]: I0225 09:39:35.928306 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nrg9l"] Feb 25 09:39:35 crc kubenswrapper[4978]: E0225 09:39:35.929640 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57853581-8f9a-497b-9d05-ab705c0fdfab" containerName="oc" Feb 25 09:39:35 crc kubenswrapper[4978]: I0225 09:39:35.929664 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="57853581-8f9a-497b-9d05-ab705c0fdfab" containerName="oc" Feb 25 09:39:35 crc kubenswrapper[4978]: E0225 09:39:35.929696 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerName="extract-utilities" Feb 25 09:39:35 crc kubenswrapper[4978]: I0225 09:39:35.929708 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerName="extract-utilities" Feb 25 09:39:35 crc kubenswrapper[4978]: E0225 09:39:35.929735 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerName="registry-server" Feb 25 09:39:35 crc kubenswrapper[4978]: I0225 09:39:35.929746 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerName="registry-server" Feb 25 09:39:35 crc kubenswrapper[4978]: E0225 09:39:35.929776 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerName="extract-content" Feb 25 09:39:35 crc kubenswrapper[4978]: I0225 09:39:35.929787 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerName="extract-content" Feb 25 09:39:35 crc kubenswrapper[4978]: I0225 09:39:35.930137 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="57853581-8f9a-497b-9d05-ab705c0fdfab" containerName="oc" Feb 25 09:39:35 crc kubenswrapper[4978]: I0225 09:39:35.930161 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f22d48dc-aa09-46f7-b2a7-e94e66bd6757" containerName="registry-server" Feb 25 09:39:35 crc kubenswrapper[4978]: I0225 09:39:35.932651 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:35 crc kubenswrapper[4978]: I0225 09:39:35.945530 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nrg9l"] Feb 25 09:39:36 crc kubenswrapper[4978]: I0225 09:39:36.021240 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-utilities\") pod \"certified-operators-nrg9l\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:36 crc kubenswrapper[4978]: I0225 09:39:36.021528 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-catalog-content\") pod \"certified-operators-nrg9l\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:36 crc kubenswrapper[4978]: I0225 09:39:36.021594 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6sd7\" (UniqueName: \"kubernetes.io/projected/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-kube-api-access-z6sd7\") pod \"certified-operators-nrg9l\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:36 crc kubenswrapper[4978]: I0225 09:39:36.122757 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-catalog-content\") pod \"certified-operators-nrg9l\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:36 crc kubenswrapper[4978]: I0225 09:39:36.122811 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6sd7\" (UniqueName: \"kubernetes.io/projected/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-kube-api-access-z6sd7\") pod \"certified-operators-nrg9l\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:36 crc kubenswrapper[4978]: I0225 09:39:36.122894 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-utilities\") pod \"certified-operators-nrg9l\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:36 crc kubenswrapper[4978]: I0225 09:39:36.123286 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-catalog-content\") pod \"certified-operators-nrg9l\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:36 crc kubenswrapper[4978]: I0225 09:39:36.123422 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-utilities\") pod \"certified-operators-nrg9l\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:36 crc kubenswrapper[4978]: I0225 09:39:36.150320 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6sd7\" (UniqueName: \"kubernetes.io/projected/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-kube-api-access-z6sd7\") pod \"certified-operators-nrg9l\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:36 crc kubenswrapper[4978]: I0225 09:39:36.286860 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:37 crc kubenswrapper[4978]: I0225 09:39:37.295179 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nrg9l"] Feb 25 09:39:38 crc kubenswrapper[4978]: I0225 09:39:38.109069 4978 generic.go:334] "Generic (PLEG): container finished" podID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerID="674ebca340d6206b773c77e26fd108c23b638c4175e409ef876fa86b1a6fbe69" exitCode=0 Feb 25 09:39:38 crc kubenswrapper[4978]: I0225 09:39:38.109184 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrg9l" event={"ID":"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93","Type":"ContainerDied","Data":"674ebca340d6206b773c77e26fd108c23b638c4175e409ef876fa86b1a6fbe69"} Feb 25 09:39:38 crc kubenswrapper[4978]: I0225 09:39:38.109441 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrg9l" event={"ID":"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93","Type":"ContainerStarted","Data":"12686dadd7488f1e5ff934ce9b1f6772c0594a6c3e5067ddebfe108a0419886e"} Feb 25 09:39:38 crc kubenswrapper[4978]: I0225 09:39:38.113389 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:39:40 crc kubenswrapper[4978]: I0225 09:39:40.137185 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrg9l" event={"ID":"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93","Type":"ContainerStarted","Data":"7ded38c63a31ccfeb78bd89654711ddf6791524a389b5b3b2f86b644cf545220"} Feb 25 09:39:41 crc kubenswrapper[4978]: I0225 09:39:41.153724 4978 generic.go:334] "Generic (PLEG): container finished" podID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerID="7ded38c63a31ccfeb78bd89654711ddf6791524a389b5b3b2f86b644cf545220" exitCode=0 Feb 25 09:39:41 crc kubenswrapper[4978]: I0225 09:39:41.153793 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrg9l" event={"ID":"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93","Type":"ContainerDied","Data":"7ded38c63a31ccfeb78bd89654711ddf6791524a389b5b3b2f86b644cf545220"} Feb 25 09:39:42 crc kubenswrapper[4978]: I0225 09:39:42.175231 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrg9l" event={"ID":"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93","Type":"ContainerStarted","Data":"6ecde531003e0ed55cc86105f8b2f7a7969dc37a61a3079a435407c69061820f"} Feb 25 09:39:42 crc kubenswrapper[4978]: I0225 09:39:42.206620 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nrg9l" podStartSLOduration=3.756265263 podStartE2EDuration="7.206602073s" podCreationTimestamp="2026-02-25 09:39:35 +0000 UTC" firstStartedPulling="2026-02-25 09:39:38.11314876 +0000 UTC m=+10471.552405219" lastFinishedPulling="2026-02-25 09:39:41.56348554 +0000 UTC m=+10475.002742029" observedRunningTime="2026-02-25 09:39:42.199916205 +0000 UTC m=+10475.639172684" watchObservedRunningTime="2026-02-25 09:39:42.206602073 +0000 UTC m=+10475.645858542" Feb 25 09:39:46 crc kubenswrapper[4978]: I0225 09:39:46.287720 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:46 crc kubenswrapper[4978]: I0225 09:39:46.288028 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:46 crc kubenswrapper[4978]: I0225 09:39:46.337228 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:46 crc kubenswrapper[4978]: I0225 09:39:46.540279 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:39:46 crc kubenswrapper[4978]: I0225 09:39:46.540364 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:39:47 crc kubenswrapper[4978]: I0225 09:39:47.349346 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:47 crc kubenswrapper[4978]: I0225 09:39:47.440950 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nrg9l"] Feb 25 09:39:49 crc kubenswrapper[4978]: I0225 09:39:49.286677 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nrg9l" podUID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerName="registry-server" containerID="cri-o://6ecde531003e0ed55cc86105f8b2f7a7969dc37a61a3079a435407c69061820f" gracePeriod=2 Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.307788 4978 generic.go:334] "Generic (PLEG): container finished" podID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerID="6ecde531003e0ed55cc86105f8b2f7a7969dc37a61a3079a435407c69061820f" exitCode=0 Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.307878 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrg9l" event={"ID":"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93","Type":"ContainerDied","Data":"6ecde531003e0ed55cc86105f8b2f7a7969dc37a61a3079a435407c69061820f"} Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.398590 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.465346 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-catalog-content\") pod \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.465650 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-utilities\") pod \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.465802 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6sd7\" (UniqueName: \"kubernetes.io/projected/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-kube-api-access-z6sd7\") pod \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\" (UID: \"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93\") " Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.466890 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-utilities" (OuterVolumeSpecName: "utilities") pod "41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" (UID: "41d5f1f0-8d7a-4c22-8e8e-357fbc748a93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.467546 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.474210 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-kube-api-access-z6sd7" (OuterVolumeSpecName: "kube-api-access-z6sd7") pod "41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" (UID: "41d5f1f0-8d7a-4c22-8e8e-357fbc748a93"). InnerVolumeSpecName "kube-api-access-z6sd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.569735 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6sd7\" (UniqueName: \"kubernetes.io/projected/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-kube-api-access-z6sd7\") on node \"crc\" DevicePath \"\"" Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.573071 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" (UID: "41d5f1f0-8d7a-4c22-8e8e-357fbc748a93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:39:50 crc kubenswrapper[4978]: I0225 09:39:50.673008 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:39:51 crc kubenswrapper[4978]: I0225 09:39:51.321246 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nrg9l" event={"ID":"41d5f1f0-8d7a-4c22-8e8e-357fbc748a93","Type":"ContainerDied","Data":"12686dadd7488f1e5ff934ce9b1f6772c0594a6c3e5067ddebfe108a0419886e"} Feb 25 09:39:51 crc kubenswrapper[4978]: I0225 09:39:51.321535 4978 scope.go:117] "RemoveContainer" containerID="6ecde531003e0ed55cc86105f8b2f7a7969dc37a61a3079a435407c69061820f" Feb 25 09:39:51 crc kubenswrapper[4978]: I0225 09:39:51.321301 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nrg9l" Feb 25 09:39:51 crc kubenswrapper[4978]: I0225 09:39:51.353561 4978 scope.go:117] "RemoveContainer" containerID="7ded38c63a31ccfeb78bd89654711ddf6791524a389b5b3b2f86b644cf545220" Feb 25 09:39:51 crc kubenswrapper[4978]: I0225 09:39:51.371504 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nrg9l"] Feb 25 09:39:51 crc kubenswrapper[4978]: I0225 09:39:51.383658 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nrg9l"] Feb 25 09:39:51 crc kubenswrapper[4978]: I0225 09:39:51.389057 4978 scope.go:117] "RemoveContainer" containerID="674ebca340d6206b773c77e26fd108c23b638c4175e409ef876fa86b1a6fbe69" Feb 25 09:39:53 crc kubenswrapper[4978]: I0225 09:39:53.338540 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" path="/var/lib/kubelet/pods/41d5f1f0-8d7a-4c22-8e8e-357fbc748a93/volumes" Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.245763 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.351951 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6207369-0c50-4031-8864-29f04e3de609\") pod \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\" (UID: \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\") " Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.352059 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmscv\" (UniqueName: \"kubernetes.io/projected/4c2aab8b-8456-4e4c-b282-6b7074a116cc-kube-api-access-rmscv\") pod \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\" (UID: \"4c2aab8b-8456-4e4c-b282-6b7074a116cc\") " Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.363748 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c2aab8b-8456-4e4c-b282-6b7074a116cc-kube-api-access-rmscv" (OuterVolumeSpecName: "kube-api-access-rmscv") pod "4c2aab8b-8456-4e4c-b282-6b7074a116cc" (UID: "4c2aab8b-8456-4e4c-b282-6b7074a116cc"). InnerVolumeSpecName "kube-api-access-rmscv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.399110 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6207369-0c50-4031-8864-29f04e3de609" (OuterVolumeSpecName: "mariadb-data") pod "4c2aab8b-8456-4e4c-b282-6b7074a116cc" (UID: "4c2aab8b-8456-4e4c-b282-6b7074a116cc"). InnerVolumeSpecName "pvc-a6207369-0c50-4031-8864-29f04e3de609". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.403845 4978 generic.go:334] "Generic (PLEG): container finished" podID="4c2aab8b-8456-4e4c-b282-6b7074a116cc" containerID="2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a" exitCode=137 Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.403906 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"4c2aab8b-8456-4e4c-b282-6b7074a116cc","Type":"ContainerDied","Data":"2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a"} Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.403942 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"4c2aab8b-8456-4e4c-b282-6b7074a116cc","Type":"ContainerDied","Data":"2699cc89291b12f2fee518daf3f8115c24c110ba0bd2753e738bfe8a72b6881b"} Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.403966 4978 scope.go:117] "RemoveContainer" containerID="2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a" Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.404146 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.459891 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmscv\" (UniqueName: \"kubernetes.io/projected/4c2aab8b-8456-4e4c-b282-6b7074a116cc-kube-api-access-rmscv\") on node \"crc\" DevicePath \"\"" Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.460199 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-a6207369-0c50-4031-8864-29f04e3de609\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6207369-0c50-4031-8864-29f04e3de609\") on node \"crc\" " Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.500740 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.515512 4978 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.515846 4978 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-a6207369-0c50-4031-8864-29f04e3de609" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6207369-0c50-4031-8864-29f04e3de609") on node "crc" Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.516167 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.517802 4978 scope.go:117] "RemoveContainer" containerID="2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a" Feb 25 09:39:58 crc kubenswrapper[4978]: E0225 09:39:58.519580 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a\": container with ID starting with 2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a not found: ID does not exist" containerID="2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a" Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.519633 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a"} err="failed to get container status \"2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a\": rpc error: code = NotFound desc = could not find container \"2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a\": container with ID starting with 2dbaf5493ae291c49179ba981fc8f48100665bb18860b61a9cfa973568d2780a not found: ID does not exist" Feb 25 09:39:58 crc kubenswrapper[4978]: I0225 09:39:58.562302 4978 reconciler_common.go:293] "Volume detached for volume \"pvc-a6207369-0c50-4031-8864-29f04e3de609\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a6207369-0c50-4031-8864-29f04e3de609\") on node \"crc\" DevicePath \"\"" Feb 25 09:39:59 crc kubenswrapper[4978]: I0225 09:39:59.163739 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Feb 25 09:39:59 crc kubenswrapper[4978]: I0225 09:39:59.164070 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="d8a2072e-806f-4e38-8cee-18b683945ad2" containerName="adoption" containerID="cri-o://449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9" gracePeriod=30 Feb 25 09:39:59 crc kubenswrapper[4978]: I0225 09:39:59.341317 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c2aab8b-8456-4e4c-b282-6b7074a116cc" path="/var/lib/kubelet/pods/4c2aab8b-8456-4e4c-b282-6b7074a116cc/volumes" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.175022 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533540-w295f"] Feb 25 09:40:00 crc kubenswrapper[4978]: E0225 09:40:00.175496 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2aab8b-8456-4e4c-b282-6b7074a116cc" containerName="adoption" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.175511 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2aab8b-8456-4e4c-b282-6b7074a116cc" containerName="adoption" Feb 25 09:40:00 crc kubenswrapper[4978]: E0225 09:40:00.175524 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerName="extract-content" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.175530 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerName="extract-content" Feb 25 09:40:00 crc kubenswrapper[4978]: E0225 09:40:00.175546 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerName="registry-server" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.175554 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerName="registry-server" Feb 25 09:40:00 crc kubenswrapper[4978]: E0225 09:40:00.175565 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerName="extract-utilities" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.175572 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerName="extract-utilities" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.175753 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c2aab8b-8456-4e4c-b282-6b7074a116cc" containerName="adoption" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.175775 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="41d5f1f0-8d7a-4c22-8e8e-357fbc748a93" containerName="registry-server" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.176447 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533540-w295f" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.179097 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.179106 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.180083 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.197381 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9rts\" (UniqueName: \"kubernetes.io/projected/78ee54e7-3195-40bb-9aaf-540a08cb620c-kube-api-access-r9rts\") pod \"auto-csr-approver-29533540-w295f\" (UID: \"78ee54e7-3195-40bb-9aaf-540a08cb620c\") " pod="openshift-infra/auto-csr-approver-29533540-w295f" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.204385 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533540-w295f"] Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.298683 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9rts\" (UniqueName: \"kubernetes.io/projected/78ee54e7-3195-40bb-9aaf-540a08cb620c-kube-api-access-r9rts\") pod \"auto-csr-approver-29533540-w295f\" (UID: \"78ee54e7-3195-40bb-9aaf-540a08cb620c\") " pod="openshift-infra/auto-csr-approver-29533540-w295f" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.325274 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9rts\" (UniqueName: \"kubernetes.io/projected/78ee54e7-3195-40bb-9aaf-540a08cb620c-kube-api-access-r9rts\") pod \"auto-csr-approver-29533540-w295f\" (UID: \"78ee54e7-3195-40bb-9aaf-540a08cb620c\") " pod="openshift-infra/auto-csr-approver-29533540-w295f" Feb 25 09:40:00 crc kubenswrapper[4978]: I0225 09:40:00.503017 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533540-w295f" Feb 25 09:40:01 crc kubenswrapper[4978]: I0225 09:40:01.026921 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533540-w295f"] Feb 25 09:40:01 crc kubenswrapper[4978]: I0225 09:40:01.432862 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533540-w295f" event={"ID":"78ee54e7-3195-40bb-9aaf-540a08cb620c","Type":"ContainerStarted","Data":"4a47b26c6d3ee8b2ab64c87fcb6c2026518da221509e3751eead58b311d5f85c"} Feb 25 09:40:03 crc kubenswrapper[4978]: I0225 09:40:03.461485 4978 generic.go:334] "Generic (PLEG): container finished" podID="78ee54e7-3195-40bb-9aaf-540a08cb620c" containerID="698f8d7ffcf03a62b5b370372410e46e26337b4089f2d94c5f15bf70f5be019f" exitCode=0 Feb 25 09:40:03 crc kubenswrapper[4978]: I0225 09:40:03.461560 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533540-w295f" event={"ID":"78ee54e7-3195-40bb-9aaf-540a08cb620c","Type":"ContainerDied","Data":"698f8d7ffcf03a62b5b370372410e46e26337b4089f2d94c5f15bf70f5be019f"} Feb 25 09:40:04 crc kubenswrapper[4978]: I0225 09:40:04.938911 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533540-w295f" Feb 25 09:40:05 crc kubenswrapper[4978]: I0225 09:40:05.101109 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9rts\" (UniqueName: \"kubernetes.io/projected/78ee54e7-3195-40bb-9aaf-540a08cb620c-kube-api-access-r9rts\") pod \"78ee54e7-3195-40bb-9aaf-540a08cb620c\" (UID: \"78ee54e7-3195-40bb-9aaf-540a08cb620c\") " Feb 25 09:40:05 crc kubenswrapper[4978]: I0225 09:40:05.108590 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78ee54e7-3195-40bb-9aaf-540a08cb620c-kube-api-access-r9rts" (OuterVolumeSpecName: "kube-api-access-r9rts") pod "78ee54e7-3195-40bb-9aaf-540a08cb620c" (UID: "78ee54e7-3195-40bb-9aaf-540a08cb620c"). InnerVolumeSpecName "kube-api-access-r9rts". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:40:05 crc kubenswrapper[4978]: I0225 09:40:05.203626 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9rts\" (UniqueName: \"kubernetes.io/projected/78ee54e7-3195-40bb-9aaf-540a08cb620c-kube-api-access-r9rts\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:05 crc kubenswrapper[4978]: I0225 09:40:05.491130 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533540-w295f" event={"ID":"78ee54e7-3195-40bb-9aaf-540a08cb620c","Type":"ContainerDied","Data":"4a47b26c6d3ee8b2ab64c87fcb6c2026518da221509e3751eead58b311d5f85c"} Feb 25 09:40:05 crc kubenswrapper[4978]: I0225 09:40:05.491907 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a47b26c6d3ee8b2ab64c87fcb6c2026518da221509e3751eead58b311d5f85c" Feb 25 09:40:05 crc kubenswrapper[4978]: I0225 09:40:05.491674 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533540-w295f" Feb 25 09:40:06 crc kubenswrapper[4978]: I0225 09:40:06.037437 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533534-wpxld"] Feb 25 09:40:06 crc kubenswrapper[4978]: I0225 09:40:06.063539 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533534-wpxld"] Feb 25 09:40:07 crc kubenswrapper[4978]: I0225 09:40:07.356967 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fed10669-c439-41b5-bab7-081176ff23b5" path="/var/lib/kubelet/pods/fed10669-c439-41b5-bab7-081176ff23b5/volumes" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.693959 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7vnhj"] Feb 25 09:40:14 crc kubenswrapper[4978]: E0225 09:40:14.695664 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78ee54e7-3195-40bb-9aaf-540a08cb620c" containerName="oc" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.695892 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="78ee54e7-3195-40bb-9aaf-540a08cb620c" containerName="oc" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.696528 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="78ee54e7-3195-40bb-9aaf-540a08cb620c" containerName="oc" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.699899 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.712362 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7vnhj"] Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.812086 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-utilities\") pod \"community-operators-7vnhj\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.812299 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p4bl\" (UniqueName: \"kubernetes.io/projected/4b049414-0424-41cf-a2ce-d2f03bbbd16c-kube-api-access-9p4bl\") pod \"community-operators-7vnhj\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.812341 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-catalog-content\") pod \"community-operators-7vnhj\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.914746 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-utilities\") pod \"community-operators-7vnhj\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.915035 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p4bl\" (UniqueName: \"kubernetes.io/projected/4b049414-0424-41cf-a2ce-d2f03bbbd16c-kube-api-access-9p4bl\") pod \"community-operators-7vnhj\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.915103 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-catalog-content\") pod \"community-operators-7vnhj\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.915433 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-utilities\") pod \"community-operators-7vnhj\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.915888 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-catalog-content\") pod \"community-operators-7vnhj\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:14 crc kubenswrapper[4978]: I0225 09:40:14.941121 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p4bl\" (UniqueName: \"kubernetes.io/projected/4b049414-0424-41cf-a2ce-d2f03bbbd16c-kube-api-access-9p4bl\") pod \"community-operators-7vnhj\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:15 crc kubenswrapper[4978]: I0225 09:40:15.052070 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:15 crc kubenswrapper[4978]: I0225 09:40:15.672437 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7vnhj"] Feb 25 09:40:16 crc kubenswrapper[4978]: I0225 09:40:16.540069 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:40:16 crc kubenswrapper[4978]: I0225 09:40:16.540470 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:40:16 crc kubenswrapper[4978]: I0225 09:40:16.661137 4978 generic.go:334] "Generic (PLEG): container finished" podID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerID="6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0" exitCode=0 Feb 25 09:40:16 crc kubenswrapper[4978]: I0225 09:40:16.661184 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vnhj" event={"ID":"4b049414-0424-41cf-a2ce-d2f03bbbd16c","Type":"ContainerDied","Data":"6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0"} Feb 25 09:40:16 crc kubenswrapper[4978]: I0225 09:40:16.661211 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vnhj" event={"ID":"4b049414-0424-41cf-a2ce-d2f03bbbd16c","Type":"ContainerStarted","Data":"33797fe43d8b4b0c2a84fe57846ff17b2f788a30148fa7bfba6b507dd878a587"} Feb 25 09:40:17 crc kubenswrapper[4978]: I0225 09:40:17.677081 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vnhj" event={"ID":"4b049414-0424-41cf-a2ce-d2f03bbbd16c","Type":"ContainerStarted","Data":"d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0"} Feb 25 09:40:18 crc kubenswrapper[4978]: I0225 09:40:18.729781 4978 generic.go:334] "Generic (PLEG): container finished" podID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerID="d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0" exitCode=0 Feb 25 09:40:18 crc kubenswrapper[4978]: I0225 09:40:18.729864 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vnhj" event={"ID":"4b049414-0424-41cf-a2ce-d2f03bbbd16c","Type":"ContainerDied","Data":"d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0"} Feb 25 09:40:19 crc kubenswrapper[4978]: I0225 09:40:19.743337 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vnhj" event={"ID":"4b049414-0424-41cf-a2ce-d2f03bbbd16c","Type":"ContainerStarted","Data":"328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7"} Feb 25 09:40:19 crc kubenswrapper[4978]: I0225 09:40:19.775641 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7vnhj" podStartSLOduration=3.298522014 podStartE2EDuration="5.775614048s" podCreationTimestamp="2026-02-25 09:40:14 +0000 UTC" firstStartedPulling="2026-02-25 09:40:16.665440094 +0000 UTC m=+10510.104696553" lastFinishedPulling="2026-02-25 09:40:19.142532138 +0000 UTC m=+10512.581788587" observedRunningTime="2026-02-25 09:40:19.770625983 +0000 UTC m=+10513.209882452" watchObservedRunningTime="2026-02-25 09:40:19.775614048 +0000 UTC m=+10513.214870547" Feb 25 09:40:25 crc kubenswrapper[4978]: I0225 09:40:25.052419 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:25 crc kubenswrapper[4978]: I0225 09:40:25.053004 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:25 crc kubenswrapper[4978]: I0225 09:40:25.123273 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:25 crc kubenswrapper[4978]: I0225 09:40:25.892254 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:25 crc kubenswrapper[4978]: I0225 09:40:25.961659 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7vnhj"] Feb 25 09:40:27 crc kubenswrapper[4978]: I0225 09:40:27.840532 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7vnhj" podUID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerName="registry-server" containerID="cri-o://328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7" gracePeriod=2 Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.337903 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.524356 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-utilities\") pod \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.524796 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p4bl\" (UniqueName: \"kubernetes.io/projected/4b049414-0424-41cf-a2ce-d2f03bbbd16c-kube-api-access-9p4bl\") pod \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.525011 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-catalog-content\") pod \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\" (UID: \"4b049414-0424-41cf-a2ce-d2f03bbbd16c\") " Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.526322 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-utilities" (OuterVolumeSpecName: "utilities") pod "4b049414-0424-41cf-a2ce-d2f03bbbd16c" (UID: "4b049414-0424-41cf-a2ce-d2f03bbbd16c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.531592 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b049414-0424-41cf-a2ce-d2f03bbbd16c-kube-api-access-9p4bl" (OuterVolumeSpecName: "kube-api-access-9p4bl") pod "4b049414-0424-41cf-a2ce-d2f03bbbd16c" (UID: "4b049414-0424-41cf-a2ce-d2f03bbbd16c"). InnerVolumeSpecName "kube-api-access-9p4bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.628329 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.628626 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p4bl\" (UniqueName: \"kubernetes.io/projected/4b049414-0424-41cf-a2ce-d2f03bbbd16c-kube-api-access-9p4bl\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.695742 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b049414-0424-41cf-a2ce-d2f03bbbd16c" (UID: "4b049414-0424-41cf-a2ce-d2f03bbbd16c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.731673 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b049414-0424-41cf-a2ce-d2f03bbbd16c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.853674 4978 generic.go:334] "Generic (PLEG): container finished" podID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerID="328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7" exitCode=0 Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.853725 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vnhj" event={"ID":"4b049414-0424-41cf-a2ce-d2f03bbbd16c","Type":"ContainerDied","Data":"328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7"} Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.853766 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7vnhj" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.853797 4978 scope.go:117] "RemoveContainer" containerID="328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.853782 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7vnhj" event={"ID":"4b049414-0424-41cf-a2ce-d2f03bbbd16c","Type":"ContainerDied","Data":"33797fe43d8b4b0c2a84fe57846ff17b2f788a30148fa7bfba6b507dd878a587"} Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.894630 4978 scope.go:117] "RemoveContainer" containerID="d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.898747 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7vnhj"] Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.912323 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7vnhj"] Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.916444 4978 scope.go:117] "RemoveContainer" containerID="6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.968293 4978 scope.go:117] "RemoveContainer" containerID="328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7" Feb 25 09:40:28 crc kubenswrapper[4978]: E0225 09:40:28.969160 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7\": container with ID starting with 328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7 not found: ID does not exist" containerID="328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.969196 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7"} err="failed to get container status \"328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7\": rpc error: code = NotFound desc = could not find container \"328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7\": container with ID starting with 328b03294bdfe3f0236579cd8bede41c8f6c23ab6a29bbd80978edb89a3474c7 not found: ID does not exist" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.969222 4978 scope.go:117] "RemoveContainer" containerID="d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0" Feb 25 09:40:28 crc kubenswrapper[4978]: E0225 09:40:28.969942 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0\": container with ID starting with d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0 not found: ID does not exist" containerID="d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.970032 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0"} err="failed to get container status \"d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0\": rpc error: code = NotFound desc = could not find container \"d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0\": container with ID starting with d9362c9db3a45387c842b45f7c712e569d68ab571f1a2797256d52dea39790d0 not found: ID does not exist" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.970076 4978 scope.go:117] "RemoveContainer" containerID="6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0" Feb 25 09:40:28 crc kubenswrapper[4978]: E0225 09:40:28.970930 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0\": container with ID starting with 6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0 not found: ID does not exist" containerID="6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0" Feb 25 09:40:28 crc kubenswrapper[4978]: I0225 09:40:28.970972 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0"} err="failed to get container status \"6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0\": rpc error: code = NotFound desc = could not find container \"6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0\": container with ID starting with 6a47c1565413d3c895f93b3cca5d88e4d8f7713cbd6df432858dee26b331ddc0 not found: ID does not exist" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.343614 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" path="/var/lib/kubelet/pods/4b049414-0424-41cf-a2ce-d2f03bbbd16c/volumes" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.746048 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.867137 4978 generic.go:334] "Generic (PLEG): container finished" podID="d8a2072e-806f-4e38-8cee-18b683945ad2" containerID="449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9" exitCode=137 Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.867221 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"d8a2072e-806f-4e38-8cee-18b683945ad2","Type":"ContainerDied","Data":"449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9"} Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.867262 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"d8a2072e-806f-4e38-8cee-18b683945ad2","Type":"ContainerDied","Data":"668da62b3b748145515f268b5e691cab63eaf7693cf3f093494274f54e6aa874"} Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.867291 4978 scope.go:117] "RemoveContainer" containerID="449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.867471 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.872306 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\") pod \"d8a2072e-806f-4e38-8cee-18b683945ad2\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.872360 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/d8a2072e-806f-4e38-8cee-18b683945ad2-ovn-data-cert\") pod \"d8a2072e-806f-4e38-8cee-18b683945ad2\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.872619 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52nnr\" (UniqueName: \"kubernetes.io/projected/d8a2072e-806f-4e38-8cee-18b683945ad2-kube-api-access-52nnr\") pod \"d8a2072e-806f-4e38-8cee-18b683945ad2\" (UID: \"d8a2072e-806f-4e38-8cee-18b683945ad2\") " Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.893098 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8a2072e-806f-4e38-8cee-18b683945ad2-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "d8a2072e-806f-4e38-8cee-18b683945ad2" (UID: "d8a2072e-806f-4e38-8cee-18b683945ad2"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.893658 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8a2072e-806f-4e38-8cee-18b683945ad2-kube-api-access-52nnr" (OuterVolumeSpecName: "kube-api-access-52nnr") pod "d8a2072e-806f-4e38-8cee-18b683945ad2" (UID: "d8a2072e-806f-4e38-8cee-18b683945ad2"). InnerVolumeSpecName "kube-api-access-52nnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.898322 4978 scope.go:117] "RemoveContainer" containerID="449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9" Feb 25 09:40:29 crc kubenswrapper[4978]: E0225 09:40:29.898872 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9\": container with ID starting with 449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9 not found: ID does not exist" containerID="449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.898913 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9"} err="failed to get container status \"449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9\": rpc error: code = NotFound desc = could not find container \"449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9\": container with ID starting with 449fe0d860025b777fb260a537c168918a2d093df11c48a4703b279f161e26b9 not found: ID does not exist" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.906735 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e" (OuterVolumeSpecName: "ovn-data") pod "d8a2072e-806f-4e38-8cee-18b683945ad2" (UID: "d8a2072e-806f-4e38-8cee-18b683945ad2"). InnerVolumeSpecName "pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.975129 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\") on node \"crc\" " Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.975176 4978 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/d8a2072e-806f-4e38-8cee-18b683945ad2-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:29 crc kubenswrapper[4978]: I0225 09:40:29.975189 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52nnr\" (UniqueName: \"kubernetes.io/projected/d8a2072e-806f-4e38-8cee-18b683945ad2-kube-api-access-52nnr\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:30 crc kubenswrapper[4978]: I0225 09:40:30.007282 4978 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Feb 25 09:40:30 crc kubenswrapper[4978]: I0225 09:40:30.007441 4978 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e") on node "crc" Feb 25 09:40:30 crc kubenswrapper[4978]: I0225 09:40:30.077232 4978 reconciler_common.go:293] "Volume detached for volume \"pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-704c8dc0-d261-4275-b832-eacaf23b9f0e\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:30 crc kubenswrapper[4978]: I0225 09:40:30.223227 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Feb 25 09:40:30 crc kubenswrapper[4978]: I0225 09:40:30.233805 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Feb 25 09:40:31 crc kubenswrapper[4978]: I0225 09:40:31.340440 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8a2072e-806f-4e38-8cee-18b683945ad2" path="/var/lib/kubelet/pods/d8a2072e-806f-4e38-8cee-18b683945ad2/volumes" Feb 25 09:40:38 crc kubenswrapper[4978]: I0225 09:40:38.589040 4978 scope.go:117] "RemoveContainer" containerID="4705e77197a4af0d64e72956803bbc8fd40c56a0dea45bc7278fa28a769d0b7e" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.735787 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-gtz5n"] Feb 25 09:40:41 crc kubenswrapper[4978]: E0225 09:40:41.737396 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8a2072e-806f-4e38-8cee-18b683945ad2" containerName="adoption" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.737420 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8a2072e-806f-4e38-8cee-18b683945ad2" containerName="adoption" Feb 25 09:40:41 crc kubenswrapper[4978]: E0225 09:40:41.737448 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerName="extract-utilities" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.737457 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerName="extract-utilities" Feb 25 09:40:41 crc kubenswrapper[4978]: E0225 09:40:41.737484 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerName="extract-content" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.737497 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerName="extract-content" Feb 25 09:40:41 crc kubenswrapper[4978]: E0225 09:40:41.737515 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerName="registry-server" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.737522 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerName="registry-server" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.738550 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b049414-0424-41cf-a2ce-d2f03bbbd16c" containerName="registry-server" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.738660 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8a2072e-806f-4e38-8cee-18b683945ad2" containerName="adoption" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.753928 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.779357 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.784175 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.806583 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-gtz5n"] Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.858110 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.858266 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-swiftconf\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.858290 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-etc-swift\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.858313 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n9g5\" (UniqueName: \"kubernetes.io/projected/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-kube-api-access-6n9g5\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.858362 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-dispersionconf\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.858444 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-scripts\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.858484 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-ring-data-devices\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.961520 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.961729 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-swiftconf\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.961763 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-etc-swift\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.961804 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n9g5\" (UniqueName: \"kubernetes.io/projected/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-kube-api-access-6n9g5\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.961884 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-dispersionconf\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.961995 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-scripts\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.962072 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-ring-data-devices\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.962451 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-etc-swift\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.962942 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-scripts\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.963030 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-ring-data-devices\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.967821 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-dispersionconf\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.972656 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-swiftconf\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.976663 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:41 crc kubenswrapper[4978]: I0225 09:40:41.999022 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n9g5\" (UniqueName: \"kubernetes.io/projected/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-kube-api-access-6n9g5\") pod \"swift-ring-rebalance-debug-gtz5n\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:42 crc kubenswrapper[4978]: I0225 09:40:42.117254 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:42 crc kubenswrapper[4978]: I0225 09:40:42.613701 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-gtz5n"] Feb 25 09:40:43 crc kubenswrapper[4978]: I0225 09:40:43.049985 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-gtz5n" event={"ID":"aaf39141-4f8a-4e38-b9f7-dc4c1151948d","Type":"ContainerStarted","Data":"e0474d67516f6bc53643dffa1a1dff5c9d424cac67f6bd0f454320eb14ac1a48"} Feb 25 09:40:43 crc kubenswrapper[4978]: I0225 09:40:43.050341 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-gtz5n" event={"ID":"aaf39141-4f8a-4e38-b9f7-dc4c1151948d","Type":"ContainerStarted","Data":"ce2bbd050a0127c046c33bdcf248a3656e2281825d1dabb6d54217a48b7c0319"} Feb 25 09:40:43 crc kubenswrapper[4978]: I0225 09:40:43.079930 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-gtz5n" podStartSLOduration=2.079914786 podStartE2EDuration="2.079914786s" podCreationTimestamp="2026-02-25 09:40:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:40:43.069909645 +0000 UTC m=+10536.509166114" watchObservedRunningTime="2026-02-25 09:40:43.079914786 +0000 UTC m=+10536.519171245" Feb 25 09:40:44 crc kubenswrapper[4978]: I0225 09:40:44.065966 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-gtz5n" event={"ID":"aaf39141-4f8a-4e38-b9f7-dc4c1151948d","Type":"ContainerDied","Data":"e0474d67516f6bc53643dffa1a1dff5c9d424cac67f6bd0f454320eb14ac1a48"} Feb 25 09:40:44 crc kubenswrapper[4978]: I0225 09:40:44.066457 4978 generic.go:334] "Generic (PLEG): container finished" podID="aaf39141-4f8a-4e38-b9f7-dc4c1151948d" containerID="e0474d67516f6bc53643dffa1a1dff5c9d424cac67f6bd0f454320eb14ac1a48" exitCode=0 Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.313103 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.349019 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-gtz5n"] Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.358232 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-gtz5n"] Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.455550 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-dispersionconf\") pod \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.455901 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-etc-swift\") pod \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.455923 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-ring-data-devices\") pod \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.455989 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-swiftconf\") pod \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.456007 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-scripts\") pod \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.456153 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-combined-ca-bundle\") pod \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.456189 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n9g5\" (UniqueName: \"kubernetes.io/projected/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-kube-api-access-6n9g5\") pod \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\" (UID: \"aaf39141-4f8a-4e38-b9f7-dc4c1151948d\") " Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.456658 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "aaf39141-4f8a-4e38-b9f7-dc4c1151948d" (UID: "aaf39141-4f8a-4e38-b9f7-dc4c1151948d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.457106 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "aaf39141-4f8a-4e38-b9f7-dc4c1151948d" (UID: "aaf39141-4f8a-4e38-b9f7-dc4c1151948d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.472030 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-kube-api-access-6n9g5" (OuterVolumeSpecName: "kube-api-access-6n9g5") pod "aaf39141-4f8a-4e38-b9f7-dc4c1151948d" (UID: "aaf39141-4f8a-4e38-b9f7-dc4c1151948d"). InnerVolumeSpecName "kube-api-access-6n9g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.485483 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-scripts" (OuterVolumeSpecName: "scripts") pod "aaf39141-4f8a-4e38-b9f7-dc4c1151948d" (UID: "aaf39141-4f8a-4e38-b9f7-dc4c1151948d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.496741 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "aaf39141-4f8a-4e38-b9f7-dc4c1151948d" (UID: "aaf39141-4f8a-4e38-b9f7-dc4c1151948d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.540895 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.540981 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.541046 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.541967 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.542072 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" gracePeriod=600 Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.560718 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n9g5\" (UniqueName: \"kubernetes.io/projected/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-kube-api-access-6n9g5\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.560752 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.560766 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.560778 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.560790 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.585896 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "aaf39141-4f8a-4e38-b9f7-dc4c1151948d" (UID: "aaf39141-4f8a-4e38-b9f7-dc4c1151948d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.600558 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aaf39141-4f8a-4e38-b9f7-dc4c1151948d" (UID: "aaf39141-4f8a-4e38-b9f7-dc4c1151948d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.662920 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:46 crc kubenswrapper[4978]: I0225 09:40:46.662962 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/aaf39141-4f8a-4e38-b9f7-dc4c1151948d-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:40:46 crc kubenswrapper[4978]: E0225 09:40:46.679672 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.125244 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-gtz5n" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.127017 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce2bbd050a0127c046c33bdcf248a3656e2281825d1dabb6d54217a48b7c0319" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.131332 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" exitCode=0 Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.131398 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136"} Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.131487 4978 scope.go:117] "RemoveContainer" containerID="859a4fb26d2b478e0589aefabf4ce096e14e697af774d19395e2b4f3b29e5be9" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.132793 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:40:47 crc kubenswrapper[4978]: E0225 09:40:47.133448 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.340662 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaf39141-4f8a-4e38-b9f7-dc4c1151948d" path="/var/lib/kubelet/pods/aaf39141-4f8a-4e38-b9f7-dc4c1151948d/volumes" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.587593 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-4s92w"] Feb 25 09:40:47 crc kubenswrapper[4978]: E0225 09:40:47.588203 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaf39141-4f8a-4e38-b9f7-dc4c1151948d" containerName="swift-ring-rebalance" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.588232 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaf39141-4f8a-4e38-b9f7-dc4c1151948d" containerName="swift-ring-rebalance" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.588619 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaf39141-4f8a-4e38-b9f7-dc4c1151948d" containerName="swift-ring-rebalance" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.589727 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.592324 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.593280 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.610340 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-4s92w"] Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.684090 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-scripts\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.684158 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/650cc49e-1aea-40fa-9b50-8c062378e9ef-etc-swift\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.684203 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.684239 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-swiftconf\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.684478 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-dispersionconf\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.684518 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-ring-data-devices\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.684587 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr4qc\" (UniqueName: \"kubernetes.io/projected/650cc49e-1aea-40fa-9b50-8c062378e9ef-kube-api-access-qr4qc\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.786895 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-dispersionconf\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.787182 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-ring-data-devices\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.787341 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr4qc\" (UniqueName: \"kubernetes.io/projected/650cc49e-1aea-40fa-9b50-8c062378e9ef-kube-api-access-qr4qc\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.787524 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-scripts\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.787669 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/650cc49e-1aea-40fa-9b50-8c062378e9ef-etc-swift\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.788193 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/650cc49e-1aea-40fa-9b50-8c062378e9ef-etc-swift\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.788360 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.788960 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-swiftconf\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.788602 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-ring-data-devices\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:47 crc kubenswrapper[4978]: I0225 09:40:47.789498 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-scripts\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:48 crc kubenswrapper[4978]: I0225 09:40:48.287063 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-dispersionconf\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:48 crc kubenswrapper[4978]: I0225 09:40:48.287090 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr4qc\" (UniqueName: \"kubernetes.io/projected/650cc49e-1aea-40fa-9b50-8c062378e9ef-kube-api-access-qr4qc\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:48 crc kubenswrapper[4978]: I0225 09:40:48.287538 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-swiftconf\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:48 crc kubenswrapper[4978]: I0225 09:40:48.294135 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-4s92w\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:48 crc kubenswrapper[4978]: I0225 09:40:48.528282 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:40:49 crc kubenswrapper[4978]: I0225 09:40:49.047836 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-4s92w"] Feb 25 09:40:49 crc kubenswrapper[4978]: I0225 09:40:49.153227 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-4s92w" event={"ID":"650cc49e-1aea-40fa-9b50-8c062378e9ef","Type":"ContainerStarted","Data":"4048ce690331e094b5eefe799acba75e572a54d44b74cd5191f5630b882105dc"} Feb 25 09:40:50 crc kubenswrapper[4978]: I0225 09:40:50.174056 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-4s92w" event={"ID":"650cc49e-1aea-40fa-9b50-8c062378e9ef","Type":"ContainerStarted","Data":"6fddf5c93eac70ec068e60efec3427719a776324a47e721526991fdac65b9c50"} Feb 25 09:40:50 crc kubenswrapper[4978]: I0225 09:40:50.213432 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-4s92w" podStartSLOduration=3.213412091 podStartE2EDuration="3.213412091s" podCreationTimestamp="2026-02-25 09:40:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:40:50.207510217 +0000 UTC m=+10543.646766676" watchObservedRunningTime="2026-02-25 09:40:50.213412091 +0000 UTC m=+10543.652668560" Feb 25 09:41:02 crc kubenswrapper[4978]: I0225 09:41:02.328538 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:41:02 crc kubenswrapper[4978]: E0225 09:41:02.329413 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.413470 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-knz87"] Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.419300 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.427027 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-knz87"] Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.485803 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-catalog-content\") pod \"redhat-marketplace-knz87\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.487197 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-utilities\") pod \"redhat-marketplace-knz87\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.487471 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skbsn\" (UniqueName: \"kubernetes.io/projected/412631e7-0c6c-4552-9ed2-4d37a36443af-kube-api-access-skbsn\") pod \"redhat-marketplace-knz87\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.589411 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-utilities\") pod \"redhat-marketplace-knz87\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.589572 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skbsn\" (UniqueName: \"kubernetes.io/projected/412631e7-0c6c-4552-9ed2-4d37a36443af-kube-api-access-skbsn\") pod \"redhat-marketplace-knz87\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.589638 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-catalog-content\") pod \"redhat-marketplace-knz87\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.589975 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-utilities\") pod \"redhat-marketplace-knz87\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.590164 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-catalog-content\") pod \"redhat-marketplace-knz87\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.611289 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skbsn\" (UniqueName: \"kubernetes.io/projected/412631e7-0c6c-4552-9ed2-4d37a36443af-kube-api-access-skbsn\") pod \"redhat-marketplace-knz87\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:05 crc kubenswrapper[4978]: I0225 09:41:05.750117 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:06 crc kubenswrapper[4978]: W0225 09:41:06.243439 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod412631e7_0c6c_4552_9ed2_4d37a36443af.slice/crio-7dafe22421f5eae397b0b049cf7589d576b708893439d715d484824526cc5472 WatchSource:0}: Error finding container 7dafe22421f5eae397b0b049cf7589d576b708893439d715d484824526cc5472: Status 404 returned error can't find the container with id 7dafe22421f5eae397b0b049cf7589d576b708893439d715d484824526cc5472 Feb 25 09:41:06 crc kubenswrapper[4978]: I0225 09:41:06.255878 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-knz87"] Feb 25 09:41:06 crc kubenswrapper[4978]: I0225 09:41:06.331065 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knz87" event={"ID":"412631e7-0c6c-4552-9ed2-4d37a36443af","Type":"ContainerStarted","Data":"7dafe22421f5eae397b0b049cf7589d576b708893439d715d484824526cc5472"} Feb 25 09:41:07 crc kubenswrapper[4978]: I0225 09:41:07.385730 4978 generic.go:334] "Generic (PLEG): container finished" podID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerID="9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2" exitCode=0 Feb 25 09:41:07 crc kubenswrapper[4978]: I0225 09:41:07.424065 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knz87" event={"ID":"412631e7-0c6c-4552-9ed2-4d37a36443af","Type":"ContainerDied","Data":"9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2"} Feb 25 09:41:09 crc kubenswrapper[4978]: I0225 09:41:09.412453 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knz87" event={"ID":"412631e7-0c6c-4552-9ed2-4d37a36443af","Type":"ContainerStarted","Data":"4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b"} Feb 25 09:41:10 crc kubenswrapper[4978]: I0225 09:41:10.424207 4978 generic.go:334] "Generic (PLEG): container finished" podID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerID="4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b" exitCode=0 Feb 25 09:41:10 crc kubenswrapper[4978]: I0225 09:41:10.424317 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knz87" event={"ID":"412631e7-0c6c-4552-9ed2-4d37a36443af","Type":"ContainerDied","Data":"4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b"} Feb 25 09:41:11 crc kubenswrapper[4978]: I0225 09:41:11.441569 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knz87" event={"ID":"412631e7-0c6c-4552-9ed2-4d37a36443af","Type":"ContainerStarted","Data":"7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f"} Feb 25 09:41:11 crc kubenswrapper[4978]: I0225 09:41:11.507533 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-knz87" podStartSLOduration=2.998604252 podStartE2EDuration="6.507508514s" podCreationTimestamp="2026-02-25 09:41:05 +0000 UTC" firstStartedPulling="2026-02-25 09:41:07.409078356 +0000 UTC m=+10560.848334815" lastFinishedPulling="2026-02-25 09:41:10.917982618 +0000 UTC m=+10564.357239077" observedRunningTime="2026-02-25 09:41:11.481089521 +0000 UTC m=+10564.920346030" watchObservedRunningTime="2026-02-25 09:41:11.507508514 +0000 UTC m=+10564.946764973" Feb 25 09:41:14 crc kubenswrapper[4978]: I0225 09:41:14.330612 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:41:14 crc kubenswrapper[4978]: E0225 09:41:14.331784 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:41:15 crc kubenswrapper[4978]: I0225 09:41:15.750280 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:15 crc kubenswrapper[4978]: I0225 09:41:15.751350 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:15 crc kubenswrapper[4978]: I0225 09:41:15.833312 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:16 crc kubenswrapper[4978]: I0225 09:41:16.501013 4978 generic.go:334] "Generic (PLEG): container finished" podID="650cc49e-1aea-40fa-9b50-8c062378e9ef" containerID="6fddf5c93eac70ec068e60efec3427719a776324a47e721526991fdac65b9c50" exitCode=0 Feb 25 09:41:16 crc kubenswrapper[4978]: I0225 09:41:16.501147 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-4s92w" event={"ID":"650cc49e-1aea-40fa-9b50-8c062378e9ef","Type":"ContainerDied","Data":"6fddf5c93eac70ec068e60efec3427719a776324a47e721526991fdac65b9c50"} Feb 25 09:41:16 crc kubenswrapper[4978]: I0225 09:41:16.607570 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:16 crc kubenswrapper[4978]: I0225 09:41:16.696463 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-knz87"] Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.129041 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.171235 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-4s92w"] Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.194734 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-4s92w"] Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.288811 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-dispersionconf\") pod \"650cc49e-1aea-40fa-9b50-8c062378e9ef\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.288925 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-combined-ca-bundle\") pod \"650cc49e-1aea-40fa-9b50-8c062378e9ef\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.289016 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-ring-data-devices\") pod \"650cc49e-1aea-40fa-9b50-8c062378e9ef\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.289046 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr4qc\" (UniqueName: \"kubernetes.io/projected/650cc49e-1aea-40fa-9b50-8c062378e9ef-kube-api-access-qr4qc\") pod \"650cc49e-1aea-40fa-9b50-8c062378e9ef\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.289147 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/650cc49e-1aea-40fa-9b50-8c062378e9ef-etc-swift\") pod \"650cc49e-1aea-40fa-9b50-8c062378e9ef\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.289176 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-scripts\") pod \"650cc49e-1aea-40fa-9b50-8c062378e9ef\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.289282 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-swiftconf\") pod \"650cc49e-1aea-40fa-9b50-8c062378e9ef\" (UID: \"650cc49e-1aea-40fa-9b50-8c062378e9ef\") " Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.290107 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/650cc49e-1aea-40fa-9b50-8c062378e9ef-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "650cc49e-1aea-40fa-9b50-8c062378e9ef" (UID: "650cc49e-1aea-40fa-9b50-8c062378e9ef"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.290134 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "650cc49e-1aea-40fa-9b50-8c062378e9ef" (UID: "650cc49e-1aea-40fa-9b50-8c062378e9ef"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.297708 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/650cc49e-1aea-40fa-9b50-8c062378e9ef-kube-api-access-qr4qc" (OuterVolumeSpecName: "kube-api-access-qr4qc") pod "650cc49e-1aea-40fa-9b50-8c062378e9ef" (UID: "650cc49e-1aea-40fa-9b50-8c062378e9ef"). InnerVolumeSpecName "kube-api-access-qr4qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.328512 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "650cc49e-1aea-40fa-9b50-8c062378e9ef" (UID: "650cc49e-1aea-40fa-9b50-8c062378e9ef"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.330543 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-scripts" (OuterVolumeSpecName: "scripts") pod "650cc49e-1aea-40fa-9b50-8c062378e9ef" (UID: "650cc49e-1aea-40fa-9b50-8c062378e9ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.332092 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "650cc49e-1aea-40fa-9b50-8c062378e9ef" (UID: "650cc49e-1aea-40fa-9b50-8c062378e9ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.339949 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "650cc49e-1aea-40fa-9b50-8c062378e9ef" (UID: "650cc49e-1aea-40fa-9b50-8c062378e9ef"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.392224 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.392496 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.392582 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.392666 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr4qc\" (UniqueName: \"kubernetes.io/projected/650cc49e-1aea-40fa-9b50-8c062378e9ef-kube-api-access-qr4qc\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.392751 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/650cc49e-1aea-40fa-9b50-8c062378e9ef-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.392824 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/650cc49e-1aea-40fa-9b50-8c062378e9ef-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.392895 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/650cc49e-1aea-40fa-9b50-8c062378e9ef-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.540765 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4048ce690331e094b5eefe799acba75e572a54d44b74cd5191f5630b882105dc" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.540797 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-4s92w" Feb 25 09:41:18 crc kubenswrapper[4978]: I0225 09:41:18.540763 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-knz87" podUID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerName="registry-server" containerID="cri-o://7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f" gracePeriod=2 Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.156422 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.311938 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-utilities\") pod \"412631e7-0c6c-4552-9ed2-4d37a36443af\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.312085 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skbsn\" (UniqueName: \"kubernetes.io/projected/412631e7-0c6c-4552-9ed2-4d37a36443af-kube-api-access-skbsn\") pod \"412631e7-0c6c-4552-9ed2-4d37a36443af\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.312140 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-catalog-content\") pod \"412631e7-0c6c-4552-9ed2-4d37a36443af\" (UID: \"412631e7-0c6c-4552-9ed2-4d37a36443af\") " Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.312867 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-utilities" (OuterVolumeSpecName: "utilities") pod "412631e7-0c6c-4552-9ed2-4d37a36443af" (UID: "412631e7-0c6c-4552-9ed2-4d37a36443af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.319040 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/412631e7-0c6c-4552-9ed2-4d37a36443af-kube-api-access-skbsn" (OuterVolumeSpecName: "kube-api-access-skbsn") pod "412631e7-0c6c-4552-9ed2-4d37a36443af" (UID: "412631e7-0c6c-4552-9ed2-4d37a36443af"). InnerVolumeSpecName "kube-api-access-skbsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.342196 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="650cc49e-1aea-40fa-9b50-8c062378e9ef" path="/var/lib/kubelet/pods/650cc49e-1aea-40fa-9b50-8c062378e9ef/volumes" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.345939 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "412631e7-0c6c-4552-9ed2-4d37a36443af" (UID: "412631e7-0c6c-4552-9ed2-4d37a36443af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.415195 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.415229 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skbsn\" (UniqueName: \"kubernetes.io/projected/412631e7-0c6c-4552-9ed2-4d37a36443af-kube-api-access-skbsn\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.415239 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412631e7-0c6c-4552-9ed2-4d37a36443af-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.551128 4978 generic.go:334] "Generic (PLEG): container finished" podID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerID="7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f" exitCode=0 Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.551248 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knz87" event={"ID":"412631e7-0c6c-4552-9ed2-4d37a36443af","Type":"ContainerDied","Data":"7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f"} Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.551282 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knz87" event={"ID":"412631e7-0c6c-4552-9ed2-4d37a36443af","Type":"ContainerDied","Data":"7dafe22421f5eae397b0b049cf7589d576b708893439d715d484824526cc5472"} Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.551300 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-knz87" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.551325 4978 scope.go:117] "RemoveContainer" containerID="7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.578469 4978 scope.go:117] "RemoveContainer" containerID="4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.583827 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-knz87"] Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.595206 4978 scope.go:117] "RemoveContainer" containerID="9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.599256 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-knz87"] Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.635952 4978 scope.go:117] "RemoveContainer" containerID="7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f" Feb 25 09:41:19 crc kubenswrapper[4978]: E0225 09:41:19.638262 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f\": container with ID starting with 7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f not found: ID does not exist" containerID="7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.638311 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f"} err="failed to get container status \"7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f\": rpc error: code = NotFound desc = could not find container \"7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f\": container with ID starting with 7f1b326107283ff9bd941ac87e323c478cdf8be9ef21afabeabfc37c18c8922f not found: ID does not exist" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.638339 4978 scope.go:117] "RemoveContainer" containerID="4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b" Feb 25 09:41:19 crc kubenswrapper[4978]: E0225 09:41:19.638673 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b\": container with ID starting with 4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b not found: ID does not exist" containerID="4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.638708 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b"} err="failed to get container status \"4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b\": rpc error: code = NotFound desc = could not find container \"4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b\": container with ID starting with 4ee4f4688bd519e538be5089ddd6e1535eca0a5fa844598781c574ccae64207b not found: ID does not exist" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.638732 4978 scope.go:117] "RemoveContainer" containerID="9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2" Feb 25 09:41:19 crc kubenswrapper[4978]: E0225 09:41:19.639026 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2\": container with ID starting with 9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2 not found: ID does not exist" containerID="9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2" Feb 25 09:41:19 crc kubenswrapper[4978]: I0225 09:41:19.639049 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2"} err="failed to get container status \"9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2\": rpc error: code = NotFound desc = could not find container \"9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2\": container with ID starting with 9f01c7ce7b91751e7c931af51c8faab97cf1e8955092b1354ab72886ebd795b2 not found: ID does not exist" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.459964 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 25 09:41:20 crc kubenswrapper[4978]: E0225 09:41:20.461012 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerName="registry-server" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.461039 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerName="registry-server" Feb 25 09:41:20 crc kubenswrapper[4978]: E0225 09:41:20.461067 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="650cc49e-1aea-40fa-9b50-8c062378e9ef" containerName="swift-ring-rebalance" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.461078 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="650cc49e-1aea-40fa-9b50-8c062378e9ef" containerName="swift-ring-rebalance" Feb 25 09:41:20 crc kubenswrapper[4978]: E0225 09:41:20.461105 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerName="extract-content" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.461119 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerName="extract-content" Feb 25 09:41:20 crc kubenswrapper[4978]: E0225 09:41:20.461164 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerName="extract-utilities" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.461175 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerName="extract-utilities" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.461552 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="412631e7-0c6c-4552-9ed2-4d37a36443af" containerName="registry-server" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.461599 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="650cc49e-1aea-40fa-9b50-8c062378e9ef" containerName="swift-ring-rebalance" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.478690 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.481317 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-2"] Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.482036 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.499743 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.502461 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.517859 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-1"] Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.526018 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.539246 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-2"] Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.570645 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-1"] Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639356 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-cache\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639472 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pwdj\" (UniqueName: \"kubernetes.io/projected/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-kube-api-access-2pwdj\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639509 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-lock\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639567 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fb9553c0-7c54-485e-b7a3-e9e1d769977d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb9553c0-7c54-485e-b7a3-e9e1d769977d\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639633 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-cache\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639838 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-etc-swift\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639868 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-etc-swift\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639894 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-cache\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639914 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-132a7f0d-fc8e-4eda-a255-95fedf4b6ee9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-132a7f0d-fc8e-4eda-a255-95fedf4b6ee9\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639939 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-etc-swift\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639956 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cghlb\" (UniqueName: \"kubernetes.io/projected/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-kube-api-access-cghlb\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.639991 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-lock\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.640006 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ccfdc5c7-12f9-418a-b1d9-adf56ab91838\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ccfdc5c7-12f9-418a-b1d9-adf56ab91838\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.640038 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqzk2\" (UniqueName: \"kubernetes.io/projected/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-kube-api-access-pqzk2\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.640068 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-combined-ca-bundle\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.640088 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-combined-ca-bundle\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.640112 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-lock\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.640133 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.741738 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-combined-ca-bundle\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.742041 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-combined-ca-bundle\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.742137 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-lock\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.742216 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.742303 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-cache\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.742394 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pwdj\" (UniqueName: \"kubernetes.io/projected/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-kube-api-access-2pwdj\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.742512 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-lock\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.742636 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fb9553c0-7c54-485e-b7a3-e9e1d769977d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb9553c0-7c54-485e-b7a3-e9e1d769977d\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.742777 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-cache\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.742851 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-cache\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.742932 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-etc-swift\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.743020 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-etc-swift\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.743130 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-cache\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.743217 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-132a7f0d-fc8e-4eda-a255-95fedf4b6ee9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-132a7f0d-fc8e-4eda-a255-95fedf4b6ee9\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.743329 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-etc-swift\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.743458 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cghlb\" (UniqueName: \"kubernetes.io/projected/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-kube-api-access-cghlb\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.743601 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-lock\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.743707 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ccfdc5c7-12f9-418a-b1d9-adf56ab91838\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ccfdc5c7-12f9-418a-b1d9-adf56ab91838\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.743845 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqzk2\" (UniqueName: \"kubernetes.io/projected/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-kube-api-access-pqzk2\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.744235 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-cache\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.744796 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-lock\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.743080 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-lock\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.745161 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-lock\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.745357 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-cache\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.746442 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-combined-ca-bundle\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.748334 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.753862 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-etc-swift\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.755387 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-etc-swift\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.763745 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-etc-swift\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.764204 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cghlb\" (UniqueName: \"kubernetes.io/projected/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-kube-api-access-cghlb\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.772114 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqzk2\" (UniqueName: \"kubernetes.io/projected/2e3f7363-8d31-4036-96bc-25f2d2c1b4bb-kube-api-access-pqzk2\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.772913 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9958cdd-7b41-48e4-b4fa-89b23d6e006c-combined-ca-bundle\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.840713 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pwdj\" (UniqueName: \"kubernetes.io/projected/f6550fa0-18f1-4c7a-a8e7-ab3960f5828b-kube-api-access-2pwdj\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.914924 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.914967 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fb9553c0-7c54-485e-b7a3-e9e1d769977d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb9553c0-7c54-485e-b7a3-e9e1d769977d\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6464d8409fa38841542ca2210b9ec0893f4863e550fbac5ee31252e31060fd45/globalmount\"" pod="openstack/swift-storage-2" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.914980 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.915012 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-132a7f0d-fc8e-4eda-a255-95fedf4b6ee9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-132a7f0d-fc8e-4eda-a255-95fedf4b6ee9\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bf212d6f4be8e91fc5dd864ca049793349306e6a70e726236d05e9f97659268e/globalmount\"" pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.914928 4978 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.915068 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ccfdc5c7-12f9-418a-b1d9-adf56ab91838\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ccfdc5c7-12f9-418a-b1d9-adf56ab91838\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/23193b0d757e12245bd9b9d9e78b6b1bdb5d55a0e13a003519b38cdf2bf7c528/globalmount\"" pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.976203 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-132a7f0d-fc8e-4eda-a255-95fedf4b6ee9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-132a7f0d-fc8e-4eda-a255-95fedf4b6ee9\") pod \"swift-storage-1\" (UID: \"c9958cdd-7b41-48e4-b4fa-89b23d6e006c\") " pod="openstack/swift-storage-1" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.984696 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ccfdc5c7-12f9-418a-b1d9-adf56ab91838\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ccfdc5c7-12f9-418a-b1d9-adf56ab91838\") pod \"swift-storage-0\" (UID: \"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b\") " pod="openstack/swift-storage-0" Feb 25 09:41:20 crc kubenswrapper[4978]: I0225 09:41:20.991258 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fb9553c0-7c54-485e-b7a3-e9e1d769977d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fb9553c0-7c54-485e-b7a3-e9e1d769977d\") pod \"swift-storage-2\" (UID: \"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb\") " pod="openstack/swift-storage-2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.126882 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.136447 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.157547 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-1" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.352843 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="412631e7-0c6c-4552-9ed2-4d37a36443af" path="/var/lib/kubelet/pods/412631e7-0c6c-4552-9ed2-4d37a36443af/volumes" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.554326 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9fw44"] Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.564591 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-9fw44"] Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.604657 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-tlbw2"] Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.607127 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.609800 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.610258 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.628551 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tlbw2"] Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.667193 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-ring-data-devices\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.667246 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-scripts\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.667307 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf7ng\" (UniqueName: \"kubernetes.io/projected/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-kube-api-access-wf7ng\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.667354 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-combined-ca-bundle\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.667435 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-dispersionconf\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.667488 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-swiftconf\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.667520 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-etc-swift\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.769464 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-dispersionconf\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.769543 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-swiftconf\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.769594 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-etc-swift\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.769695 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-ring-data-devices\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.769740 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-scripts\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.769820 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf7ng\" (UniqueName: \"kubernetes.io/projected/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-kube-api-access-wf7ng\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.769888 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-combined-ca-bundle\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.770159 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-etc-swift\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.770970 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-scripts\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.770995 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-ring-data-devices\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.891174 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-2"] Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.989129 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-dispersionconf\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.989281 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-combined-ca-bundle\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.989587 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-swiftconf\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:21 crc kubenswrapper[4978]: I0225 09:41:21.989618 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf7ng\" (UniqueName: \"kubernetes.io/projected/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-kube-api-access-wf7ng\") pod \"swift-ring-rebalance-tlbw2\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:22 crc kubenswrapper[4978]: I0225 09:41:22.235293 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:22 crc kubenswrapper[4978]: I0225 09:41:22.528065 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-1"] Feb 25 09:41:22 crc kubenswrapper[4978]: I0225 09:41:22.611684 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"d38c6cefae9eddb0b8f84e456835ae1f10b3341f575d0f40427e0461c4a951a2"} Feb 25 09:41:22 crc kubenswrapper[4978]: I0225 09:41:22.615896 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"de818fe86cadbb9d297eb049340fcf48e0985d217ed917443e4be69198986b46"} Feb 25 09:41:22 crc kubenswrapper[4978]: I0225 09:41:22.739624 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 25 09:41:22 crc kubenswrapper[4978]: W0225 09:41:22.747726 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6550fa0_18f1_4c7a_a8e7_ab3960f5828b.slice/crio-c2150ac23c0cdf223323463f41df730f1c3f5aad252db9491b0916659fe0bf3c WatchSource:0}: Error finding container c2150ac23c0cdf223323463f41df730f1c3f5aad252db9491b0916659fe0bf3c: Status 404 returned error can't find the container with id c2150ac23c0cdf223323463f41df730f1c3f5aad252db9491b0916659fe0bf3c Feb 25 09:41:22 crc kubenswrapper[4978]: I0225 09:41:22.787948 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tlbw2"] Feb 25 09:41:22 crc kubenswrapper[4978]: W0225 09:41:22.795625 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod751b1bbb_a2d3_4d1b_bc2e_60bb99fbd859.slice/crio-04d33f7ce6b50ddc86e44c9f6c99264b98c0fc534aad659e475f9d6abcb1809f WatchSource:0}: Error finding container 04d33f7ce6b50ddc86e44c9f6c99264b98c0fc534aad659e475f9d6abcb1809f: Status 404 returned error can't find the container with id 04d33f7ce6b50ddc86e44c9f6c99264b98c0fc534aad659e475f9d6abcb1809f Feb 25 09:41:23 crc kubenswrapper[4978]: I0225 09:41:23.347195 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0976f41-a781-43a1-be78-efbb126c9910" path="/var/lib/kubelet/pods/f0976f41-a781-43a1-be78-efbb126c9910/volumes" Feb 25 09:41:23 crc kubenswrapper[4978]: I0225 09:41:23.629897 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tlbw2" event={"ID":"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859","Type":"ContainerStarted","Data":"04d33f7ce6b50ddc86e44c9f6c99264b98c0fc534aad659e475f9d6abcb1809f"} Feb 25 09:41:23 crc kubenswrapper[4978]: I0225 09:41:23.631762 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"c2150ac23c0cdf223323463f41df730f1c3f5aad252db9491b0916659fe0bf3c"} Feb 25 09:41:24 crc kubenswrapper[4978]: I0225 09:41:24.641769 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tlbw2" event={"ID":"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859","Type":"ContainerStarted","Data":"173b35239f6824f87f4feac0676964b501baca63b758633a0e27e4fef1a42540"} Feb 25 09:41:24 crc kubenswrapper[4978]: I0225 09:41:24.665556 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"cefd345055849699eefb8c1751e28a6d91276fe0e2314f9a70af5dc4c3655edb"} Feb 25 09:41:24 crc kubenswrapper[4978]: I0225 09:41:24.665599 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"57d68d8c81cf165cd1335a5eb227ce1d8cd8f5145bc23542a545d637e41dee98"} Feb 25 09:41:24 crc kubenswrapper[4978]: I0225 09:41:24.667739 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"51c7238b9cf63730a686aea725370f8ed86368b5cf433eabe23fa64daba472e5"} Feb 25 09:41:24 crc kubenswrapper[4978]: I0225 09:41:24.667769 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"ef089c9b95edb82f704bb6a9dff5dd31df9db5cea028853d172675adab9e0571"} Feb 25 09:41:24 crc kubenswrapper[4978]: I0225 09:41:24.674957 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-tlbw2" podStartSLOduration=3.674940966 podStartE2EDuration="3.674940966s" podCreationTimestamp="2026-02-25 09:41:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:41:24.665323807 +0000 UTC m=+10578.104580266" watchObservedRunningTime="2026-02-25 09:41:24.674940966 +0000 UTC m=+10578.114197425" Feb 25 09:41:24 crc kubenswrapper[4978]: I0225 09:41:24.676452 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"68ca114068809d532d573a3977519159f902315da64ef07a514aa9c65aa4ad6e"} Feb 25 09:41:24 crc kubenswrapper[4978]: I0225 09:41:24.676505 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"e00a7bd4a68a1fefab1066243a8581f69b77873ce1837e7209216a0f7efc077d"} Feb 25 09:41:25 crc kubenswrapper[4978]: I0225 09:41:25.699942 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"9284c4a810ec0f141336e3cb19abab7d8b648fb066bab004d3dd3cbd95fbadc3"} Feb 25 09:41:25 crc kubenswrapper[4978]: I0225 09:41:25.700213 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"44f45313ffc9fe2a0d91228d9752b1b123828c3d37d1cafaa10f6c18e0877505"} Feb 25 09:41:25 crc kubenswrapper[4978]: I0225 09:41:25.709389 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"963e074f4c25460b01d9c5720b349f2c77f9e954e3de77313136ea3e9a2e8b02"} Feb 25 09:41:25 crc kubenswrapper[4978]: I0225 09:41:25.709432 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"a51fb10afb99fcf9afe20db07e668908a5c79040038a186fd7222471fef2d436"} Feb 25 09:41:25 crc kubenswrapper[4978]: I0225 09:41:25.712311 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"b3582a571b9df9aab22ea6573941852b72d80bd43bd50904b35a82b295b7937d"} Feb 25 09:41:25 crc kubenswrapper[4978]: I0225 09:41:25.712355 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"cefaf97e0197d692e71c2771a390332d654ec9cfbd3010e94efced632662609b"} Feb 25 09:41:26 crc kubenswrapper[4978]: I0225 09:41:26.762023 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"37acc5838cede9fbccc4288501495f6cadd80eac2404e47a9262cfed0ef8f8c3"} Feb 25 09:41:26 crc kubenswrapper[4978]: I0225 09:41:26.777089 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"984b25bb4638299ebd5a35f4bc087a635980462f80ec3e7a70962aa7f89efb06"} Feb 25 09:41:26 crc kubenswrapper[4978]: I0225 09:41:26.780942 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"cf7a77fcc81afcd6a02783212a553ca63fe2b652b23250947cfbd1c357848c22"} Feb 25 09:41:27 crc kubenswrapper[4978]: I0225 09:41:27.805686 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"76d1ab62f81d67b70cc6b0670a3fb543fb8aab1d5cd5134bd76a0056727c19aa"} Feb 25 09:41:27 crc kubenswrapper[4978]: I0225 09:41:27.806247 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"96b836173fc29d8fd97f3c9a1c8bfdd554e9031b9fe4e14a9781da4fabd5c6e4"} Feb 25 09:41:27 crc kubenswrapper[4978]: I0225 09:41:27.806267 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"6b6e392d004fd220f20cc698ebb867a71a56224fc27a3025a86986b9b8919837"} Feb 25 09:41:27 crc kubenswrapper[4978]: I0225 09:41:27.814841 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"388a2b7d8aec301896d2b9430cdddd21b7715d5021c2d8c82b1125d96f2878cb"} Feb 25 09:41:27 crc kubenswrapper[4978]: I0225 09:41:27.814884 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"b38ae2274b4d5555fd8d20227d26eef1e26fc7ba81d0dabe8f4080142f4be0d3"} Feb 25 09:41:27 crc kubenswrapper[4978]: I0225 09:41:27.814897 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"01e841534d40331441145a1e4ff9d63cf03c688cfc021544e094d30ddc603738"} Feb 25 09:41:27 crc kubenswrapper[4978]: I0225 09:41:27.818985 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"657594a8da23dd1ca2d4f7fe8146f29840009903b1fa589949f8b9c8e2663d20"} Feb 25 09:41:27 crc kubenswrapper[4978]: I0225 09:41:27.819034 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"35ba5232347a62ab81d156de43747c93664cfaeca005fe0d212fb9bbc503f977"} Feb 25 09:41:27 crc kubenswrapper[4978]: I0225 09:41:27.819043 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"57ce5105b2106eae3957033d76741362137a2aad06c3d41f33e2c92f4cac22e6"} Feb 25 09:41:28 crc kubenswrapper[4978]: I0225 09:41:28.327183 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:41:28 crc kubenswrapper[4978]: E0225 09:41:28.327566 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:41:29 crc kubenswrapper[4978]: I0225 09:41:29.893716 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"a6fd8940af011e44c7bf2695f94b8e601161859d072d9757e047c615199fc4c5"} Feb 25 09:41:29 crc kubenswrapper[4978]: I0225 09:41:29.894299 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"810bf300e92d49c6a9878a0313bd04c25434abfa8d853f294e203b0982dc83b4"} Feb 25 09:41:29 crc kubenswrapper[4978]: I0225 09:41:29.894316 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"4cc856580ac4f0051d67f040fcf80f607d506b29a36597785b9fdc0881d91af6"} Feb 25 09:41:29 crc kubenswrapper[4978]: I0225 09:41:29.934690 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"affc0a21d6a3fc991d57f8730acb30ed520fd118da7b7146f5a21b8fbd1e41ea"} Feb 25 09:41:29 crc kubenswrapper[4978]: I0225 09:41:29.934734 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"49cdbcd3f8dcd44b5c1d0bd7feb154270b9322a15ebb07e284642854f359806c"} Feb 25 09:41:29 crc kubenswrapper[4978]: I0225 09:41:29.934744 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"2f568783e6512796e78400558d4fde62dfd5a3e777dc315b5f9ca04d0190fb3a"} Feb 25 09:41:29 crc kubenswrapper[4978]: I0225 09:41:29.944982 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"0f06d1174c1c868bf849983fd3133a2f4fa37094dd598b8c0d3072b2b457967b"} Feb 25 09:41:29 crc kubenswrapper[4978]: I0225 09:41:29.945014 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"03c6f5166652935a067659428163d9677fe903c60ff43b527965879b0b754e8a"} Feb 25 09:41:29 crc kubenswrapper[4978]: I0225 09:41:29.945042 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"ee09a36a9ff54cd0b34bd2eedb6a4ca67fd6c582c81fbc2e9a377d8ce5aa2e4a"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.959759 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"b3c917e915878f69172a183af0c279e822fc370ccbbbfae5f185a7944d59fc1d"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.960735 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"5a3e90ba69c2c52a2f6d5919346eb83008323819055d0102368768a6c924ffb8"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.960832 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"163cc862bf7d458901359fea43296b87a1f80cfde91ac50f1747ab8807754069"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.960919 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f6550fa0-18f1-4c7a-a8e7-ab3960f5828b","Type":"ContainerStarted","Data":"50187458c7350414ae2b39521b26f00083767e2a004cde86e2475e8a8161c89b"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.969713 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"95670402712ed380f157432251235a896f8c8949eea1211b753ee67f98e3571e"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.969757 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"24e8f7d6fdba4212df00bd3a0ee2674f92ca61204b55b635021fa9971f816599"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.969769 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"99c89f2667ddf1ddebd04db88fed2347aabf4afdda57929cd87a7d6bdfb7d994"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.969778 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-1" event={"ID":"c9958cdd-7b41-48e4-b4fa-89b23d6e006c","Type":"ContainerStarted","Data":"f999fc9a933e2d602910a6bbd1014fc7cc0474daf847016ea65ed7cf3c4d7664"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.982146 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"24193fe16e91983310c2b7d5f20308440aa03cc375480eda8adc044abc23d6c9"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.982188 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"b2051eb4815219ba7bc3be7694df7ab63b531761755e1cc7688bfc2a814e269e"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.982199 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"a3eb8bbff0eaf3b761dd65bb0aa03daa1070c49f972204c209014e990cf79f12"} Feb 25 09:41:30 crc kubenswrapper[4978]: I0225 09:41:30.982211 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-2" event={"ID":"2e3f7363-8d31-4036-96bc-25f2d2c1b4bb","Type":"ContainerStarted","Data":"0c0ab97f00e9f0f0f635e6677f36df40e1787028f943d7f195b60009dbd4f3fb"} Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.000605 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=5.993775748 podStartE2EDuration="12.000585501s" podCreationTimestamp="2026-02-25 09:41:19 +0000 UTC" firstStartedPulling="2026-02-25 09:41:22.753482603 +0000 UTC m=+10576.192739062" lastFinishedPulling="2026-02-25 09:41:28.760292346 +0000 UTC m=+10582.199548815" observedRunningTime="2026-02-25 09:41:30.998807376 +0000 UTC m=+10584.438063865" watchObservedRunningTime="2026-02-25 09:41:31.000585501 +0000 UTC m=+10584.439841960" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.052401 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-2" podStartSLOduration=5.398320178 podStartE2EDuration="12.052386783s" podCreationTimestamp="2026-02-25 09:41:19 +0000 UTC" firstStartedPulling="2026-02-25 09:41:22.106124938 +0000 UTC m=+10575.545381397" lastFinishedPulling="2026-02-25 09:41:28.760191543 +0000 UTC m=+10582.199448002" observedRunningTime="2026-02-25 09:41:31.035122007 +0000 UTC m=+10584.474378486" watchObservedRunningTime="2026-02-25 09:41:31.052386783 +0000 UTC m=+10584.491643242" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.093480 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-1" podStartSLOduration=5.892888129 podStartE2EDuration="12.093424931s" podCreationTimestamp="2026-02-25 09:41:19 +0000 UTC" firstStartedPulling="2026-02-25 09:41:22.561896481 +0000 UTC m=+10576.001152940" lastFinishedPulling="2026-02-25 09:41:28.762433273 +0000 UTC m=+10582.201689742" observedRunningTime="2026-02-25 09:41:31.075897445 +0000 UTC m=+10584.515153904" watchObservedRunningTime="2026-02-25 09:41:31.093424931 +0000 UTC m=+10584.532681390" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.468308 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b8dcfd55-xnv42"] Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.470122 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.471869 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.508194 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b8dcfd55-xnv42"] Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.539753 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-config\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.539829 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-svc\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.539853 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.539899 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-swift-storage-0\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.539937 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-networker\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.539962 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.540027 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-cell1\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.540152 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h29rv\" (UniqueName: \"kubernetes.io/projected/05958022-1fd8-4cb3-885a-4d49b69034cb-kube-api-access-h29rv\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.553562 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b8dcfd55-xnv42"] Feb 25 09:41:31 crc kubenswrapper[4978]: E0225 09:41:31.554415 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-h29rv openstack-cell1 openstack-networker ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" podUID="05958022-1fd8-4cb3-885a-4d49b69034cb" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.569114 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6997548745-tb5n6"] Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.570987 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.578717 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-1" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.578940 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-2" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.631137 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6997548745-tb5n6"] Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643461 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-cell1\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643541 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-swift-storage-0\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643563 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-2\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-swift-storage-2\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643580 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-openstack-networker\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643596 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-1\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-swift-storage-1\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643617 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-openstack-cell1\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643640 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h29rv\" (UniqueName: \"kubernetes.io/projected/05958022-1fd8-4cb3-885a-4d49b69034cb-kube-api-access-h29rv\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643666 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-ovsdbserver-nb\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643684 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-config\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643736 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-config\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643763 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-svc\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643776 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643797 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-swift-storage-0\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643813 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-ovsdbserver-sb\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643843 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-networker\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643858 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643879 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqw8h\" (UniqueName: \"kubernetes.io/projected/605215fb-7a4c-49fb-a715-5c7251cad82e-kube-api-access-bqw8h\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.643899 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-svc\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.644663 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-cell1\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.645413 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-config\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.645914 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-svc\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.646969 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-sb\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.647030 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-nb\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.647323 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-swift-storage-0\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.648002 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-networker\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.665130 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h29rv\" (UniqueName: \"kubernetes.io/projected/05958022-1fd8-4cb3-885a-4d49b69034cb-kube-api-access-h29rv\") pod \"dnsmasq-dns-5b8dcfd55-xnv42\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.729468 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-rm6xq"] Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.739005 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.747243 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-ovsdbserver-nb\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.747279 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-config\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.747377 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-ovsdbserver-sb\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.747418 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqw8h\" (UniqueName: \"kubernetes.io/projected/605215fb-7a4c-49fb-a715-5c7251cad82e-kube-api-access-bqw8h\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.747443 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-svc\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.747531 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-swift-storage-0\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.747554 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-2\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-swift-storage-2\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.747573 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-openstack-networker\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.747592 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-1\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-swift-storage-1\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.747615 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-openstack-cell1\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.764104 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-rm6xq"] Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.786430 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-ovsdbserver-nb\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.789540 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-ovsdbserver-sb\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.789759 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-svc\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.790912 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-1\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-swift-storage-1\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.790934 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-swift-storage-0\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.791061 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-openstack-networker\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.791437 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-openstack-cell1\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.791669 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-2\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-dns-swift-storage-2\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.794686 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/605215fb-7a4c-49fb-a715-5c7251cad82e-config\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.794733 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqw8h\" (UniqueName: \"kubernetes.io/projected/605215fb-7a4c-49fb-a715-5c7251cad82e-kube-api-access-bqw8h\") pod \"dnsmasq-dns-6997548745-tb5n6\" (UID: \"605215fb-7a4c-49fb-a715-5c7251cad82e\") " pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.849319 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-dispersionconf\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.849381 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-ring-data-devices\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.849411 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-swiftconf\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.849504 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-scripts\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.849525 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.849585 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgvmz\" (UniqueName: \"kubernetes.io/projected/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-kube-api-access-zgvmz\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.849638 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-etc-swift\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.904743 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.950900 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-ring-data-devices\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.950953 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-swiftconf\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.951031 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-scripts\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.951051 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.951111 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgvmz\" (UniqueName: \"kubernetes.io/projected/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-kube-api-access-zgvmz\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.951141 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-etc-swift\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.951193 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-dispersionconf\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.954486 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-scripts\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.954556 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-ring-data-devices\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:31 crc kubenswrapper[4978]: I0225 09:41:31.954861 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-etc-swift\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.000062 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.386323 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-swiftconf\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.386470 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-dispersionconf\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.399345 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.399856 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgvmz\" (UniqueName: \"kubernetes.io/projected/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-kube-api-access-zgvmz\") pod \"swift-ring-rebalance-debug-rm6xq\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.510014 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.564968 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-cell1\") pod \"05958022-1fd8-4cb3-885a-4d49b69034cb\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.565074 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h29rv\" (UniqueName: \"kubernetes.io/projected/05958022-1fd8-4cb3-885a-4d49b69034cb-kube-api-access-h29rv\") pod \"05958022-1fd8-4cb3-885a-4d49b69034cb\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.565114 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-networker\") pod \"05958022-1fd8-4cb3-885a-4d49b69034cb\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.565350 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-swift-storage-0\") pod \"05958022-1fd8-4cb3-885a-4d49b69034cb\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.565526 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "05958022-1fd8-4cb3-885a-4d49b69034cb" (UID: "05958022-1fd8-4cb3-885a-4d49b69034cb"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.565710 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-networker" (OuterVolumeSpecName: "openstack-networker") pod "05958022-1fd8-4cb3-885a-4d49b69034cb" (UID: "05958022-1fd8-4cb3-885a-4d49b69034cb"). InnerVolumeSpecName "openstack-networker". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.565772 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "05958022-1fd8-4cb3-885a-4d49b69034cb" (UID: "05958022-1fd8-4cb3-885a-4d49b69034cb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.566821 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-nb\") pod \"05958022-1fd8-4cb3-885a-4d49b69034cb\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.566886 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-config\") pod \"05958022-1fd8-4cb3-885a-4d49b69034cb\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.566991 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-svc\") pod \"05958022-1fd8-4cb3-885a-4d49b69034cb\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.567049 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-sb\") pod \"05958022-1fd8-4cb3-885a-4d49b69034cb\" (UID: \"05958022-1fd8-4cb3-885a-4d49b69034cb\") " Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.567230 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "05958022-1fd8-4cb3-885a-4d49b69034cb" (UID: "05958022-1fd8-4cb3-885a-4d49b69034cb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.567290 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "05958022-1fd8-4cb3-885a-4d49b69034cb" (UID: "05958022-1fd8-4cb3-885a-4d49b69034cb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.567552 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-config" (OuterVolumeSpecName: "config") pod "05958022-1fd8-4cb3-885a-4d49b69034cb" (UID: "05958022-1fd8-4cb3-885a-4d49b69034cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.567656 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "05958022-1fd8-4cb3-885a-4d49b69034cb" (UID: "05958022-1fd8-4cb3-885a-4d49b69034cb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.567967 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.567990 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.568002 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.568011 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.568021 4978 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.568031 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.568040 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05958022-1fd8-4cb3-885a-4d49b69034cb-config\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.588668 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05958022-1fd8-4cb3-885a-4d49b69034cb-kube-api-access-h29rv" (OuterVolumeSpecName: "kube-api-access-h29rv") pod "05958022-1fd8-4cb3-885a-4d49b69034cb" (UID: "05958022-1fd8-4cb3-885a-4d49b69034cb"). InnerVolumeSpecName "kube-api-access-h29rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.670667 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h29rv\" (UniqueName: \"kubernetes.io/projected/05958022-1fd8-4cb3-885a-4d49b69034cb-kube-api-access-h29rv\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.678729 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:32 crc kubenswrapper[4978]: I0225 09:41:32.774440 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-rm6xq"] Feb 25 09:41:33 crc kubenswrapper[4978]: I0225 09:41:33.007270 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b8dcfd55-xnv42" Feb 25 09:41:33 crc kubenswrapper[4978]: I0225 09:41:33.032556 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6997548745-tb5n6"] Feb 25 09:41:33 crc kubenswrapper[4978]: I0225 09:41:33.217803 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-rm6xq"] Feb 25 09:41:33 crc kubenswrapper[4978]: I0225 09:41:33.290803 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b8dcfd55-xnv42"] Feb 25 09:41:33 crc kubenswrapper[4978]: I0225 09:41:33.311815 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b8dcfd55-xnv42"] Feb 25 09:41:33 crc kubenswrapper[4978]: I0225 09:41:33.344099 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05958022-1fd8-4cb3-885a-4d49b69034cb" path="/var/lib/kubelet/pods/05958022-1fd8-4cb3-885a-4d49b69034cb/volumes" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.021605 4978 generic.go:334] "Generic (PLEG): container finished" podID="605215fb-7a4c-49fb-a715-5c7251cad82e" containerID="ebb21290083f4538d6883c918556edfb33be1e09d1f8db97e71d78187a23f189" exitCode=0 Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.021793 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6997548745-tb5n6" event={"ID":"605215fb-7a4c-49fb-a715-5c7251cad82e","Type":"ContainerDied","Data":"ebb21290083f4538d6883c918556edfb33be1e09d1f8db97e71d78187a23f189"} Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.023650 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6997548745-tb5n6" event={"ID":"605215fb-7a4c-49fb-a715-5c7251cad82e","Type":"ContainerStarted","Data":"7b1fd2e8764d8f2ec808aad5b92d768fed0074f1b0ed124f285ab947df32575a"} Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.028774 4978 generic.go:334] "Generic (PLEG): container finished" podID="b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" containerID="237a82ebb374672ad89788f86b443ffb69d45bbb1f8b09c52090f2ad2c1a4afa" exitCode=2 Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.028807 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-rm6xq" event={"ID":"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f","Type":"ContainerDied","Data":"237a82ebb374672ad89788f86b443ffb69d45bbb1f8b09c52090f2ad2c1a4afa"} Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.028827 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-rm6xq" event={"ID":"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f","Type":"ContainerStarted","Data":"e00ad7216f315645846bffa46378f4a0cb1a83af2baf75cc14c1c6ab27ab9539"} Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.134615 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-rm6xq"] Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.150624 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-rm6xq"] Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.754135 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.830069 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-combined-ca-bundle\") pod \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.830181 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgvmz\" (UniqueName: \"kubernetes.io/projected/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-kube-api-access-zgvmz\") pod \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.830312 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-scripts\") pod \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.830396 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-swiftconf\") pod \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.830480 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-etc-swift\") pod \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.830515 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-ring-data-devices\") pod \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.830547 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-dispersionconf\") pod \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\" (UID: \"b12264ee-a5e1-4eb4-84ed-486b9bf4d11f\") " Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.831341 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" (UID: "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.831810 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" (UID: "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.832933 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.832992 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.838836 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-kube-api-access-zgvmz" (OuterVolumeSpecName: "kube-api-access-zgvmz") pod "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" (UID: "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f"). InnerVolumeSpecName "kube-api-access-zgvmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.870517 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" (UID: "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.889651 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" (UID: "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.894026 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-scripts" (OuterVolumeSpecName: "scripts") pod "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" (UID: "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.894408 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" (UID: "b12264ee-a5e1-4eb4-84ed-486b9bf4d11f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.934595 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.934631 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgvmz\" (UniqueName: \"kubernetes.io/projected/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-kube-api-access-zgvmz\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.934644 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.934653 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:34 crc kubenswrapper[4978]: I0225 09:41:34.934664 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:35 crc kubenswrapper[4978]: I0225 09:41:35.040254 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6997548745-tb5n6" event={"ID":"605215fb-7a4c-49fb-a715-5c7251cad82e","Type":"ContainerStarted","Data":"1fbd54b59d1cd85a59b945e081d012d9063ed1ab3f8b487cca75c54ea4edaf07"} Feb 25 09:41:35 crc kubenswrapper[4978]: I0225 09:41:35.041335 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:35 crc kubenswrapper[4978]: I0225 09:41:35.044110 4978 scope.go:117] "RemoveContainer" containerID="237a82ebb374672ad89788f86b443ffb69d45bbb1f8b09c52090f2ad2c1a4afa" Feb 25 09:41:35 crc kubenswrapper[4978]: I0225 09:41:35.044209 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-rm6xq" Feb 25 09:41:35 crc kubenswrapper[4978]: I0225 09:41:35.081224 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6997548745-tb5n6" podStartSLOduration=4.081200824 podStartE2EDuration="4.081200824s" podCreationTimestamp="2026-02-25 09:41:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:41:35.072225856 +0000 UTC m=+10588.511482355" watchObservedRunningTime="2026-02-25 09:41:35.081200824 +0000 UTC m=+10588.520457303" Feb 25 09:41:35 crc kubenswrapper[4978]: I0225 09:41:35.355242 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" path="/var/lib/kubelet/pods/b12264ee-a5e1-4eb4-84ed-486b9bf4d11f/volumes" Feb 25 09:41:36 crc kubenswrapper[4978]: I0225 09:41:36.074818 4978 generic.go:334] "Generic (PLEG): container finished" podID="751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" containerID="173b35239f6824f87f4feac0676964b501baca63b758633a0e27e4fef1a42540" exitCode=0 Feb 25 09:41:36 crc kubenswrapper[4978]: I0225 09:41:36.074911 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tlbw2" event={"ID":"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859","Type":"ContainerDied","Data":"173b35239f6824f87f4feac0676964b501baca63b758633a0e27e4fef1a42540"} Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.550728 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.589255 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf7ng\" (UniqueName: \"kubernetes.io/projected/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-kube-api-access-wf7ng\") pod \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.589322 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-ring-data-devices\") pod \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.589393 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-etc-swift\") pod \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.589459 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-scripts\") pod \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.589539 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-swiftconf\") pod \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.589587 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-dispersionconf\") pod \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.589711 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-combined-ca-bundle\") pod \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\" (UID: \"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859\") " Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.590637 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" (UID: "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.591611 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" (UID: "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.595286 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-kube-api-access-wf7ng" (OuterVolumeSpecName: "kube-api-access-wf7ng") pod "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" (UID: "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859"). InnerVolumeSpecName "kube-api-access-wf7ng". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.624109 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-scripts" (OuterVolumeSpecName: "scripts") pod "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" (UID: "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.633866 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" (UID: "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.634797 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" (UID: "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.651321 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" (UID: "751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.692365 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf7ng\" (UniqueName: \"kubernetes.io/projected/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-kube-api-access-wf7ng\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.692423 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.692439 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.692452 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.692464 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.692474 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:37 crc kubenswrapper[4978]: I0225 09:41:37.692486 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:38 crc kubenswrapper[4978]: I0225 09:41:38.111834 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tlbw2" event={"ID":"751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859","Type":"ContainerDied","Data":"04d33f7ce6b50ddc86e44c9f6c99264b98c0fc534aad659e475f9d6abcb1809f"} Feb 25 09:41:38 crc kubenswrapper[4978]: I0225 09:41:38.111914 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04d33f7ce6b50ddc86e44c9f6c99264b98c0fc534aad659e475f9d6abcb1809f" Feb 25 09:41:38 crc kubenswrapper[4978]: I0225 09:41:38.112009 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tlbw2" Feb 25 09:41:38 crc kubenswrapper[4978]: I0225 09:41:38.715933 4978 scope.go:117] "RemoveContainer" containerID="ba1fd6abbe8ff9ab0430c4fd0d93d1f9f6a597a2bc2ec7efeaaa32902a86ea89" Feb 25 09:41:41 crc kubenswrapper[4978]: I0225 09:41:41.906645 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6997548745-tb5n6" Feb 25 09:41:42 crc kubenswrapper[4978]: I0225 09:41:42.031880 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65cf54b65f-6jwcs"] Feb 25 09:41:42 crc kubenswrapper[4978]: I0225 09:41:42.032157 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" podUID="ae3313f2-822b-42cd-826a-daac63f0926c" containerName="dnsmasq-dns" containerID="cri-o://b6d84f4c7ef5cc6a0a8d1062144e6bef8c995b07806707f434b173641cd2f362" gracePeriod=10 Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.213744 4978 generic.go:334] "Generic (PLEG): container finished" podID="ae3313f2-822b-42cd-826a-daac63f0926c" containerID="b6d84f4c7ef5cc6a0a8d1062144e6bef8c995b07806707f434b173641cd2f362" exitCode=0 Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.214313 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" event={"ID":"ae3313f2-822b-42cd-826a-daac63f0926c","Type":"ContainerDied","Data":"b6d84f4c7ef5cc6a0a8d1062144e6bef8c995b07806707f434b173641cd2f362"} Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.214355 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" event={"ID":"ae3313f2-822b-42cd-826a-daac63f0926c","Type":"ContainerDied","Data":"5bda5b472b78c8549b087f0ae648ee30d742ad71773ff0f840e5e436533df37e"} Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.214380 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bda5b472b78c8549b087f0ae648ee30d742ad71773ff0f840e5e436533df37e" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.224174 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.328242 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:41:43 crc kubenswrapper[4978]: E0225 09:41:43.328474 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.372732 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-dns-svc\") pod \"ae3313f2-822b-42cd-826a-daac63f0926c\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.372808 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-config\") pod \"ae3313f2-822b-42cd-826a-daac63f0926c\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.372978 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-cell1\") pod \"ae3313f2-822b-42cd-826a-daac63f0926c\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.373009 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-sb\") pod \"ae3313f2-822b-42cd-826a-daac63f0926c\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.373065 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-nb\") pod \"ae3313f2-822b-42cd-826a-daac63f0926c\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.373105 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-networker\") pod \"ae3313f2-822b-42cd-826a-daac63f0926c\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.373186 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw9v6\" (UniqueName: \"kubernetes.io/projected/ae3313f2-822b-42cd-826a-daac63f0926c-kube-api-access-bw9v6\") pod \"ae3313f2-822b-42cd-826a-daac63f0926c\" (UID: \"ae3313f2-822b-42cd-826a-daac63f0926c\") " Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.381853 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae3313f2-822b-42cd-826a-daac63f0926c-kube-api-access-bw9v6" (OuterVolumeSpecName: "kube-api-access-bw9v6") pod "ae3313f2-822b-42cd-826a-daac63f0926c" (UID: "ae3313f2-822b-42cd-826a-daac63f0926c"). InnerVolumeSpecName "kube-api-access-bw9v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.438446 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ae3313f2-822b-42cd-826a-daac63f0926c" (UID: "ae3313f2-822b-42cd-826a-daac63f0926c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.441830 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-networker" (OuterVolumeSpecName: "openstack-networker") pod "ae3313f2-822b-42cd-826a-daac63f0926c" (UID: "ae3313f2-822b-42cd-826a-daac63f0926c"). InnerVolumeSpecName "openstack-networker". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.445828 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ae3313f2-822b-42cd-826a-daac63f0926c" (UID: "ae3313f2-822b-42cd-826a-daac63f0926c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.451047 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ae3313f2-822b-42cd-826a-daac63f0926c" (UID: "ae3313f2-822b-42cd-826a-daac63f0926c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.454055 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "ae3313f2-822b-42cd-826a-daac63f0926c" (UID: "ae3313f2-822b-42cd-826a-daac63f0926c"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.465326 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-config" (OuterVolumeSpecName: "config") pod "ae3313f2-822b-42cd-826a-daac63f0926c" (UID: "ae3313f2-822b-42cd-826a-daac63f0926c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.476121 4978 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-config\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.476261 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-cell1\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.476404 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.476491 4978 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.476566 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-networker\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-openstack-networker\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.476659 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw9v6\" (UniqueName: \"kubernetes.io/projected/ae3313f2-822b-42cd-826a-daac63f0926c-kube-api-access-bw9v6\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:43 crc kubenswrapper[4978]: I0225 09:41:43.476754 4978 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae3313f2-822b-42cd-826a-daac63f0926c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 25 09:41:44 crc kubenswrapper[4978]: I0225 09:41:44.223548 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65cf54b65f-6jwcs" Feb 25 09:41:44 crc kubenswrapper[4978]: I0225 09:41:44.268509 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65cf54b65f-6jwcs"] Feb 25 09:41:44 crc kubenswrapper[4978]: I0225 09:41:44.280935 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65cf54b65f-6jwcs"] Feb 25 09:41:45 crc kubenswrapper[4978]: I0225 09:41:45.365276 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae3313f2-822b-42cd-826a-daac63f0926c" path="/var/lib/kubelet/pods/ae3313f2-822b-42cd-826a-daac63f0926c/volumes" Feb 25 09:41:57 crc kubenswrapper[4978]: I0225 09:41:57.349067 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:41:57 crc kubenswrapper[4978]: E0225 09:41:57.350067 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.165585 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533542-6w6zb"] Feb 25 09:42:00 crc kubenswrapper[4978]: E0225 09:42:00.166584 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3313f2-822b-42cd-826a-daac63f0926c" containerName="dnsmasq-dns" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.166605 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3313f2-822b-42cd-826a-daac63f0926c" containerName="dnsmasq-dns" Feb 25 09:42:00 crc kubenswrapper[4978]: E0225 09:42:00.166638 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3313f2-822b-42cd-826a-daac63f0926c" containerName="init" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.166646 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3313f2-822b-42cd-826a-daac63f0926c" containerName="init" Feb 25 09:42:00 crc kubenswrapper[4978]: E0225 09:42:00.166662 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" containerName="swift-ring-rebalance" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.166671 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" containerName="swift-ring-rebalance" Feb 25 09:42:00 crc kubenswrapper[4978]: E0225 09:42:00.166691 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" containerName="swift-ring-rebalance" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.166699 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" containerName="swift-ring-rebalance" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.167027 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b12264ee-a5e1-4eb4-84ed-486b9bf4d11f" containerName="swift-ring-rebalance" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.167058 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859" containerName="swift-ring-rebalance" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.167092 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3313f2-822b-42cd-826a-daac63f0926c" containerName="dnsmasq-dns" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.168178 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533542-6w6zb" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.174431 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.174672 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.174819 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.182693 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533542-6w6zb"] Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.187831 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz9g9\" (UniqueName: \"kubernetes.io/projected/83565f13-d815-4310-b957-0c4c204bd7f5-kube-api-access-jz9g9\") pod \"auto-csr-approver-29533542-6w6zb\" (UID: \"83565f13-d815-4310-b957-0c4c204bd7f5\") " pod="openshift-infra/auto-csr-approver-29533542-6w6zb" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.291080 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz9g9\" (UniqueName: \"kubernetes.io/projected/83565f13-d815-4310-b957-0c4c204bd7f5-kube-api-access-jz9g9\") pod \"auto-csr-approver-29533542-6w6zb\" (UID: \"83565f13-d815-4310-b957-0c4c204bd7f5\") " pod="openshift-infra/auto-csr-approver-29533542-6w6zb" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.316803 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz9g9\" (UniqueName: \"kubernetes.io/projected/83565f13-d815-4310-b957-0c4c204bd7f5-kube-api-access-jz9g9\") pod \"auto-csr-approver-29533542-6w6zb\" (UID: \"83565f13-d815-4310-b957-0c4c204bd7f5\") " pod="openshift-infra/auto-csr-approver-29533542-6w6zb" Feb 25 09:42:00 crc kubenswrapper[4978]: I0225 09:42:00.504335 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533542-6w6zb" Feb 25 09:42:01 crc kubenswrapper[4978]: I0225 09:42:01.030962 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533542-6w6zb"] Feb 25 09:42:01 crc kubenswrapper[4978]: W0225 09:42:01.033698 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83565f13_d815_4310_b957_0c4c204bd7f5.slice/crio-b4b7039b800f9ce655d057815e65d846dd308e499f77c2e133d47e194b0f9de1 WatchSource:0}: Error finding container b4b7039b800f9ce655d057815e65d846dd308e499f77c2e133d47e194b0f9de1: Status 404 returned error can't find the container with id b4b7039b800f9ce655d057815e65d846dd308e499f77c2e133d47e194b0f9de1 Feb 25 09:42:01 crc kubenswrapper[4978]: I0225 09:42:01.444019 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533542-6w6zb" event={"ID":"83565f13-d815-4310-b957-0c4c204bd7f5","Type":"ContainerStarted","Data":"b4b7039b800f9ce655d057815e65d846dd308e499f77c2e133d47e194b0f9de1"} Feb 25 09:42:03 crc kubenswrapper[4978]: I0225 09:42:03.472463 4978 generic.go:334] "Generic (PLEG): container finished" podID="83565f13-d815-4310-b957-0c4c204bd7f5" containerID="fd87ff67314e8bfd306f7e65ee9d288807e6b8fe583d2129495293c47340ec7c" exitCode=0 Feb 25 09:42:03 crc kubenswrapper[4978]: I0225 09:42:03.472631 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533542-6w6zb" event={"ID":"83565f13-d815-4310-b957-0c4c204bd7f5","Type":"ContainerDied","Data":"fd87ff67314e8bfd306f7e65ee9d288807e6b8fe583d2129495293c47340ec7c"} Feb 25 09:42:04 crc kubenswrapper[4978]: I0225 09:42:04.929353 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533542-6w6zb" Feb 25 09:42:05 crc kubenswrapper[4978]: I0225 09:42:05.020968 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz9g9\" (UniqueName: \"kubernetes.io/projected/83565f13-d815-4310-b957-0c4c204bd7f5-kube-api-access-jz9g9\") pod \"83565f13-d815-4310-b957-0c4c204bd7f5\" (UID: \"83565f13-d815-4310-b957-0c4c204bd7f5\") " Feb 25 09:42:05 crc kubenswrapper[4978]: I0225 09:42:05.026213 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83565f13-d815-4310-b957-0c4c204bd7f5-kube-api-access-jz9g9" (OuterVolumeSpecName: "kube-api-access-jz9g9") pod "83565f13-d815-4310-b957-0c4c204bd7f5" (UID: "83565f13-d815-4310-b957-0c4c204bd7f5"). InnerVolumeSpecName "kube-api-access-jz9g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:42:05 crc kubenswrapper[4978]: I0225 09:42:05.124074 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz9g9\" (UniqueName: \"kubernetes.io/projected/83565f13-d815-4310-b957-0c4c204bd7f5-kube-api-access-jz9g9\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:05 crc kubenswrapper[4978]: I0225 09:42:05.504186 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533542-6w6zb" event={"ID":"83565f13-d815-4310-b957-0c4c204bd7f5","Type":"ContainerDied","Data":"b4b7039b800f9ce655d057815e65d846dd308e499f77c2e133d47e194b0f9de1"} Feb 25 09:42:05 crc kubenswrapper[4978]: I0225 09:42:05.504236 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533542-6w6zb" Feb 25 09:42:05 crc kubenswrapper[4978]: I0225 09:42:05.504251 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4b7039b800f9ce655d057815e65d846dd308e499f77c2e133d47e194b0f9de1" Feb 25 09:42:06 crc kubenswrapper[4978]: I0225 09:42:06.029410 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533536-jdbn9"] Feb 25 09:42:06 crc kubenswrapper[4978]: I0225 09:42:06.041794 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533536-jdbn9"] Feb 25 09:42:07 crc kubenswrapper[4978]: I0225 09:42:07.351908 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de6a92a4-917e-49a6-ad0d-e4455c8a0eb8" path="/var/lib/kubelet/pods/de6a92a4-917e-49a6-ad0d-e4455c8a0eb8/volumes" Feb 25 09:42:10 crc kubenswrapper[4978]: I0225 09:42:10.329467 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:42:10 crc kubenswrapper[4978]: E0225 09:42:10.329788 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:42:22 crc kubenswrapper[4978]: I0225 09:42:22.328161 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:42:22 crc kubenswrapper[4978]: E0225 09:42:22.329050 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.309675 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-x96bj"] Feb 25 09:42:34 crc kubenswrapper[4978]: E0225 09:42:34.310610 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83565f13-d815-4310-b957-0c4c204bd7f5" containerName="oc" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.310622 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="83565f13-d815-4310-b957-0c4c204bd7f5" containerName="oc" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.310818 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="83565f13-d815-4310-b957-0c4c204bd7f5" containerName="oc" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.311531 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.315183 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.315699 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.334531 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-x96bj"] Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.351500 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-dispersionconf\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.351800 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-swiftconf\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.351990 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-etc-swift\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.352167 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-scripts\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.352288 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.352429 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-ring-data-devices\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.352813 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptjk6\" (UniqueName: \"kubernetes.io/projected/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-kube-api-access-ptjk6\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.455200 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-scripts\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.455287 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-ring-data-devices\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.455341 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.455610 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptjk6\" (UniqueName: \"kubernetes.io/projected/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-kube-api-access-ptjk6\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.455798 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-dispersionconf\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.455840 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-swiftconf\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.455937 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-etc-swift\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.456283 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-scripts\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.456660 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-ring-data-devices\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.457180 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-etc-swift\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.462626 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-dispersionconf\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.463667 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-swiftconf\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.465062 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.479335 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptjk6\" (UniqueName: \"kubernetes.io/projected/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-kube-api-access-ptjk6\") pod \"swift-ring-rebalance-debug-x96bj\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.651220 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:34 crc kubenswrapper[4978]: I0225 09:42:34.983902 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-x96bj"] Feb 25 09:42:35 crc kubenswrapper[4978]: I0225 09:42:35.933045 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-x96bj" event={"ID":"4647734b-2db6-4a7f-bdf0-8a37d0ae6443","Type":"ContainerStarted","Data":"ef90200e91e588db9aabdfe18f64cc449c8463ff06878e5617906a6a93b9fd99"} Feb 25 09:42:36 crc kubenswrapper[4978]: I0225 09:42:36.327859 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:42:36 crc kubenswrapper[4978]: E0225 09:42:36.328804 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:42:36 crc kubenswrapper[4978]: I0225 09:42:36.946625 4978 generic.go:334] "Generic (PLEG): container finished" podID="4647734b-2db6-4a7f-bdf0-8a37d0ae6443" containerID="5e60e0b8b91c21393ad1013d898fed088dcddc7070f65ead3b2cce7e9bc4925f" exitCode=0 Feb 25 09:42:36 crc kubenswrapper[4978]: I0225 09:42:36.946718 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-x96bj" event={"ID":"4647734b-2db6-4a7f-bdf0-8a37d0ae6443","Type":"ContainerDied","Data":"5e60e0b8b91c21393ad1013d898fed088dcddc7070f65ead3b2cce7e9bc4925f"} Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.011530 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-x96bj"] Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.025694 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-x96bj"] Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.482772 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-vngkf"] Feb 25 09:42:37 crc kubenswrapper[4978]: E0225 09:42:37.483890 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4647734b-2db6-4a7f-bdf0-8a37d0ae6443" containerName="swift-ring-rebalance" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.483928 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4647734b-2db6-4a7f-bdf0-8a37d0ae6443" containerName="swift-ring-rebalance" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.484347 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4647734b-2db6-4a7f-bdf0-8a37d0ae6443" containerName="swift-ring-rebalance" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.485702 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.499135 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-vngkf"] Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.623134 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-swiftconf\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.623475 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.623676 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-dispersionconf\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.623802 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-scripts\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.623928 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-ring-data-devices\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.624046 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbgdb\" (UniqueName: \"kubernetes.io/projected/44f1cb94-2071-4c34-8588-6a7d26aadc24-kube-api-access-xbgdb\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.624175 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/44f1cb94-2071-4c34-8588-6a7d26aadc24-etc-swift\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.726446 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-swiftconf\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.726529 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.726617 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-dispersionconf\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.726669 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-scripts\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.726721 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-ring-data-devices\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.726758 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbgdb\" (UniqueName: \"kubernetes.io/projected/44f1cb94-2071-4c34-8588-6a7d26aadc24-kube-api-access-xbgdb\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.726814 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/44f1cb94-2071-4c34-8588-6a7d26aadc24-etc-swift\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.727806 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/44f1cb94-2071-4c34-8588-6a7d26aadc24-etc-swift\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.728470 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-ring-data-devices\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:37 crc kubenswrapper[4978]: I0225 09:42:37.728535 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-scripts\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.286664 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-dispersionconf\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.286887 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-swiftconf\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.287180 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.288412 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbgdb\" (UniqueName: \"kubernetes.io/projected/44f1cb94-2071-4c34-8588-6a7d26aadc24-kube-api-access-xbgdb\") pod \"swift-ring-rebalance-debug-vngkf\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.419750 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.570765 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.751167 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-combined-ca-bundle\") pod \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.751219 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-swiftconf\") pod \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.751265 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-dispersionconf\") pod \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.751318 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-scripts\") pod \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.751361 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptjk6\" (UniqueName: \"kubernetes.io/projected/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-kube-api-access-ptjk6\") pod \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.751407 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-ring-data-devices\") pod \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.751484 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-etc-swift\") pod \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\" (UID: \"4647734b-2db6-4a7f-bdf0-8a37d0ae6443\") " Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.753034 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4647734b-2db6-4a7f-bdf0-8a37d0ae6443" (UID: "4647734b-2db6-4a7f-bdf0-8a37d0ae6443"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.753144 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4647734b-2db6-4a7f-bdf0-8a37d0ae6443" (UID: "4647734b-2db6-4a7f-bdf0-8a37d0ae6443"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.760056 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-kube-api-access-ptjk6" (OuterVolumeSpecName: "kube-api-access-ptjk6") pod "4647734b-2db6-4a7f-bdf0-8a37d0ae6443" (UID: "4647734b-2db6-4a7f-bdf0-8a37d0ae6443"). InnerVolumeSpecName "kube-api-access-ptjk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.790014 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-scripts" (OuterVolumeSpecName: "scripts") pod "4647734b-2db6-4a7f-bdf0-8a37d0ae6443" (UID: "4647734b-2db6-4a7f-bdf0-8a37d0ae6443"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.792149 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4647734b-2db6-4a7f-bdf0-8a37d0ae6443" (UID: "4647734b-2db6-4a7f-bdf0-8a37d0ae6443"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.798714 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4647734b-2db6-4a7f-bdf0-8a37d0ae6443" (UID: "4647734b-2db6-4a7f-bdf0-8a37d0ae6443"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.822621 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4647734b-2db6-4a7f-bdf0-8a37d0ae6443" (UID: "4647734b-2db6-4a7f-bdf0-8a37d0ae6443"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.829979 4978 scope.go:117] "RemoveContainer" containerID="f7a06736fe3496594a681d6fd01f1970c4c0db5e15cc79375fd86f62b61f1859" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.855414 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.855987 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.856018 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.856047 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.856074 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptjk6\" (UniqueName: \"kubernetes.io/projected/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-kube-api-access-ptjk6\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.856099 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.856123 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4647734b-2db6-4a7f-bdf0-8a37d0ae6443-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.861179 4978 scope.go:117] "RemoveContainer" containerID="b25b3cb54f308bc7cf51fa83778276997329ef0e190f6c54b92979e345caa0bf" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.915931 4978 scope.go:117] "RemoveContainer" containerID="b6d84f4c7ef5cc6a0a8d1062144e6bef8c995b07806707f434b173641cd2f362" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.973174 4978 scope.go:117] "RemoveContainer" containerID="5e60e0b8b91c21393ad1013d898fed088dcddc7070f65ead3b2cce7e9bc4925f" Feb 25 09:42:38 crc kubenswrapper[4978]: I0225 09:42:38.973208 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-x96bj" Feb 25 09:42:39 crc kubenswrapper[4978]: I0225 09:42:39.121488 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-vngkf"] Feb 25 09:42:39 crc kubenswrapper[4978]: W0225 09:42:39.125144 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44f1cb94_2071_4c34_8588_6a7d26aadc24.slice/crio-05bd5a6cd22f030407e8305e776368b2f478e4a8e831a57bdd4beab9ca49e38b WatchSource:0}: Error finding container 05bd5a6cd22f030407e8305e776368b2f478e4a8e831a57bdd4beab9ca49e38b: Status 404 returned error can't find the container with id 05bd5a6cd22f030407e8305e776368b2f478e4a8e831a57bdd4beab9ca49e38b Feb 25 09:42:39 crc kubenswrapper[4978]: I0225 09:42:39.363284 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4647734b-2db6-4a7f-bdf0-8a37d0ae6443" path="/var/lib/kubelet/pods/4647734b-2db6-4a7f-bdf0-8a37d0ae6443/volumes" Feb 25 09:42:39 crc kubenswrapper[4978]: I0225 09:42:39.991573 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-vngkf" event={"ID":"44f1cb94-2071-4c34-8588-6a7d26aadc24","Type":"ContainerStarted","Data":"99964fba2fac870611f765e084cf039b3d4601a2cf718b386563a48ab5f39822"} Feb 25 09:42:39 crc kubenswrapper[4978]: I0225 09:42:39.991618 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-vngkf" event={"ID":"44f1cb94-2071-4c34-8588-6a7d26aadc24","Type":"ContainerStarted","Data":"05bd5a6cd22f030407e8305e776368b2f478e4a8e831a57bdd4beab9ca49e38b"} Feb 25 09:42:47 crc kubenswrapper[4978]: I0225 09:42:47.077540 4978 generic.go:334] "Generic (PLEG): container finished" podID="44f1cb94-2071-4c34-8588-6a7d26aadc24" containerID="99964fba2fac870611f765e084cf039b3d4601a2cf718b386563a48ab5f39822" exitCode=0 Feb 25 09:42:47 crc kubenswrapper[4978]: I0225 09:42:47.077644 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-vngkf" event={"ID":"44f1cb94-2071-4c34-8588-6a7d26aadc24","Type":"ContainerDied","Data":"99964fba2fac870611f765e084cf039b3d4601a2cf718b386563a48ab5f39822"} Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.499382 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.585015 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-vngkf"] Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.593292 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-combined-ca-bundle\") pod \"44f1cb94-2071-4c34-8588-6a7d26aadc24\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.593943 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbgdb\" (UniqueName: \"kubernetes.io/projected/44f1cb94-2071-4c34-8588-6a7d26aadc24-kube-api-access-xbgdb\") pod \"44f1cb94-2071-4c34-8588-6a7d26aadc24\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.594116 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-ring-data-devices\") pod \"44f1cb94-2071-4c34-8588-6a7d26aadc24\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.594979 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "44f1cb94-2071-4c34-8588-6a7d26aadc24" (UID: "44f1cb94-2071-4c34-8588-6a7d26aadc24"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.595094 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-swiftconf\") pod \"44f1cb94-2071-4c34-8588-6a7d26aadc24\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.595409 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/44f1cb94-2071-4c34-8588-6a7d26aadc24-etc-swift\") pod \"44f1cb94-2071-4c34-8588-6a7d26aadc24\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.595677 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-scripts\") pod \"44f1cb94-2071-4c34-8588-6a7d26aadc24\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.595897 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-dispersionconf\") pod \"44f1cb94-2071-4c34-8588-6a7d26aadc24\" (UID: \"44f1cb94-2071-4c34-8588-6a7d26aadc24\") " Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.596661 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44f1cb94-2071-4c34-8588-6a7d26aadc24-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "44f1cb94-2071-4c34-8588-6a7d26aadc24" (UID: "44f1cb94-2071-4c34-8588-6a7d26aadc24"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.597395 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/44f1cb94-2071-4c34-8588-6a7d26aadc24-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.597421 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.605516 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44f1cb94-2071-4c34-8588-6a7d26aadc24-kube-api-access-xbgdb" (OuterVolumeSpecName: "kube-api-access-xbgdb") pod "44f1cb94-2071-4c34-8588-6a7d26aadc24" (UID: "44f1cb94-2071-4c34-8588-6a7d26aadc24"). InnerVolumeSpecName "kube-api-access-xbgdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.613252 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-vngkf"] Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.628429 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44f1cb94-2071-4c34-8588-6a7d26aadc24" (UID: "44f1cb94-2071-4c34-8588-6a7d26aadc24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.632746 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-scripts" (OuterVolumeSpecName: "scripts") pod "44f1cb94-2071-4c34-8588-6a7d26aadc24" (UID: "44f1cb94-2071-4c34-8588-6a7d26aadc24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.636408 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "44f1cb94-2071-4c34-8588-6a7d26aadc24" (UID: "44f1cb94-2071-4c34-8588-6a7d26aadc24"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.638262 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "44f1cb94-2071-4c34-8588-6a7d26aadc24" (UID: "44f1cb94-2071-4c34-8588-6a7d26aadc24"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.700110 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.700145 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbgdb\" (UniqueName: \"kubernetes.io/projected/44f1cb94-2071-4c34-8588-6a7d26aadc24-kube-api-access-xbgdb\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.700161 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.700174 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44f1cb94-2071-4c34-8588-6a7d26aadc24-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:48 crc kubenswrapper[4978]: I0225 09:42:48.700186 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/44f1cb94-2071-4c34-8588-6a7d26aadc24-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:42:49 crc kubenswrapper[4978]: I0225 09:42:49.102562 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05bd5a6cd22f030407e8305e776368b2f478e4a8e831a57bdd4beab9ca49e38b" Feb 25 09:42:49 crc kubenswrapper[4978]: I0225 09:42:49.102627 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-vngkf" Feb 25 09:42:49 crc kubenswrapper[4978]: I0225 09:42:49.344152 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44f1cb94-2071-4c34-8588-6a7d26aadc24" path="/var/lib/kubelet/pods/44f1cb94-2071-4c34-8588-6a7d26aadc24/volumes" Feb 25 09:42:51 crc kubenswrapper[4978]: I0225 09:42:51.328327 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:42:51 crc kubenswrapper[4978]: E0225 09:42:51.328837 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.138822 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-dsmgq"] Feb 25 09:42:52 crc kubenswrapper[4978]: E0225 09:42:52.139725 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44f1cb94-2071-4c34-8588-6a7d26aadc24" containerName="swift-ring-rebalance" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.139742 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="44f1cb94-2071-4c34-8588-6a7d26aadc24" containerName="swift-ring-rebalance" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.140013 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="44f1cb94-2071-4c34-8588-6a7d26aadc24" containerName="swift-ring-rebalance" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.140884 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.149065 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.151472 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.158870 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-dsmgq"] Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.291170 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e8479ccb-429a-4ce8-834c-977bf150fb33-etc-swift\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.291332 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.291421 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-scripts\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.291444 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-dispersionconf\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.291705 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jshv\" (UniqueName: \"kubernetes.io/projected/e8479ccb-429a-4ce8-834c-977bf150fb33-kube-api-access-6jshv\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.291758 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-ring-data-devices\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.291845 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-swiftconf\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.394486 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e8479ccb-429a-4ce8-834c-977bf150fb33-etc-swift\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.394619 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.394672 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-dispersionconf\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.394712 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-scripts\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.394856 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jshv\" (UniqueName: \"kubernetes.io/projected/e8479ccb-429a-4ce8-834c-977bf150fb33-kube-api-access-6jshv\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.394892 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-ring-data-devices\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.394957 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-swiftconf\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.395218 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e8479ccb-429a-4ce8-834c-977bf150fb33-etc-swift\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.396556 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-scripts\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.397185 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-ring-data-devices\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.402164 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-dispersionconf\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.402295 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-swiftconf\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.402463 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.418125 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jshv\" (UniqueName: \"kubernetes.io/projected/e8479ccb-429a-4ce8-834c-977bf150fb33-kube-api-access-6jshv\") pod \"swift-ring-rebalance-debug-dsmgq\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:52 crc kubenswrapper[4978]: I0225 09:42:52.480004 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:42:53 crc kubenswrapper[4978]: I0225 09:42:53.013200 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-dsmgq"] Feb 25 09:42:53 crc kubenswrapper[4978]: I0225 09:42:53.177910 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-dsmgq" event={"ID":"e8479ccb-429a-4ce8-834c-977bf150fb33","Type":"ContainerStarted","Data":"ce5bc940cae41a847f9d1ec77cb5b07e1c461e5638a96c756394cf8b5694f4cf"} Feb 25 09:42:54 crc kubenswrapper[4978]: I0225 09:42:54.191021 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-dsmgq" event={"ID":"e8479ccb-429a-4ce8-834c-977bf150fb33","Type":"ContainerStarted","Data":"4c8626fa4fcab0ab57df1558940121fe1f9b7e43ee9753873ad98e2baf118846"} Feb 25 09:42:54 crc kubenswrapper[4978]: I0225 09:42:54.220504 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-dsmgq" podStartSLOduration=2.22047776 podStartE2EDuration="2.22047776s" podCreationTimestamp="2026-02-25 09:42:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:42:54.207651731 +0000 UTC m=+10667.646908220" watchObservedRunningTime="2026-02-25 09:42:54.22047776 +0000 UTC m=+10667.659734229" Feb 25 09:43:02 crc kubenswrapper[4978]: I0225 09:43:02.293568 4978 generic.go:334] "Generic (PLEG): container finished" podID="e8479ccb-429a-4ce8-834c-977bf150fb33" containerID="4c8626fa4fcab0ab57df1558940121fe1f9b7e43ee9753873ad98e2baf118846" exitCode=0 Feb 25 09:43:02 crc kubenswrapper[4978]: I0225 09:43:02.293695 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-dsmgq" event={"ID":"e8479ccb-429a-4ce8-834c-977bf150fb33","Type":"ContainerDied","Data":"4c8626fa4fcab0ab57df1558940121fe1f9b7e43ee9753873ad98e2baf118846"} Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.104459 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.161165 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-dsmgq"] Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.173420 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-scripts\") pod \"e8479ccb-429a-4ce8-834c-977bf150fb33\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.173516 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jshv\" (UniqueName: \"kubernetes.io/projected/e8479ccb-429a-4ce8-834c-977bf150fb33-kube-api-access-6jshv\") pod \"e8479ccb-429a-4ce8-834c-977bf150fb33\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.173546 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-combined-ca-bundle\") pod \"e8479ccb-429a-4ce8-834c-977bf150fb33\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.173572 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-ring-data-devices\") pod \"e8479ccb-429a-4ce8-834c-977bf150fb33\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.173678 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-swiftconf\") pod \"e8479ccb-429a-4ce8-834c-977bf150fb33\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.173731 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-dispersionconf\") pod \"e8479ccb-429a-4ce8-834c-977bf150fb33\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.173781 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e8479ccb-429a-4ce8-834c-977bf150fb33-etc-swift\") pod \"e8479ccb-429a-4ce8-834c-977bf150fb33\" (UID: \"e8479ccb-429a-4ce8-834c-977bf150fb33\") " Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.174292 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e8479ccb-429a-4ce8-834c-977bf150fb33" (UID: "e8479ccb-429a-4ce8-834c-977bf150fb33"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.174548 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-dsmgq"] Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.175407 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8479ccb-429a-4ce8-834c-977bf150fb33-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e8479ccb-429a-4ce8-834c-977bf150fb33" (UID: "e8479ccb-429a-4ce8-834c-977bf150fb33"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.200845 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8479ccb-429a-4ce8-834c-977bf150fb33-kube-api-access-6jshv" (OuterVolumeSpecName: "kube-api-access-6jshv") pod "e8479ccb-429a-4ce8-834c-977bf150fb33" (UID: "e8479ccb-429a-4ce8-834c-977bf150fb33"). InnerVolumeSpecName "kube-api-access-6jshv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.210603 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e8479ccb-429a-4ce8-834c-977bf150fb33" (UID: "e8479ccb-429a-4ce8-834c-977bf150fb33"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.214163 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e8479ccb-429a-4ce8-834c-977bf150fb33" (UID: "e8479ccb-429a-4ce8-834c-977bf150fb33"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.225902 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-scripts" (OuterVolumeSpecName: "scripts") pod "e8479ccb-429a-4ce8-834c-977bf150fb33" (UID: "e8479ccb-429a-4ce8-834c-977bf150fb33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.230580 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8479ccb-429a-4ce8-834c-977bf150fb33" (UID: "e8479ccb-429a-4ce8-834c-977bf150fb33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.276102 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.276138 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jshv\" (UniqueName: \"kubernetes.io/projected/e8479ccb-429a-4ce8-834c-977bf150fb33-kube-api-access-6jshv\") on node \"crc\" DevicePath \"\"" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.276150 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.276160 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e8479ccb-429a-4ce8-834c-977bf150fb33-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.276168 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.276179 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e8479ccb-429a-4ce8-834c-977bf150fb33-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.276191 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e8479ccb-429a-4ce8-834c-977bf150fb33-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.319305 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce5bc940cae41a847f9d1ec77cb5b07e1c461e5638a96c756394cf8b5694f4cf" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.319356 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-dsmgq" Feb 25 09:43:04 crc kubenswrapper[4978]: I0225 09:43:04.328350 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:43:04 crc kubenswrapper[4978]: E0225 09:43:04.328798 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:43:05 crc kubenswrapper[4978]: I0225 09:43:05.348227 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8479ccb-429a-4ce8-834c-977bf150fb33" path="/var/lib/kubelet/pods/e8479ccb-429a-4ce8-834c-977bf150fb33/volumes" Feb 25 09:43:17 crc kubenswrapper[4978]: I0225 09:43:17.344012 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:43:17 crc kubenswrapper[4978]: E0225 09:43:17.345239 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:43:31 crc kubenswrapper[4978]: I0225 09:43:31.327160 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:43:31 crc kubenswrapper[4978]: E0225 09:43:31.328521 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:43:45 crc kubenswrapper[4978]: I0225 09:43:45.328762 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:43:45 crc kubenswrapper[4978]: E0225 09:43:45.332236 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.162850 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533544-bjw6s"] Feb 25 09:44:00 crc kubenswrapper[4978]: E0225 09:44:00.164083 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8479ccb-429a-4ce8-834c-977bf150fb33" containerName="swift-ring-rebalance" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.164106 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8479ccb-429a-4ce8-834c-977bf150fb33" containerName="swift-ring-rebalance" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.164518 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8479ccb-429a-4ce8-834c-977bf150fb33" containerName="swift-ring-rebalance" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.165701 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533544-bjw6s" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.170287 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.175254 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.175464 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.183999 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533544-bjw6s"] Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.275337 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rvnz\" (UniqueName: \"kubernetes.io/projected/677ce558-3f07-465e-b6a2-97126eaea074-kube-api-access-7rvnz\") pod \"auto-csr-approver-29533544-bjw6s\" (UID: \"677ce558-3f07-465e-b6a2-97126eaea074\") " pod="openshift-infra/auto-csr-approver-29533544-bjw6s" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.327501 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:44:00 crc kubenswrapper[4978]: E0225 09:44:00.327844 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.377661 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rvnz\" (UniqueName: \"kubernetes.io/projected/677ce558-3f07-465e-b6a2-97126eaea074-kube-api-access-7rvnz\") pod \"auto-csr-approver-29533544-bjw6s\" (UID: \"677ce558-3f07-465e-b6a2-97126eaea074\") " pod="openshift-infra/auto-csr-approver-29533544-bjw6s" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.408019 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rvnz\" (UniqueName: \"kubernetes.io/projected/677ce558-3f07-465e-b6a2-97126eaea074-kube-api-access-7rvnz\") pod \"auto-csr-approver-29533544-bjw6s\" (UID: \"677ce558-3f07-465e-b6a2-97126eaea074\") " pod="openshift-infra/auto-csr-approver-29533544-bjw6s" Feb 25 09:44:00 crc kubenswrapper[4978]: I0225 09:44:00.498877 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533544-bjw6s" Feb 25 09:44:01 crc kubenswrapper[4978]: I0225 09:44:01.009596 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533544-bjw6s"] Feb 25 09:44:01 crc kubenswrapper[4978]: I0225 09:44:01.211514 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533544-bjw6s" event={"ID":"677ce558-3f07-465e-b6a2-97126eaea074","Type":"ContainerStarted","Data":"18ed0b7c418212a8f8dde5da59a8bd957f7696c942050a35fbfc3d5b363a67ce"} Feb 25 09:44:03 crc kubenswrapper[4978]: I0225 09:44:03.233798 4978 generic.go:334] "Generic (PLEG): container finished" podID="677ce558-3f07-465e-b6a2-97126eaea074" containerID="9e8674c1a831eddf11614a8d71a99ffcd693a78212eca2b09d49dea60d79075c" exitCode=0 Feb 25 09:44:03 crc kubenswrapper[4978]: I0225 09:44:03.233879 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533544-bjw6s" event={"ID":"677ce558-3f07-465e-b6a2-97126eaea074","Type":"ContainerDied","Data":"9e8674c1a831eddf11614a8d71a99ffcd693a78212eca2b09d49dea60d79075c"} Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.412035 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-stk2k"] Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.416046 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.419799 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.422962 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.424420 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-stk2k"] Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.470514 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-dispersionconf\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.470654 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/55b65695-0237-4eb0-83bc-2c955671206f-etc-swift\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.470706 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-ring-data-devices\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.470764 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.470794 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-scripts\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.470896 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-swiftconf\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.470929 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv56d\" (UniqueName: \"kubernetes.io/projected/55b65695-0237-4eb0-83bc-2c955671206f-kube-api-access-cv56d\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.572977 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.573501 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-scripts\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.573552 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-swiftconf\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.573587 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv56d\" (UniqueName: \"kubernetes.io/projected/55b65695-0237-4eb0-83bc-2c955671206f-kube-api-access-cv56d\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.573672 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-dispersionconf\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.573764 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/55b65695-0237-4eb0-83bc-2c955671206f-etc-swift\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.573817 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-ring-data-devices\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.574196 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/55b65695-0237-4eb0-83bc-2c955671206f-etc-swift\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.577539 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-ring-data-devices\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.577720 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-scripts\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.582305 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-swiftconf\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.592749 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv56d\" (UniqueName: \"kubernetes.io/projected/55b65695-0237-4eb0-83bc-2c955671206f-kube-api-access-cv56d\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.594186 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-dispersionconf\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.594523 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-stk2k\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.747154 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:04 crc kubenswrapper[4978]: I0225 09:44:04.904544 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533544-bjw6s" Feb 25 09:44:05 crc kubenswrapper[4978]: I0225 09:44:04.999956 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rvnz\" (UniqueName: \"kubernetes.io/projected/677ce558-3f07-465e-b6a2-97126eaea074-kube-api-access-7rvnz\") pod \"677ce558-3f07-465e-b6a2-97126eaea074\" (UID: \"677ce558-3f07-465e-b6a2-97126eaea074\") " Feb 25 09:44:05 crc kubenswrapper[4978]: I0225 09:44:05.005159 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/677ce558-3f07-465e-b6a2-97126eaea074-kube-api-access-7rvnz" (OuterVolumeSpecName: "kube-api-access-7rvnz") pod "677ce558-3f07-465e-b6a2-97126eaea074" (UID: "677ce558-3f07-465e-b6a2-97126eaea074"). InnerVolumeSpecName "kube-api-access-7rvnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:44:05 crc kubenswrapper[4978]: I0225 09:44:05.108523 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rvnz\" (UniqueName: \"kubernetes.io/projected/677ce558-3f07-465e-b6a2-97126eaea074-kube-api-access-7rvnz\") on node \"crc\" DevicePath \"\"" Feb 25 09:44:05 crc kubenswrapper[4978]: I0225 09:44:05.274593 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533544-bjw6s" event={"ID":"677ce558-3f07-465e-b6a2-97126eaea074","Type":"ContainerDied","Data":"18ed0b7c418212a8f8dde5da59a8bd957f7696c942050a35fbfc3d5b363a67ce"} Feb 25 09:44:05 crc kubenswrapper[4978]: I0225 09:44:05.275135 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18ed0b7c418212a8f8dde5da59a8bd957f7696c942050a35fbfc3d5b363a67ce" Feb 25 09:44:05 crc kubenswrapper[4978]: I0225 09:44:05.274660 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533544-bjw6s" Feb 25 09:44:05 crc kubenswrapper[4978]: I0225 09:44:05.310335 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-stk2k"] Feb 25 09:44:05 crc kubenswrapper[4978]: W0225 09:44:05.312064 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55b65695_0237_4eb0_83bc_2c955671206f.slice/crio-136ac76af825bc2db87f2ce84394b2a722297d49c9f528f043c18dff5ae970a4 WatchSource:0}: Error finding container 136ac76af825bc2db87f2ce84394b2a722297d49c9f528f043c18dff5ae970a4: Status 404 returned error can't find the container with id 136ac76af825bc2db87f2ce84394b2a722297d49c9f528f043c18dff5ae970a4 Feb 25 09:44:06 crc kubenswrapper[4978]: I0225 09:44:06.003413 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533538-7swrx"] Feb 25 09:44:06 crc kubenswrapper[4978]: I0225 09:44:06.020666 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533538-7swrx"] Feb 25 09:44:06 crc kubenswrapper[4978]: I0225 09:44:06.289618 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-stk2k" event={"ID":"55b65695-0237-4eb0-83bc-2c955671206f","Type":"ContainerStarted","Data":"776daad305ecf8c4c0101a3a22dd8314e950bf51e4dc4967f3092ef6dc9d8237"} Feb 25 09:44:06 crc kubenswrapper[4978]: I0225 09:44:06.289669 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-stk2k" event={"ID":"55b65695-0237-4eb0-83bc-2c955671206f","Type":"ContainerStarted","Data":"136ac76af825bc2db87f2ce84394b2a722297d49c9f528f043c18dff5ae970a4"} Feb 25 09:44:06 crc kubenswrapper[4978]: I0225 09:44:06.308165 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-stk2k" podStartSLOduration=2.308147478 podStartE2EDuration="2.308147478s" podCreationTimestamp="2026-02-25 09:44:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:44:06.306027833 +0000 UTC m=+10739.745284292" watchObservedRunningTime="2026-02-25 09:44:06.308147478 +0000 UTC m=+10739.747403937" Feb 25 09:44:07 crc kubenswrapper[4978]: I0225 09:44:07.341533 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57853581-8f9a-497b-9d05-ab705c0fdfab" path="/var/lib/kubelet/pods/57853581-8f9a-497b-9d05-ab705c0fdfab/volumes" Feb 25 09:44:11 crc kubenswrapper[4978]: I0225 09:44:11.328419 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:44:11 crc kubenswrapper[4978]: E0225 09:44:11.329166 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:44:15 crc kubenswrapper[4978]: I0225 09:44:15.401503 4978 generic.go:334] "Generic (PLEG): container finished" podID="55b65695-0237-4eb0-83bc-2c955671206f" containerID="776daad305ecf8c4c0101a3a22dd8314e950bf51e4dc4967f3092ef6dc9d8237" exitCode=0 Feb 25 09:44:15 crc kubenswrapper[4978]: I0225 09:44:15.401603 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-stk2k" event={"ID":"55b65695-0237-4eb0-83bc-2c955671206f","Type":"ContainerDied","Data":"776daad305ecf8c4c0101a3a22dd8314e950bf51e4dc4967f3092ef6dc9d8237"} Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.369030 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.427836 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-stk2k"] Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.433434 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-stk2k"] Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.434669 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="136ac76af825bc2db87f2ce84394b2a722297d49c9f528f043c18dff5ae970a4" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.434754 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-stk2k" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.535531 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-swiftconf\") pod \"55b65695-0237-4eb0-83bc-2c955671206f\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.535718 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/55b65695-0237-4eb0-83bc-2c955671206f-etc-swift\") pod \"55b65695-0237-4eb0-83bc-2c955671206f\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.535847 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-combined-ca-bundle\") pod \"55b65695-0237-4eb0-83bc-2c955671206f\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.535898 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-ring-data-devices\") pod \"55b65695-0237-4eb0-83bc-2c955671206f\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.536003 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv56d\" (UniqueName: \"kubernetes.io/projected/55b65695-0237-4eb0-83bc-2c955671206f-kube-api-access-cv56d\") pod \"55b65695-0237-4eb0-83bc-2c955671206f\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.536180 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-scripts\") pod \"55b65695-0237-4eb0-83bc-2c955671206f\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.536230 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-dispersionconf\") pod \"55b65695-0237-4eb0-83bc-2c955671206f\" (UID: \"55b65695-0237-4eb0-83bc-2c955671206f\") " Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.536788 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "55b65695-0237-4eb0-83bc-2c955671206f" (UID: "55b65695-0237-4eb0-83bc-2c955671206f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.537287 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55b65695-0237-4eb0-83bc-2c955671206f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "55b65695-0237-4eb0-83bc-2c955671206f" (UID: "55b65695-0237-4eb0-83bc-2c955671206f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.537466 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.548564 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55b65695-0237-4eb0-83bc-2c955671206f-kube-api-access-cv56d" (OuterVolumeSpecName: "kube-api-access-cv56d") pod "55b65695-0237-4eb0-83bc-2c955671206f" (UID: "55b65695-0237-4eb0-83bc-2c955671206f"). InnerVolumeSpecName "kube-api-access-cv56d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.577079 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "55b65695-0237-4eb0-83bc-2c955671206f" (UID: "55b65695-0237-4eb0-83bc-2c955671206f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.584219 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-scripts" (OuterVolumeSpecName: "scripts") pod "55b65695-0237-4eb0-83bc-2c955671206f" (UID: "55b65695-0237-4eb0-83bc-2c955671206f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.584695 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55b65695-0237-4eb0-83bc-2c955671206f" (UID: "55b65695-0237-4eb0-83bc-2c955671206f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.586308 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "55b65695-0237-4eb0-83bc-2c955671206f" (UID: "55b65695-0237-4eb0-83bc-2c955671206f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.652244 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.652284 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/55b65695-0237-4eb0-83bc-2c955671206f-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.652298 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.652315 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv56d\" (UniqueName: \"kubernetes.io/projected/55b65695-0237-4eb0-83bc-2c955671206f-kube-api-access-cv56d\") on node \"crc\" DevicePath \"\"" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.652327 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55b65695-0237-4eb0-83bc-2c955671206f-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:44:17 crc kubenswrapper[4978]: I0225 09:44:17.652338 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/55b65695-0237-4eb0-83bc-2c955671206f-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:44:19 crc kubenswrapper[4978]: I0225 09:44:19.341057 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55b65695-0237-4eb0-83bc-2c955671206f" path="/var/lib/kubelet/pods/55b65695-0237-4eb0-83bc-2c955671206f/volumes" Feb 25 09:44:26 crc kubenswrapper[4978]: I0225 09:44:26.328214 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:44:26 crc kubenswrapper[4978]: E0225 09:44:26.329035 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:44:37 crc kubenswrapper[4978]: I0225 09:44:37.341856 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:44:37 crc kubenswrapper[4978]: E0225 09:44:37.343095 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:44:39 crc kubenswrapper[4978]: I0225 09:44:39.047091 4978 scope.go:117] "RemoveContainer" containerID="4bc469cac38f40815aa233bedb9c6696570cd77ac80bd2387af5171deabb9161" Feb 25 09:44:52 crc kubenswrapper[4978]: I0225 09:44:52.327599 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:44:52 crc kubenswrapper[4978]: E0225 09:44:52.328430 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.159330 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs"] Feb 25 09:45:00 crc kubenswrapper[4978]: E0225 09:45:00.160403 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b65695-0237-4eb0-83bc-2c955671206f" containerName="swift-ring-rebalance" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.160420 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b65695-0237-4eb0-83bc-2c955671206f" containerName="swift-ring-rebalance" Feb 25 09:45:00 crc kubenswrapper[4978]: E0225 09:45:00.160460 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="677ce558-3f07-465e-b6a2-97126eaea074" containerName="oc" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.160467 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="677ce558-3f07-465e-b6a2-97126eaea074" containerName="oc" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.160685 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="55b65695-0237-4eb0-83bc-2c955671206f" containerName="swift-ring-rebalance" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.160699 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="677ce558-3f07-465e-b6a2-97126eaea074" containerName="oc" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.161504 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.164226 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.165606 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.186194 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs"] Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.293841 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64b5569b-078c-45d8-89c9-8a4f32e2aad7-config-volume\") pod \"collect-profiles-29533545-s74zs\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.294529 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64b5569b-078c-45d8-89c9-8a4f32e2aad7-secret-volume\") pod \"collect-profiles-29533545-s74zs\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.294600 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn44k\" (UniqueName: \"kubernetes.io/projected/64b5569b-078c-45d8-89c9-8a4f32e2aad7-kube-api-access-rn44k\") pod \"collect-profiles-29533545-s74zs\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.397116 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64b5569b-078c-45d8-89c9-8a4f32e2aad7-secret-volume\") pod \"collect-profiles-29533545-s74zs\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.397192 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn44k\" (UniqueName: \"kubernetes.io/projected/64b5569b-078c-45d8-89c9-8a4f32e2aad7-kube-api-access-rn44k\") pod \"collect-profiles-29533545-s74zs\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.397467 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64b5569b-078c-45d8-89c9-8a4f32e2aad7-config-volume\") pod \"collect-profiles-29533545-s74zs\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.399493 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64b5569b-078c-45d8-89c9-8a4f32e2aad7-config-volume\") pod \"collect-profiles-29533545-s74zs\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.408258 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64b5569b-078c-45d8-89c9-8a4f32e2aad7-secret-volume\") pod \"collect-profiles-29533545-s74zs\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.414911 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn44k\" (UniqueName: \"kubernetes.io/projected/64b5569b-078c-45d8-89c9-8a4f32e2aad7-kube-api-access-rn44k\") pod \"collect-profiles-29533545-s74zs\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.489207 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:00 crc kubenswrapper[4978]: I0225 09:45:00.985359 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs"] Feb 25 09:45:00 crc kubenswrapper[4978]: W0225 09:45:00.986218 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64b5569b_078c_45d8_89c9_8a4f32e2aad7.slice/crio-569399e32d8e8c28a9d71601024b0520e2ab1dd772d6773864741466611014ca WatchSource:0}: Error finding container 569399e32d8e8c28a9d71601024b0520e2ab1dd772d6773864741466611014ca: Status 404 returned error can't find the container with id 569399e32d8e8c28a9d71601024b0520e2ab1dd772d6773864741466611014ca Feb 25 09:45:01 crc kubenswrapper[4978]: I0225 09:45:01.992704 4978 generic.go:334] "Generic (PLEG): container finished" podID="64b5569b-078c-45d8-89c9-8a4f32e2aad7" containerID="9ea691c0ccee30301c453c41eacc320eaee2233143b7e11d76b1408816e9ca9a" exitCode=0 Feb 25 09:45:01 crc kubenswrapper[4978]: I0225 09:45:01.992830 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" event={"ID":"64b5569b-078c-45d8-89c9-8a4f32e2aad7","Type":"ContainerDied","Data":"9ea691c0ccee30301c453c41eacc320eaee2233143b7e11d76b1408816e9ca9a"} Feb 25 09:45:01 crc kubenswrapper[4978]: I0225 09:45:01.993203 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" event={"ID":"64b5569b-078c-45d8-89c9-8a4f32e2aad7","Type":"ContainerStarted","Data":"569399e32d8e8c28a9d71601024b0520e2ab1dd772d6773864741466611014ca"} Feb 25 09:45:03 crc kubenswrapper[4978]: I0225 09:45:03.523847 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:03 crc kubenswrapper[4978]: I0225 09:45:03.671928 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64b5569b-078c-45d8-89c9-8a4f32e2aad7-config-volume\") pod \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " Feb 25 09:45:03 crc kubenswrapper[4978]: I0225 09:45:03.672082 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64b5569b-078c-45d8-89c9-8a4f32e2aad7-secret-volume\") pod \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " Feb 25 09:45:03 crc kubenswrapper[4978]: I0225 09:45:03.672234 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn44k\" (UniqueName: \"kubernetes.io/projected/64b5569b-078c-45d8-89c9-8a4f32e2aad7-kube-api-access-rn44k\") pod \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\" (UID: \"64b5569b-078c-45d8-89c9-8a4f32e2aad7\") " Feb 25 09:45:03 crc kubenswrapper[4978]: I0225 09:45:03.672804 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64b5569b-078c-45d8-89c9-8a4f32e2aad7-config-volume" (OuterVolumeSpecName: "config-volume") pod "64b5569b-078c-45d8-89c9-8a4f32e2aad7" (UID: "64b5569b-078c-45d8-89c9-8a4f32e2aad7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:45:03 crc kubenswrapper[4978]: I0225 09:45:03.673483 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64b5569b-078c-45d8-89c9-8a4f32e2aad7-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 09:45:03 crc kubenswrapper[4978]: I0225 09:45:03.680136 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64b5569b-078c-45d8-89c9-8a4f32e2aad7-kube-api-access-rn44k" (OuterVolumeSpecName: "kube-api-access-rn44k") pod "64b5569b-078c-45d8-89c9-8a4f32e2aad7" (UID: "64b5569b-078c-45d8-89c9-8a4f32e2aad7"). InnerVolumeSpecName "kube-api-access-rn44k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:45:03 crc kubenswrapper[4978]: I0225 09:45:03.681579 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64b5569b-078c-45d8-89c9-8a4f32e2aad7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "64b5569b-078c-45d8-89c9-8a4f32e2aad7" (UID: "64b5569b-078c-45d8-89c9-8a4f32e2aad7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:45:03 crc kubenswrapper[4978]: I0225 09:45:03.775678 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64b5569b-078c-45d8-89c9-8a4f32e2aad7-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 09:45:03 crc kubenswrapper[4978]: I0225 09:45:03.775711 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn44k\" (UniqueName: \"kubernetes.io/projected/64b5569b-078c-45d8-89c9-8a4f32e2aad7-kube-api-access-rn44k\") on node \"crc\" DevicePath \"\"" Feb 25 09:45:04 crc kubenswrapper[4978]: I0225 09:45:04.020518 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" event={"ID":"64b5569b-078c-45d8-89c9-8a4f32e2aad7","Type":"ContainerDied","Data":"569399e32d8e8c28a9d71601024b0520e2ab1dd772d6773864741466611014ca"} Feb 25 09:45:04 crc kubenswrapper[4978]: I0225 09:45:04.020562 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="569399e32d8e8c28a9d71601024b0520e2ab1dd772d6773864741466611014ca" Feb 25 09:45:04 crc kubenswrapper[4978]: I0225 09:45:04.020627 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs" Feb 25 09:45:04 crc kubenswrapper[4978]: I0225 09:45:04.628889 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2"] Feb 25 09:45:04 crc kubenswrapper[4978]: I0225 09:45:04.647244 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533500-4rsv2"] Feb 25 09:45:05 crc kubenswrapper[4978]: I0225 09:45:05.342765 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d7114d5-5975-480f-acbc-66ffd0aff346" path="/var/lib/kubelet/pods/0d7114d5-5975-480f-acbc-66ffd0aff346/volumes" Feb 25 09:45:06 crc kubenswrapper[4978]: I0225 09:45:06.328108 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:45:06 crc kubenswrapper[4978]: E0225 09:45:06.328742 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.639023 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-nm2qk"] Feb 25 09:45:17 crc kubenswrapper[4978]: E0225 09:45:17.640188 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64b5569b-078c-45d8-89c9-8a4f32e2aad7" containerName="collect-profiles" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.640206 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="64b5569b-078c-45d8-89c9-8a4f32e2aad7" containerName="collect-profiles" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.640562 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="64b5569b-078c-45d8-89c9-8a4f32e2aad7" containerName="collect-profiles" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.641542 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.645034 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.645312 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.660071 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-dispersionconf\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.660154 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/63d9d3a3-abc3-4a67-ad58-c181aada1514-etc-swift\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.660186 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-ring-data-devices\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.660228 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd6z4\" (UniqueName: \"kubernetes.io/projected/63d9d3a3-abc3-4a67-ad58-c181aada1514-kube-api-access-wd6z4\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.660663 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-swiftconf\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.660704 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-scripts\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.660845 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.675487 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-nm2qk"] Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.762787 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.763052 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-dispersionconf\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.763106 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-ring-data-devices\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.763138 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/63d9d3a3-abc3-4a67-ad58-c181aada1514-etc-swift\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.763180 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd6z4\" (UniqueName: \"kubernetes.io/projected/63d9d3a3-abc3-4a67-ad58-c181aada1514-kube-api-access-wd6z4\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.763240 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-swiftconf\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.763273 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-scripts\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.763686 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/63d9d3a3-abc3-4a67-ad58-c181aada1514-etc-swift\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.765431 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-scripts\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.765491 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-ring-data-devices\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.772207 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-dispersionconf\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.772394 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.772807 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-swiftconf\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.780272 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd6z4\" (UniqueName: \"kubernetes.io/projected/63d9d3a3-abc3-4a67-ad58-c181aada1514-kube-api-access-wd6z4\") pod \"swift-ring-rebalance-debug-nm2qk\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:17 crc kubenswrapper[4978]: I0225 09:45:17.965569 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:18 crc kubenswrapper[4978]: I0225 09:45:18.327947 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:45:18 crc kubenswrapper[4978]: E0225 09:45:18.328720 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:45:18 crc kubenswrapper[4978]: I0225 09:45:18.524131 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-nm2qk"] Feb 25 09:45:19 crc kubenswrapper[4978]: I0225 09:45:19.234519 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nm2qk" event={"ID":"63d9d3a3-abc3-4a67-ad58-c181aada1514","Type":"ContainerStarted","Data":"35bb28f592ad20de882ef44283459b18da4aa95778cdb97f669674f957e9791d"} Feb 25 09:45:19 crc kubenswrapper[4978]: I0225 09:45:19.235144 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nm2qk" event={"ID":"63d9d3a3-abc3-4a67-ad58-c181aada1514","Type":"ContainerStarted","Data":"6908a6f61e710fada66b2913317fc65956d77409551ce21172f6a02a91c3f580"} Feb 25 09:45:19 crc kubenswrapper[4978]: I0225 09:45:19.261681 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-nm2qk" podStartSLOduration=2.261663212 podStartE2EDuration="2.261663212s" podCreationTimestamp="2026-02-25 09:45:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:45:19.254405836 +0000 UTC m=+10812.693662325" watchObservedRunningTime="2026-02-25 09:45:19.261663212 +0000 UTC m=+10812.700919681" Feb 25 09:45:28 crc kubenswrapper[4978]: I0225 09:45:28.339097 4978 generic.go:334] "Generic (PLEG): container finished" podID="63d9d3a3-abc3-4a67-ad58-c181aada1514" containerID="35bb28f592ad20de882ef44283459b18da4aa95778cdb97f669674f957e9791d" exitCode=0 Feb 25 09:45:28 crc kubenswrapper[4978]: I0225 09:45:28.339152 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nm2qk" event={"ID":"63d9d3a3-abc3-4a67-ad58-c181aada1514","Type":"ContainerDied","Data":"35bb28f592ad20de882ef44283459b18da4aa95778cdb97f669674f957e9791d"} Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.304958 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.357818 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-nm2qk"] Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.360099 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nm2qk" event={"ID":"63d9d3a3-abc3-4a67-ad58-c181aada1514","Type":"ContainerDied","Data":"6908a6f61e710fada66b2913317fc65956d77409551ce21172f6a02a91c3f580"} Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.360128 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6908a6f61e710fada66b2913317fc65956d77409551ce21172f6a02a91c3f580" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.360177 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nm2qk" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.367872 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-nm2qk"] Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.486086 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-dispersionconf\") pod \"63d9d3a3-abc3-4a67-ad58-c181aada1514\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.486221 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-scripts\") pod \"63d9d3a3-abc3-4a67-ad58-c181aada1514\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.486262 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd6z4\" (UniqueName: \"kubernetes.io/projected/63d9d3a3-abc3-4a67-ad58-c181aada1514-kube-api-access-wd6z4\") pod \"63d9d3a3-abc3-4a67-ad58-c181aada1514\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.486312 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-combined-ca-bundle\") pod \"63d9d3a3-abc3-4a67-ad58-c181aada1514\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.486398 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/63d9d3a3-abc3-4a67-ad58-c181aada1514-etc-swift\") pod \"63d9d3a3-abc3-4a67-ad58-c181aada1514\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.486455 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-ring-data-devices\") pod \"63d9d3a3-abc3-4a67-ad58-c181aada1514\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.486583 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-swiftconf\") pod \"63d9d3a3-abc3-4a67-ad58-c181aada1514\" (UID: \"63d9d3a3-abc3-4a67-ad58-c181aada1514\") " Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.487090 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "63d9d3a3-abc3-4a67-ad58-c181aada1514" (UID: "63d9d3a3-abc3-4a67-ad58-c181aada1514"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.487478 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63d9d3a3-abc3-4a67-ad58-c181aada1514-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "63d9d3a3-abc3-4a67-ad58-c181aada1514" (UID: "63d9d3a3-abc3-4a67-ad58-c181aada1514"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.487778 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/63d9d3a3-abc3-4a67-ad58-c181aada1514-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.487925 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.493672 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63d9d3a3-abc3-4a67-ad58-c181aada1514-kube-api-access-wd6z4" (OuterVolumeSpecName: "kube-api-access-wd6z4") pod "63d9d3a3-abc3-4a67-ad58-c181aada1514" (UID: "63d9d3a3-abc3-4a67-ad58-c181aada1514"). InnerVolumeSpecName "kube-api-access-wd6z4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.518718 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "63d9d3a3-abc3-4a67-ad58-c181aada1514" (UID: "63d9d3a3-abc3-4a67-ad58-c181aada1514"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.523613 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "63d9d3a3-abc3-4a67-ad58-c181aada1514" (UID: "63d9d3a3-abc3-4a67-ad58-c181aada1514"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.529449 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-scripts" (OuterVolumeSpecName: "scripts") pod "63d9d3a3-abc3-4a67-ad58-c181aada1514" (UID: "63d9d3a3-abc3-4a67-ad58-c181aada1514"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.533939 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63d9d3a3-abc3-4a67-ad58-c181aada1514" (UID: "63d9d3a3-abc3-4a67-ad58-c181aada1514"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.590796 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.590827 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63d9d3a3-abc3-4a67-ad58-c181aada1514-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.590875 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd6z4\" (UniqueName: \"kubernetes.io/projected/63d9d3a3-abc3-4a67-ad58-c181aada1514-kube-api-access-wd6z4\") on node \"crc\" DevicePath \"\"" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.590887 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:45:30 crc kubenswrapper[4978]: I0225 09:45:30.590896 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/63d9d3a3-abc3-4a67-ad58-c181aada1514-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:45:30 crc kubenswrapper[4978]: E0225 09:45:30.869080 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63d9d3a3_abc3_4a67_ad58_c181aada1514.slice\": RecentStats: unable to find data in memory cache]" Feb 25 09:45:31 crc kubenswrapper[4978]: I0225 09:45:31.341148 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63d9d3a3-abc3-4a67-ad58-c181aada1514" path="/var/lib/kubelet/pods/63d9d3a3-abc3-4a67-ad58-c181aada1514/volumes" Feb 25 09:45:33 crc kubenswrapper[4978]: I0225 09:45:33.327861 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:45:33 crc kubenswrapper[4978]: E0225 09:45:33.328556 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:45:39 crc kubenswrapper[4978]: I0225 09:45:39.177114 4978 scope.go:117] "RemoveContainer" containerID="643169a3c71c7a40fb55504469b404bbde60718bc95d306ab88068f5a5127cd4" Feb 25 09:45:46 crc kubenswrapper[4978]: I0225 09:45:46.328545 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:45:46 crc kubenswrapper[4978]: E0225 09:45:46.329154 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:45:58 crc kubenswrapper[4978]: I0225 09:45:58.328296 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:45:58 crc kubenswrapper[4978]: I0225 09:45:58.719483 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"deecfe5b697e078cff808a3eda1a38a3bf48b1db8ad7591055a352645c854593"} Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.177955 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533546-6d24g"] Feb 25 09:46:00 crc kubenswrapper[4978]: E0225 09:46:00.178945 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d9d3a3-abc3-4a67-ad58-c181aada1514" containerName="swift-ring-rebalance" Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.178970 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d9d3a3-abc3-4a67-ad58-c181aada1514" containerName="swift-ring-rebalance" Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.179314 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d9d3a3-abc3-4a67-ad58-c181aada1514" containerName="swift-ring-rebalance" Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.180459 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533546-6d24g" Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.189489 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.190612 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.219966 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.243195 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533546-6d24g"] Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.280195 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bmqx\" (UniqueName: \"kubernetes.io/projected/341213eb-7544-4b05-883b-f4925bc39f03-kube-api-access-6bmqx\") pod \"auto-csr-approver-29533546-6d24g\" (UID: \"341213eb-7544-4b05-883b-f4925bc39f03\") " pod="openshift-infra/auto-csr-approver-29533546-6d24g" Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.382133 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bmqx\" (UniqueName: \"kubernetes.io/projected/341213eb-7544-4b05-883b-f4925bc39f03-kube-api-access-6bmqx\") pod \"auto-csr-approver-29533546-6d24g\" (UID: \"341213eb-7544-4b05-883b-f4925bc39f03\") " pod="openshift-infra/auto-csr-approver-29533546-6d24g" Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.596075 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bmqx\" (UniqueName: \"kubernetes.io/projected/341213eb-7544-4b05-883b-f4925bc39f03-kube-api-access-6bmqx\") pod \"auto-csr-approver-29533546-6d24g\" (UID: \"341213eb-7544-4b05-883b-f4925bc39f03\") " pod="openshift-infra/auto-csr-approver-29533546-6d24g" Feb 25 09:46:00 crc kubenswrapper[4978]: I0225 09:46:00.842901 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533546-6d24g" Feb 25 09:46:01 crc kubenswrapper[4978]: I0225 09:46:01.370503 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533546-6d24g"] Feb 25 09:46:01 crc kubenswrapper[4978]: I0225 09:46:01.373998 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:46:01 crc kubenswrapper[4978]: I0225 09:46:01.781789 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533546-6d24g" event={"ID":"341213eb-7544-4b05-883b-f4925bc39f03","Type":"ContainerStarted","Data":"90aa1a469b8c0c2cc49f5c450ed413a4ed253e36de3269cb9803addb5f8f647a"} Feb 25 09:46:03 crc kubenswrapper[4978]: I0225 09:46:03.812109 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533546-6d24g" event={"ID":"341213eb-7544-4b05-883b-f4925bc39f03","Type":"ContainerStarted","Data":"7ddeb26c059979325ce291599ee23a29779593d64790c534c75a9444acfdb540"} Feb 25 09:46:03 crc kubenswrapper[4978]: I0225 09:46:03.830031 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533546-6d24g" podStartSLOduration=2.682983773 podStartE2EDuration="3.830007537s" podCreationTimestamp="2026-02-25 09:46:00 +0000 UTC" firstStartedPulling="2026-02-25 09:46:01.373758142 +0000 UTC m=+10854.813014601" lastFinishedPulling="2026-02-25 09:46:02.520781896 +0000 UTC m=+10855.960038365" observedRunningTime="2026-02-25 09:46:03.827923482 +0000 UTC m=+10857.267179941" watchObservedRunningTime="2026-02-25 09:46:03.830007537 +0000 UTC m=+10857.269264006" Feb 25 09:46:04 crc kubenswrapper[4978]: I0225 09:46:04.832836 4978 generic.go:334] "Generic (PLEG): container finished" podID="341213eb-7544-4b05-883b-f4925bc39f03" containerID="7ddeb26c059979325ce291599ee23a29779593d64790c534c75a9444acfdb540" exitCode=0 Feb 25 09:46:04 crc kubenswrapper[4978]: I0225 09:46:04.832937 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533546-6d24g" event={"ID":"341213eb-7544-4b05-883b-f4925bc39f03","Type":"ContainerDied","Data":"7ddeb26c059979325ce291599ee23a29779593d64790c534c75a9444acfdb540"} Feb 25 09:46:06 crc kubenswrapper[4978]: I0225 09:46:06.707352 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533546-6d24g" Feb 25 09:46:06 crc kubenswrapper[4978]: I0225 09:46:06.838136 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bmqx\" (UniqueName: \"kubernetes.io/projected/341213eb-7544-4b05-883b-f4925bc39f03-kube-api-access-6bmqx\") pod \"341213eb-7544-4b05-883b-f4925bc39f03\" (UID: \"341213eb-7544-4b05-883b-f4925bc39f03\") " Feb 25 09:46:06 crc kubenswrapper[4978]: I0225 09:46:06.848127 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/341213eb-7544-4b05-883b-f4925bc39f03-kube-api-access-6bmqx" (OuterVolumeSpecName: "kube-api-access-6bmqx") pod "341213eb-7544-4b05-883b-f4925bc39f03" (UID: "341213eb-7544-4b05-883b-f4925bc39f03"). InnerVolumeSpecName "kube-api-access-6bmqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:46:06 crc kubenswrapper[4978]: I0225 09:46:06.872108 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533546-6d24g" event={"ID":"341213eb-7544-4b05-883b-f4925bc39f03","Type":"ContainerDied","Data":"90aa1a469b8c0c2cc49f5c450ed413a4ed253e36de3269cb9803addb5f8f647a"} Feb 25 09:46:06 crc kubenswrapper[4978]: I0225 09:46:06.872161 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90aa1a469b8c0c2cc49f5c450ed413a4ed253e36de3269cb9803addb5f8f647a" Feb 25 09:46:06 crc kubenswrapper[4978]: I0225 09:46:06.872231 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533546-6d24g" Feb 25 09:46:06 crc kubenswrapper[4978]: I0225 09:46:06.939726 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533540-w295f"] Feb 25 09:46:06 crc kubenswrapper[4978]: I0225 09:46:06.941826 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bmqx\" (UniqueName: \"kubernetes.io/projected/341213eb-7544-4b05-883b-f4925bc39f03-kube-api-access-6bmqx\") on node \"crc\" DevicePath \"\"" Feb 25 09:46:06 crc kubenswrapper[4978]: I0225 09:46:06.949803 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533540-w295f"] Feb 25 09:46:07 crc kubenswrapper[4978]: I0225 09:46:07.349039 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78ee54e7-3195-40bb-9aaf-540a08cb620c" path="/var/lib/kubelet/pods/78ee54e7-3195-40bb-9aaf-540a08cb620c/volumes" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.543008 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-2lvmk"] Feb 25 09:46:30 crc kubenswrapper[4978]: E0225 09:46:30.544855 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341213eb-7544-4b05-883b-f4925bc39f03" containerName="oc" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.544879 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="341213eb-7544-4b05-883b-f4925bc39f03" containerName="oc" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.545202 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="341213eb-7544-4b05-883b-f4925bc39f03" containerName="oc" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.546499 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.548675 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.549693 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.590453 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-2lvmk"] Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.618928 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-swiftconf\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.619035 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-dispersionconf\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.619093 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.619114 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fbsm\" (UniqueName: \"kubernetes.io/projected/1d355df9-898c-4de1-a211-4552c4c97fad-kube-api-access-5fbsm\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.619136 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-scripts\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.619176 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d355df9-898c-4de1-a211-4552c4c97fad-etc-swift\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.619215 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-ring-data-devices\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.721211 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-dispersionconf\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.721306 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.721336 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fbsm\" (UniqueName: \"kubernetes.io/projected/1d355df9-898c-4de1-a211-4552c4c97fad-kube-api-access-5fbsm\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.721361 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-scripts\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.721419 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d355df9-898c-4de1-a211-4552c4c97fad-etc-swift\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.721463 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-ring-data-devices\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.721483 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-swiftconf\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.722758 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d355df9-898c-4de1-a211-4552c4c97fad-etc-swift\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.723975 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-scripts\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.724208 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-ring-data-devices\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.730412 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-dispersionconf\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.730710 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.737728 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-swiftconf\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.751847 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fbsm\" (UniqueName: \"kubernetes.io/projected/1d355df9-898c-4de1-a211-4552c4c97fad-kube-api-access-5fbsm\") pod \"swift-ring-rebalance-debug-2lvmk\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:30 crc kubenswrapper[4978]: I0225 09:46:30.873127 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:31 crc kubenswrapper[4978]: I0225 09:46:31.409495 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-2lvmk"] Feb 25 09:46:32 crc kubenswrapper[4978]: I0225 09:46:32.220345 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2lvmk" event={"ID":"1d355df9-898c-4de1-a211-4552c4c97fad","Type":"ContainerStarted","Data":"c9a54f53cc8cab85d14c843ac080e82b9a21b029c5684a7a36e7328aca364e63"} Feb 25 09:46:32 crc kubenswrapper[4978]: I0225 09:46:32.221499 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2lvmk" event={"ID":"1d355df9-898c-4de1-a211-4552c4c97fad","Type":"ContainerStarted","Data":"d4b69da8fec7604658003b8a2ec993684fec83f642e5f80aabb017da9e9d43dd"} Feb 25 09:46:32 crc kubenswrapper[4978]: I0225 09:46:32.250854 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-2lvmk" podStartSLOduration=2.250835223 podStartE2EDuration="2.250835223s" podCreationTimestamp="2026-02-25 09:46:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:46:32.241238824 +0000 UTC m=+10885.680495273" watchObservedRunningTime="2026-02-25 09:46:32.250835223 +0000 UTC m=+10885.690091682" Feb 25 09:46:39 crc kubenswrapper[4978]: I0225 09:46:39.444817 4978 scope.go:117] "RemoveContainer" containerID="698f8d7ffcf03a62b5b370372410e46e26337b4089f2d94c5f15bf70f5be019f" Feb 25 09:46:41 crc kubenswrapper[4978]: I0225 09:46:41.312594 4978 generic.go:334] "Generic (PLEG): container finished" podID="1d355df9-898c-4de1-a211-4552c4c97fad" containerID="c9a54f53cc8cab85d14c843ac080e82b9a21b029c5684a7a36e7328aca364e63" exitCode=0 Feb 25 09:46:41 crc kubenswrapper[4978]: I0225 09:46:41.312827 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2lvmk" event={"ID":"1d355df9-898c-4de1-a211-4552c4c97fad","Type":"ContainerDied","Data":"c9a54f53cc8cab85d14c843ac080e82b9a21b029c5684a7a36e7328aca364e63"} Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.314735 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.342674 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2lvmk" event={"ID":"1d355df9-898c-4de1-a211-4552c4c97fad","Type":"ContainerDied","Data":"d4b69da8fec7604658003b8a2ec993684fec83f642e5f80aabb017da9e9d43dd"} Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.342716 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4b69da8fec7604658003b8a2ec993684fec83f642e5f80aabb017da9e9d43dd" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.342749 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2lvmk" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.382492 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-2lvmk"] Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.394760 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-2lvmk"] Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.465507 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d355df9-898c-4de1-a211-4552c4c97fad-etc-swift\") pod \"1d355df9-898c-4de1-a211-4552c4c97fad\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.465573 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-ring-data-devices\") pod \"1d355df9-898c-4de1-a211-4552c4c97fad\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.465611 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fbsm\" (UniqueName: \"kubernetes.io/projected/1d355df9-898c-4de1-a211-4552c4c97fad-kube-api-access-5fbsm\") pod \"1d355df9-898c-4de1-a211-4552c4c97fad\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.465673 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-dispersionconf\") pod \"1d355df9-898c-4de1-a211-4552c4c97fad\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.465728 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-swiftconf\") pod \"1d355df9-898c-4de1-a211-4552c4c97fad\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.465822 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-scripts\") pod \"1d355df9-898c-4de1-a211-4552c4c97fad\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.465878 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-combined-ca-bundle\") pod \"1d355df9-898c-4de1-a211-4552c4c97fad\" (UID: \"1d355df9-898c-4de1-a211-4552c4c97fad\") " Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.467363 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d355df9-898c-4de1-a211-4552c4c97fad-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1d355df9-898c-4de1-a211-4552c4c97fad" (UID: "1d355df9-898c-4de1-a211-4552c4c97fad"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.467544 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1d355df9-898c-4de1-a211-4552c4c97fad" (UID: "1d355df9-898c-4de1-a211-4552c4c97fad"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.486121 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d355df9-898c-4de1-a211-4552c4c97fad-kube-api-access-5fbsm" (OuterVolumeSpecName: "kube-api-access-5fbsm") pod "1d355df9-898c-4de1-a211-4552c4c97fad" (UID: "1d355df9-898c-4de1-a211-4552c4c97fad"). InnerVolumeSpecName "kube-api-access-5fbsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.498695 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1d355df9-898c-4de1-a211-4552c4c97fad" (UID: "1d355df9-898c-4de1-a211-4552c4c97fad"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.500559 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1d355df9-898c-4de1-a211-4552c4c97fad" (UID: "1d355df9-898c-4de1-a211-4552c4c97fad"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.504587 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d355df9-898c-4de1-a211-4552c4c97fad" (UID: "1d355df9-898c-4de1-a211-4552c4c97fad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.521414 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-scripts" (OuterVolumeSpecName: "scripts") pod "1d355df9-898c-4de1-a211-4552c4c97fad" (UID: "1d355df9-898c-4de1-a211-4552c4c97fad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.568465 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.568501 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.568512 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d355df9-898c-4de1-a211-4552c4c97fad-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.568521 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d355df9-898c-4de1-a211-4552c4c97fad-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.568532 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fbsm\" (UniqueName: \"kubernetes.io/projected/1d355df9-898c-4de1-a211-4552c4c97fad-kube-api-access-5fbsm\") on node \"crc\" DevicePath \"\"" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.568543 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:46:44 crc kubenswrapper[4978]: I0225 09:46:44.568551 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d355df9-898c-4de1-a211-4552c4c97fad-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:46:45 crc kubenswrapper[4978]: I0225 09:46:45.368520 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d355df9-898c-4de1-a211-4552c4c97fad" path="/var/lib/kubelet/pods/1d355df9-898c-4de1-a211-4552c4c97fad/volumes" Feb 25 09:47:39 crc kubenswrapper[4978]: I0225 09:47:39.517037 4978 scope.go:117] "RemoveContainer" containerID="6fddf5c93eac70ec068e60efec3427719a776324a47e721526991fdac65b9c50" Feb 25 09:47:39 crc kubenswrapper[4978]: I0225 09:47:39.578150 4978 scope.go:117] "RemoveContainer" containerID="e0474d67516f6bc53643dffa1a1dff5c9d424cac67f6bd0f454320eb14ac1a48" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.599864 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-7vwlp"] Feb 25 09:47:44 crc kubenswrapper[4978]: E0225 09:47:44.601091 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d355df9-898c-4de1-a211-4552c4c97fad" containerName="swift-ring-rebalance" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.601113 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d355df9-898c-4de1-a211-4552c4c97fad" containerName="swift-ring-rebalance" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.601481 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d355df9-898c-4de1-a211-4552c4c97fad" containerName="swift-ring-rebalance" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.602817 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.605075 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.605592 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.626542 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-7vwlp"] Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.705831 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-scripts\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.706138 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-ring-data-devices\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.706176 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d53688c-8fe4-4acf-96ba-90befe280c8a-etc-swift\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.706470 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sgm9\" (UniqueName: \"kubernetes.io/projected/2d53688c-8fe4-4acf-96ba-90befe280c8a-kube-api-access-6sgm9\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.706611 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-swiftconf\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.706746 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-dispersionconf\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.706901 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.809061 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-dispersionconf\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.809153 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.809191 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-scripts\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.809217 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-ring-data-devices\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.809249 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d53688c-8fe4-4acf-96ba-90befe280c8a-etc-swift\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.809335 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sgm9\" (UniqueName: \"kubernetes.io/projected/2d53688c-8fe4-4acf-96ba-90befe280c8a-kube-api-access-6sgm9\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.809393 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-swiftconf\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.809978 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-ring-data-devices\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.810210 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-scripts\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.810301 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d53688c-8fe4-4acf-96ba-90befe280c8a-etc-swift\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.816165 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-dispersionconf\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.816632 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.818442 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-swiftconf\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.832999 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sgm9\" (UniqueName: \"kubernetes.io/projected/2d53688c-8fe4-4acf-96ba-90befe280c8a-kube-api-access-6sgm9\") pod \"swift-ring-rebalance-debug-7vwlp\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:44 crc kubenswrapper[4978]: I0225 09:47:44.932244 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:45 crc kubenswrapper[4978]: I0225 09:47:45.411890 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-7vwlp"] Feb 25 09:47:45 crc kubenswrapper[4978]: I0225 09:47:45.469295 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-7vwlp" event={"ID":"2d53688c-8fe4-4acf-96ba-90befe280c8a","Type":"ContainerStarted","Data":"bd1fb1cb0c7dbea2cabdeaeaf1d53b7bce06811c72cb1c842d448fd4d14555a3"} Feb 25 09:47:46 crc kubenswrapper[4978]: I0225 09:47:46.506104 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-7vwlp" event={"ID":"2d53688c-8fe4-4acf-96ba-90befe280c8a","Type":"ContainerStarted","Data":"65a1631b6d75383c8585ac91c4bfaf91a59ff77ae5576d8ae577dd3a69e5d296"} Feb 25 09:47:46 crc kubenswrapper[4978]: I0225 09:47:46.537142 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-7vwlp" podStartSLOduration=2.5371121199999997 podStartE2EDuration="2.53711212s" podCreationTimestamp="2026-02-25 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:47:46.527335137 +0000 UTC m=+10959.966591626" watchObservedRunningTime="2026-02-25 09:47:46.53711212 +0000 UTC m=+10959.976368599" Feb 25 09:47:55 crc kubenswrapper[4978]: E0225 09:47:55.050345 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d53688c_8fe4_4acf_96ba_90befe280c8a.slice/crio-conmon-65a1631b6d75383c8585ac91c4bfaf91a59ff77ae5576d8ae577dd3a69e5d296.scope\": RecentStats: unable to find data in memory cache]" Feb 25 09:47:55 crc kubenswrapper[4978]: I0225 09:47:55.616044 4978 generic.go:334] "Generic (PLEG): container finished" podID="2d53688c-8fe4-4acf-96ba-90befe280c8a" containerID="65a1631b6d75383c8585ac91c4bfaf91a59ff77ae5576d8ae577dd3a69e5d296" exitCode=0 Feb 25 09:47:55 crc kubenswrapper[4978]: I0225 09:47:55.616098 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-7vwlp" event={"ID":"2d53688c-8fe4-4acf-96ba-90befe280c8a","Type":"ContainerDied","Data":"65a1631b6d75383c8585ac91c4bfaf91a59ff77ae5576d8ae577dd3a69e5d296"} Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.638316 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-7vwlp" event={"ID":"2d53688c-8fe4-4acf-96ba-90befe280c8a","Type":"ContainerDied","Data":"bd1fb1cb0c7dbea2cabdeaeaf1d53b7bce06811c72cb1c842d448fd4d14555a3"} Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.638739 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd1fb1cb0c7dbea2cabdeaeaf1d53b7bce06811c72cb1c842d448fd4d14555a3" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.725853 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.807323 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-7vwlp"] Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.825697 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-7vwlp"] Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.863358 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-scripts\") pod \"2d53688c-8fe4-4acf-96ba-90befe280c8a\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.863600 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d53688c-8fe4-4acf-96ba-90befe280c8a-etc-swift\") pod \"2d53688c-8fe4-4acf-96ba-90befe280c8a\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.863663 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-ring-data-devices\") pod \"2d53688c-8fe4-4acf-96ba-90befe280c8a\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.863736 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-combined-ca-bundle\") pod \"2d53688c-8fe4-4acf-96ba-90befe280c8a\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.863786 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-swiftconf\") pod \"2d53688c-8fe4-4acf-96ba-90befe280c8a\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.863854 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-dispersionconf\") pod \"2d53688c-8fe4-4acf-96ba-90befe280c8a\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.863886 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sgm9\" (UniqueName: \"kubernetes.io/projected/2d53688c-8fe4-4acf-96ba-90befe280c8a-kube-api-access-6sgm9\") pod \"2d53688c-8fe4-4acf-96ba-90befe280c8a\" (UID: \"2d53688c-8fe4-4acf-96ba-90befe280c8a\") " Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.865536 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2d53688c-8fe4-4acf-96ba-90befe280c8a" (UID: "2d53688c-8fe4-4acf-96ba-90befe280c8a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.867132 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d53688c-8fe4-4acf-96ba-90befe280c8a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2d53688c-8fe4-4acf-96ba-90befe280c8a" (UID: "2d53688c-8fe4-4acf-96ba-90befe280c8a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.869838 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d53688c-8fe4-4acf-96ba-90befe280c8a-kube-api-access-6sgm9" (OuterVolumeSpecName: "kube-api-access-6sgm9") pod "2d53688c-8fe4-4acf-96ba-90befe280c8a" (UID: "2d53688c-8fe4-4acf-96ba-90befe280c8a"). InnerVolumeSpecName "kube-api-access-6sgm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.893991 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-scripts" (OuterVolumeSpecName: "scripts") pod "2d53688c-8fe4-4acf-96ba-90befe280c8a" (UID: "2d53688c-8fe4-4acf-96ba-90befe280c8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.901354 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d53688c-8fe4-4acf-96ba-90befe280c8a" (UID: "2d53688c-8fe4-4acf-96ba-90befe280c8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.913577 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2d53688c-8fe4-4acf-96ba-90befe280c8a" (UID: "2d53688c-8fe4-4acf-96ba-90befe280c8a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.920869 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2d53688c-8fe4-4acf-96ba-90befe280c8a" (UID: "2d53688c-8fe4-4acf-96ba-90befe280c8a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.966370 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.966409 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.966419 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d53688c-8fe4-4acf-96ba-90befe280c8a-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.966429 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sgm9\" (UniqueName: \"kubernetes.io/projected/2d53688c-8fe4-4acf-96ba-90befe280c8a-kube-api-access-6sgm9\") on node \"crc\" DevicePath \"\"" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.966438 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.966446 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d53688c-8fe4-4acf-96ba-90befe280c8a-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:47:57 crc kubenswrapper[4978]: I0225 09:47:57.966454 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d53688c-8fe4-4acf-96ba-90befe280c8a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:47:58 crc kubenswrapper[4978]: I0225 09:47:58.652743 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-7vwlp" Feb 25 09:47:59 crc kubenswrapper[4978]: I0225 09:47:59.344363 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d53688c-8fe4-4acf-96ba-90befe280c8a" path="/var/lib/kubelet/pods/2d53688c-8fe4-4acf-96ba-90befe280c8a/volumes" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.176931 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533548-4khrw"] Feb 25 09:48:00 crc kubenswrapper[4978]: E0225 09:48:00.177761 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d53688c-8fe4-4acf-96ba-90befe280c8a" containerName="swift-ring-rebalance" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.177776 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d53688c-8fe4-4acf-96ba-90befe280c8a" containerName="swift-ring-rebalance" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.178094 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d53688c-8fe4-4acf-96ba-90befe280c8a" containerName="swift-ring-rebalance" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.178938 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533548-4khrw" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.182212 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.183235 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.183681 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.190196 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533548-4khrw"] Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.324044 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f658z\" (UniqueName: \"kubernetes.io/projected/bd80fd9a-eb0d-4555-8822-a34bb6085184-kube-api-access-f658z\") pod \"auto-csr-approver-29533548-4khrw\" (UID: \"bd80fd9a-eb0d-4555-8822-a34bb6085184\") " pod="openshift-infra/auto-csr-approver-29533548-4khrw" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.427347 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f658z\" (UniqueName: \"kubernetes.io/projected/bd80fd9a-eb0d-4555-8822-a34bb6085184-kube-api-access-f658z\") pod \"auto-csr-approver-29533548-4khrw\" (UID: \"bd80fd9a-eb0d-4555-8822-a34bb6085184\") " pod="openshift-infra/auto-csr-approver-29533548-4khrw" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.448163 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f658z\" (UniqueName: \"kubernetes.io/projected/bd80fd9a-eb0d-4555-8822-a34bb6085184-kube-api-access-f658z\") pod \"auto-csr-approver-29533548-4khrw\" (UID: \"bd80fd9a-eb0d-4555-8822-a34bb6085184\") " pod="openshift-infra/auto-csr-approver-29533548-4khrw" Feb 25 09:48:00 crc kubenswrapper[4978]: I0225 09:48:00.506970 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533548-4khrw" Feb 25 09:48:01 crc kubenswrapper[4978]: I0225 09:48:01.007185 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533548-4khrw"] Feb 25 09:48:01 crc kubenswrapper[4978]: I0225 09:48:01.724747 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533548-4khrw" event={"ID":"bd80fd9a-eb0d-4555-8822-a34bb6085184","Type":"ContainerStarted","Data":"53666a9eef3894a117b089fa6e1764d23d516a126458da9643912314906df63c"} Feb 25 09:48:02 crc kubenswrapper[4978]: I0225 09:48:02.736029 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533548-4khrw" event={"ID":"bd80fd9a-eb0d-4555-8822-a34bb6085184","Type":"ContainerStarted","Data":"3381cf5e13bb15501678cb41908eb08881530c85c268e5df66b8afc6c9bb5933"} Feb 25 09:48:02 crc kubenswrapper[4978]: I0225 09:48:02.757796 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533548-4khrw" podStartSLOduration=1.8180070910000001 podStartE2EDuration="2.757773616s" podCreationTimestamp="2026-02-25 09:48:00 +0000 UTC" firstStartedPulling="2026-02-25 09:48:01.018731579 +0000 UTC m=+10974.457988038" lastFinishedPulling="2026-02-25 09:48:01.958498094 +0000 UTC m=+10975.397754563" observedRunningTime="2026-02-25 09:48:02.748173058 +0000 UTC m=+10976.187429527" watchObservedRunningTime="2026-02-25 09:48:02.757773616 +0000 UTC m=+10976.197030085" Feb 25 09:48:03 crc kubenswrapper[4978]: I0225 09:48:03.750969 4978 generic.go:334] "Generic (PLEG): container finished" podID="bd80fd9a-eb0d-4555-8822-a34bb6085184" containerID="3381cf5e13bb15501678cb41908eb08881530c85c268e5df66b8afc6c9bb5933" exitCode=0 Feb 25 09:48:03 crc kubenswrapper[4978]: I0225 09:48:03.751035 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533548-4khrw" event={"ID":"bd80fd9a-eb0d-4555-8822-a34bb6085184","Type":"ContainerDied","Data":"3381cf5e13bb15501678cb41908eb08881530c85c268e5df66b8afc6c9bb5933"} Feb 25 09:48:05 crc kubenswrapper[4978]: I0225 09:48:05.668853 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533548-4khrw" Feb 25 09:48:05 crc kubenswrapper[4978]: I0225 09:48:05.757817 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f658z\" (UniqueName: \"kubernetes.io/projected/bd80fd9a-eb0d-4555-8822-a34bb6085184-kube-api-access-f658z\") pod \"bd80fd9a-eb0d-4555-8822-a34bb6085184\" (UID: \"bd80fd9a-eb0d-4555-8822-a34bb6085184\") " Feb 25 09:48:05 crc kubenswrapper[4978]: I0225 09:48:05.788391 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd80fd9a-eb0d-4555-8822-a34bb6085184-kube-api-access-f658z" (OuterVolumeSpecName: "kube-api-access-f658z") pod "bd80fd9a-eb0d-4555-8822-a34bb6085184" (UID: "bd80fd9a-eb0d-4555-8822-a34bb6085184"). InnerVolumeSpecName "kube-api-access-f658z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:48:05 crc kubenswrapper[4978]: I0225 09:48:05.802765 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533548-4khrw" event={"ID":"bd80fd9a-eb0d-4555-8822-a34bb6085184","Type":"ContainerDied","Data":"53666a9eef3894a117b089fa6e1764d23d516a126458da9643912314906df63c"} Feb 25 09:48:05 crc kubenswrapper[4978]: I0225 09:48:05.802805 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53666a9eef3894a117b089fa6e1764d23d516a126458da9643912314906df63c" Feb 25 09:48:05 crc kubenswrapper[4978]: I0225 09:48:05.802853 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533548-4khrw" Feb 25 09:48:05 crc kubenswrapper[4978]: I0225 09:48:05.833012 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533542-6w6zb"] Feb 25 09:48:05 crc kubenswrapper[4978]: I0225 09:48:05.870613 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f658z\" (UniqueName: \"kubernetes.io/projected/bd80fd9a-eb0d-4555-8822-a34bb6085184-kube-api-access-f658z\") on node \"crc\" DevicePath \"\"" Feb 25 09:48:05 crc kubenswrapper[4978]: I0225 09:48:05.874839 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533542-6w6zb"] Feb 25 09:48:07 crc kubenswrapper[4978]: I0225 09:48:07.344885 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83565f13-d815-4310-b957-0c4c204bd7f5" path="/var/lib/kubelet/pods/83565f13-d815-4310-b957-0c4c204bd7f5/volumes" Feb 25 09:48:16 crc kubenswrapper[4978]: I0225 09:48:16.540751 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:48:16 crc kubenswrapper[4978]: I0225 09:48:16.541509 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:48:39 crc kubenswrapper[4978]: I0225 09:48:39.714639 4978 scope.go:117] "RemoveContainer" containerID="fd87ff67314e8bfd306f7e65ee9d288807e6b8fe583d2129495293c47340ec7c" Feb 25 09:48:39 crc kubenswrapper[4978]: I0225 09:48:39.778246 4978 scope.go:117] "RemoveContainer" containerID="99964fba2fac870611f765e084cf039b3d4601a2cf718b386563a48ab5f39822" Feb 25 09:48:46 crc kubenswrapper[4978]: I0225 09:48:46.540480 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:48:46 crc kubenswrapper[4978]: I0225 09:48:46.541318 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:48:57 crc kubenswrapper[4978]: I0225 09:48:57.989968 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-nzdnv"] Feb 25 09:48:57 crc kubenswrapper[4978]: E0225 09:48:57.991301 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd80fd9a-eb0d-4555-8822-a34bb6085184" containerName="oc" Feb 25 09:48:57 crc kubenswrapper[4978]: I0225 09:48:57.991322 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd80fd9a-eb0d-4555-8822-a34bb6085184" containerName="oc" Feb 25 09:48:57 crc kubenswrapper[4978]: I0225 09:48:57.991665 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd80fd9a-eb0d-4555-8822-a34bb6085184" containerName="oc" Feb 25 09:48:57 crc kubenswrapper[4978]: I0225 09:48:57.992608 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:57 crc kubenswrapper[4978]: I0225 09:48:57.995090 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:48:57 crc kubenswrapper[4978]: I0225 09:48:57.995143 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.002137 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-nzdnv"] Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.106605 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-swiftconf\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.106842 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-scripts\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.106901 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-dispersionconf\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.107165 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl8kr\" (UniqueName: \"kubernetes.io/projected/9b6de1c3-bc6d-464d-99a5-13643a735738-kube-api-access-wl8kr\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.107238 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-ring-data-devices\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.107436 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9b6de1c3-bc6d-464d-99a5-13643a735738-etc-swift\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.107577 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.209376 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl8kr\" (UniqueName: \"kubernetes.io/projected/9b6de1c3-bc6d-464d-99a5-13643a735738-kube-api-access-wl8kr\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.209578 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-ring-data-devices\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.209779 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9b6de1c3-bc6d-464d-99a5-13643a735738-etc-swift\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.209862 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.210006 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-swiftconf\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.210311 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9b6de1c3-bc6d-464d-99a5-13643a735738-etc-swift\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.210345 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-scripts\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.210448 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-dispersionconf\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.210926 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-ring-data-devices\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.210966 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-scripts\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.217547 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.218882 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-swiftconf\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.227716 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl8kr\" (UniqueName: \"kubernetes.io/projected/9b6de1c3-bc6d-464d-99a5-13643a735738-kube-api-access-wl8kr\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.229897 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-dispersionconf\") pod \"swift-ring-rebalance-debug-nzdnv\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.326984 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:48:58 crc kubenswrapper[4978]: I0225 09:48:58.830707 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-nzdnv"] Feb 25 09:48:59 crc kubenswrapper[4978]: I0225 09:48:59.469669 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nzdnv" event={"ID":"9b6de1c3-bc6d-464d-99a5-13643a735738","Type":"ContainerStarted","Data":"21c38642b544dc4523d811bb78efe5121ec40f769c78080834a8abb8a19d38fe"} Feb 25 09:48:59 crc kubenswrapper[4978]: I0225 09:48:59.470003 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nzdnv" event={"ID":"9b6de1c3-bc6d-464d-99a5-13643a735738","Type":"ContainerStarted","Data":"cc487810129f7e624ae41e013e916970c5efbdab74d61c6e36bd7c12584ccaa8"} Feb 25 09:48:59 crc kubenswrapper[4978]: I0225 09:48:59.499811 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-nzdnv" podStartSLOduration=2.499781219 podStartE2EDuration="2.499781219s" podCreationTimestamp="2026-02-25 09:48:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:48:59.486824866 +0000 UTC m=+11032.926081325" watchObservedRunningTime="2026-02-25 09:48:59.499781219 +0000 UTC m=+11032.939037718" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.442352 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jr4n7"] Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.447201 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.462824 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jr4n7"] Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.595360 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1af080e-55f9-4d48-82c6-02160123e37f-catalog-content\") pod \"redhat-operators-jr4n7\" (UID: \"a1af080e-55f9-4d48-82c6-02160123e37f\") " pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.595426 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7jgt\" (UniqueName: \"kubernetes.io/projected/a1af080e-55f9-4d48-82c6-02160123e37f-kube-api-access-p7jgt\") pod \"redhat-operators-jr4n7\" (UID: \"a1af080e-55f9-4d48-82c6-02160123e37f\") " pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.595463 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1af080e-55f9-4d48-82c6-02160123e37f-utilities\") pod \"redhat-operators-jr4n7\" (UID: \"a1af080e-55f9-4d48-82c6-02160123e37f\") " pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.697332 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1af080e-55f9-4d48-82c6-02160123e37f-catalog-content\") pod \"redhat-operators-jr4n7\" (UID: \"a1af080e-55f9-4d48-82c6-02160123e37f\") " pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.697401 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7jgt\" (UniqueName: \"kubernetes.io/projected/a1af080e-55f9-4d48-82c6-02160123e37f-kube-api-access-p7jgt\") pod \"redhat-operators-jr4n7\" (UID: \"a1af080e-55f9-4d48-82c6-02160123e37f\") " pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.697445 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1af080e-55f9-4d48-82c6-02160123e37f-utilities\") pod \"redhat-operators-jr4n7\" (UID: \"a1af080e-55f9-4d48-82c6-02160123e37f\") " pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.697859 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1af080e-55f9-4d48-82c6-02160123e37f-catalog-content\") pod \"redhat-operators-jr4n7\" (UID: \"a1af080e-55f9-4d48-82c6-02160123e37f\") " pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.697882 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1af080e-55f9-4d48-82c6-02160123e37f-utilities\") pod \"redhat-operators-jr4n7\" (UID: \"a1af080e-55f9-4d48-82c6-02160123e37f\") " pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.720288 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7jgt\" (UniqueName: \"kubernetes.io/projected/a1af080e-55f9-4d48-82c6-02160123e37f-kube-api-access-p7jgt\") pod \"redhat-operators-jr4n7\" (UID: \"a1af080e-55f9-4d48-82c6-02160123e37f\") " pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:01 crc kubenswrapper[4978]: I0225 09:49:01.779661 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:02 crc kubenswrapper[4978]: I0225 09:49:02.328104 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jr4n7"] Feb 25 09:49:02 crc kubenswrapper[4978]: W0225 09:49:02.336647 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1af080e_55f9_4d48_82c6_02160123e37f.slice/crio-898a596f9bbabab8681c835c5071be08512b83d209f237c03b7abe8868e76769 WatchSource:0}: Error finding container 898a596f9bbabab8681c835c5071be08512b83d209f237c03b7abe8868e76769: Status 404 returned error can't find the container with id 898a596f9bbabab8681c835c5071be08512b83d209f237c03b7abe8868e76769 Feb 25 09:49:02 crc kubenswrapper[4978]: I0225 09:49:02.509459 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr4n7" event={"ID":"a1af080e-55f9-4d48-82c6-02160123e37f","Type":"ContainerStarted","Data":"898a596f9bbabab8681c835c5071be08512b83d209f237c03b7abe8868e76769"} Feb 25 09:49:03 crc kubenswrapper[4978]: I0225 09:49:03.526148 4978 generic.go:334] "Generic (PLEG): container finished" podID="a1af080e-55f9-4d48-82c6-02160123e37f" containerID="86a1ce737187236a76e6cf29c21e52f0cb68f8475a27be3ab0fbc31292107328" exitCode=0 Feb 25 09:49:03 crc kubenswrapper[4978]: I0225 09:49:03.526328 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr4n7" event={"ID":"a1af080e-55f9-4d48-82c6-02160123e37f","Type":"ContainerDied","Data":"86a1ce737187236a76e6cf29c21e52f0cb68f8475a27be3ab0fbc31292107328"} Feb 25 09:49:09 crc kubenswrapper[4978]: I0225 09:49:09.615689 4978 generic.go:334] "Generic (PLEG): container finished" podID="9b6de1c3-bc6d-464d-99a5-13643a735738" containerID="21c38642b544dc4523d811bb78efe5121ec40f769c78080834a8abb8a19d38fe" exitCode=0 Feb 25 09:49:09 crc kubenswrapper[4978]: I0225 09:49:09.615745 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nzdnv" event={"ID":"9b6de1c3-bc6d-464d-99a5-13643a735738","Type":"ContainerDied","Data":"21c38642b544dc4523d811bb78efe5121ec40f769c78080834a8abb8a19d38fe"} Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.787232 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.846641 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-nzdnv"] Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.861098 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-nzdnv"] Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.947892 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-swiftconf\") pod \"9b6de1c3-bc6d-464d-99a5-13643a735738\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.947949 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9b6de1c3-bc6d-464d-99a5-13643a735738-etc-swift\") pod \"9b6de1c3-bc6d-464d-99a5-13643a735738\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.948012 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-combined-ca-bundle\") pod \"9b6de1c3-bc6d-464d-99a5-13643a735738\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.948576 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-scripts\") pod \"9b6de1c3-bc6d-464d-99a5-13643a735738\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.948645 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl8kr\" (UniqueName: \"kubernetes.io/projected/9b6de1c3-bc6d-464d-99a5-13643a735738-kube-api-access-wl8kr\") pod \"9b6de1c3-bc6d-464d-99a5-13643a735738\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.948971 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b6de1c3-bc6d-464d-99a5-13643a735738-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9b6de1c3-bc6d-464d-99a5-13643a735738" (UID: "9b6de1c3-bc6d-464d-99a5-13643a735738"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.949055 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-dispersionconf\") pod \"9b6de1c3-bc6d-464d-99a5-13643a735738\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.949131 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-ring-data-devices\") pod \"9b6de1c3-bc6d-464d-99a5-13643a735738\" (UID: \"9b6de1c3-bc6d-464d-99a5-13643a735738\") " Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.949635 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9b6de1c3-bc6d-464d-99a5-13643a735738-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.950213 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9b6de1c3-bc6d-464d-99a5-13643a735738" (UID: "9b6de1c3-bc6d-464d-99a5-13643a735738"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.954685 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b6de1c3-bc6d-464d-99a5-13643a735738-kube-api-access-wl8kr" (OuterVolumeSpecName: "kube-api-access-wl8kr") pod "9b6de1c3-bc6d-464d-99a5-13643a735738" (UID: "9b6de1c3-bc6d-464d-99a5-13643a735738"). InnerVolumeSpecName "kube-api-access-wl8kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.985835 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9b6de1c3-bc6d-464d-99a5-13643a735738" (UID: "9b6de1c3-bc6d-464d-99a5-13643a735738"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:49:11 crc kubenswrapper[4978]: I0225 09:49:11.990932 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9b6de1c3-bc6d-464d-99a5-13643a735738" (UID: "9b6de1c3-bc6d-464d-99a5-13643a735738"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:49:12 crc kubenswrapper[4978]: I0225 09:49:12.000955 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b6de1c3-bc6d-464d-99a5-13643a735738" (UID: "9b6de1c3-bc6d-464d-99a5-13643a735738"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:49:12 crc kubenswrapper[4978]: I0225 09:49:12.013930 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-scripts" (OuterVolumeSpecName: "scripts") pod "9b6de1c3-bc6d-464d-99a5-13643a735738" (UID: "9b6de1c3-bc6d-464d-99a5-13643a735738"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:49:12 crc kubenswrapper[4978]: I0225 09:49:12.052088 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:49:12 crc kubenswrapper[4978]: I0225 09:49:12.052127 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:49:12 crc kubenswrapper[4978]: I0225 09:49:12.052137 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl8kr\" (UniqueName: \"kubernetes.io/projected/9b6de1c3-bc6d-464d-99a5-13643a735738-kube-api-access-wl8kr\") on node \"crc\" DevicePath \"\"" Feb 25 09:49:12 crc kubenswrapper[4978]: I0225 09:49:12.052148 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:49:12 crc kubenswrapper[4978]: I0225 09:49:12.052159 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9b6de1c3-bc6d-464d-99a5-13643a735738-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:49:12 crc kubenswrapper[4978]: I0225 09:49:12.052167 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9b6de1c3-bc6d-464d-99a5-13643a735738-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:49:12 crc kubenswrapper[4978]: I0225 09:49:12.648753 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc487810129f7e624ae41e013e916970c5efbdab74d61c6e36bd7c12584ccaa8" Feb 25 09:49:12 crc kubenswrapper[4978]: I0225 09:49:12.648794 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nzdnv" Feb 25 09:49:13 crc kubenswrapper[4978]: I0225 09:49:13.350822 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b6de1c3-bc6d-464d-99a5-13643a735738" path="/var/lib/kubelet/pods/9b6de1c3-bc6d-464d-99a5-13643a735738/volumes" Feb 25 09:49:16 crc kubenswrapper[4978]: I0225 09:49:16.540642 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:49:16 crc kubenswrapper[4978]: I0225 09:49:16.540731 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:49:16 crc kubenswrapper[4978]: I0225 09:49:16.540797 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 09:49:16 crc kubenswrapper[4978]: I0225 09:49:16.542075 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"deecfe5b697e078cff808a3eda1a38a3bf48b1db8ad7591055a352645c854593"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 09:49:16 crc kubenswrapper[4978]: I0225 09:49:16.542480 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://deecfe5b697e078cff808a3eda1a38a3bf48b1db8ad7591055a352645c854593" gracePeriod=600 Feb 25 09:49:16 crc kubenswrapper[4978]: I0225 09:49:16.704516 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="deecfe5b697e078cff808a3eda1a38a3bf48b1db8ad7591055a352645c854593" exitCode=0 Feb 25 09:49:16 crc kubenswrapper[4978]: I0225 09:49:16.704563 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"deecfe5b697e078cff808a3eda1a38a3bf48b1db8ad7591055a352645c854593"} Feb 25 09:49:16 crc kubenswrapper[4978]: I0225 09:49:16.704642 4978 scope.go:117] "RemoveContainer" containerID="929b23417ab41ed72a476fa10aadf67f677ddd1e645905ee515dc145e28c6136" Feb 25 09:49:18 crc kubenswrapper[4978]: I0225 09:49:18.731241 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr4n7" event={"ID":"a1af080e-55f9-4d48-82c6-02160123e37f","Type":"ContainerStarted","Data":"1c4389ce7887405e51898453b6a2cc2d7d09f702d35e44fe02ebead0a7b56d2c"} Feb 25 09:49:18 crc kubenswrapper[4978]: I0225 09:49:18.733521 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c"} Feb 25 09:49:20 crc kubenswrapper[4978]: I0225 09:49:20.779866 4978 generic.go:334] "Generic (PLEG): container finished" podID="a1af080e-55f9-4d48-82c6-02160123e37f" containerID="1c4389ce7887405e51898453b6a2cc2d7d09f702d35e44fe02ebead0a7b56d2c" exitCode=0 Feb 25 09:49:20 crc kubenswrapper[4978]: I0225 09:49:20.780097 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr4n7" event={"ID":"a1af080e-55f9-4d48-82c6-02160123e37f","Type":"ContainerDied","Data":"1c4389ce7887405e51898453b6a2cc2d7d09f702d35e44fe02ebead0a7b56d2c"} Feb 25 09:49:21 crc kubenswrapper[4978]: I0225 09:49:21.793780 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jr4n7" event={"ID":"a1af080e-55f9-4d48-82c6-02160123e37f","Type":"ContainerStarted","Data":"52a85ac2abffe31056a209b2d25aa8aab3f760f7d4c350a89daf39478ffeecbb"} Feb 25 09:49:21 crc kubenswrapper[4978]: I0225 09:49:21.819962 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jr4n7" podStartSLOduration=2.887482738 podStartE2EDuration="20.819935941s" podCreationTimestamp="2026-02-25 09:49:01 +0000 UTC" firstStartedPulling="2026-02-25 09:49:03.529124897 +0000 UTC m=+11036.968381356" lastFinishedPulling="2026-02-25 09:49:21.46157809 +0000 UTC m=+11054.900834559" observedRunningTime="2026-02-25 09:49:21.814383428 +0000 UTC m=+11055.253639917" watchObservedRunningTime="2026-02-25 09:49:21.819935941 +0000 UTC m=+11055.259192420" Feb 25 09:49:31 crc kubenswrapper[4978]: I0225 09:49:31.780088 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:31 crc kubenswrapper[4978]: I0225 09:49:31.782955 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:31 crc kubenswrapper[4978]: I0225 09:49:31.846661 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:32 crc kubenswrapper[4978]: I0225 09:49:32.821641 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jr4n7" Feb 25 09:49:32 crc kubenswrapper[4978]: I0225 09:49:32.927873 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jr4n7"] Feb 25 09:49:32 crc kubenswrapper[4978]: I0225 09:49:32.989659 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bwjnj"] Feb 25 09:49:32 crc kubenswrapper[4978]: I0225 09:49:32.989903 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bwjnj" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerName="registry-server" containerID="cri-o://cffdaf1ed686416ad7d1f7c33ca3c4b1c96e5970f3949c9b2071aaeffe295cbd" gracePeriod=2 Feb 25 09:49:33 crc kubenswrapper[4978]: I0225 09:49:33.773003 4978 generic.go:334] "Generic (PLEG): container finished" podID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerID="cffdaf1ed686416ad7d1f7c33ca3c4b1c96e5970f3949c9b2071aaeffe295cbd" exitCode=0 Feb 25 09:49:33 crc kubenswrapper[4978]: I0225 09:49:33.773075 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwjnj" event={"ID":"b052d38e-931f-4eb1-85ea-142ef3da9881","Type":"ContainerDied","Data":"cffdaf1ed686416ad7d1f7c33ca3c4b1c96e5970f3949c9b2071aaeffe295cbd"} Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.177683 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.353816 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-utilities\") pod \"b052d38e-931f-4eb1-85ea-142ef3da9881\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.353990 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smsvj\" (UniqueName: \"kubernetes.io/projected/b052d38e-931f-4eb1-85ea-142ef3da9881-kube-api-access-smsvj\") pod \"b052d38e-931f-4eb1-85ea-142ef3da9881\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.354191 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-catalog-content\") pod \"b052d38e-931f-4eb1-85ea-142ef3da9881\" (UID: \"b052d38e-931f-4eb1-85ea-142ef3da9881\") " Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.355403 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-utilities" (OuterVolumeSpecName: "utilities") pod "b052d38e-931f-4eb1-85ea-142ef3da9881" (UID: "b052d38e-931f-4eb1-85ea-142ef3da9881"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.364352 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b052d38e-931f-4eb1-85ea-142ef3da9881-kube-api-access-smsvj" (OuterVolumeSpecName: "kube-api-access-smsvj") pod "b052d38e-931f-4eb1-85ea-142ef3da9881" (UID: "b052d38e-931f-4eb1-85ea-142ef3da9881"). InnerVolumeSpecName "kube-api-access-smsvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.456675 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.456707 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smsvj\" (UniqueName: \"kubernetes.io/projected/b052d38e-931f-4eb1-85ea-142ef3da9881-kube-api-access-smsvj\") on node \"crc\" DevicePath \"\"" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.514254 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b052d38e-931f-4eb1-85ea-142ef3da9881" (UID: "b052d38e-931f-4eb1-85ea-142ef3da9881"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.559080 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b052d38e-931f-4eb1-85ea-142ef3da9881-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.786075 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwjnj" event={"ID":"b052d38e-931f-4eb1-85ea-142ef3da9881","Type":"ContainerDied","Data":"c68878c14403d9b5655354b7c8052649095c8f102429053cec66c5675279fbdb"} Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.786150 4978 scope.go:117] "RemoveContainer" containerID="cffdaf1ed686416ad7d1f7c33ca3c4b1c96e5970f3949c9b2071aaeffe295cbd" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.786153 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwjnj" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.810402 4978 scope.go:117] "RemoveContainer" containerID="d443f24582cf5b7de4d32d1b3624d1fa54799a46d0e10d2aca168fcfdc53db2d" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.822937 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bwjnj"] Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.838588 4978 scope.go:117] "RemoveContainer" containerID="b4bc7b9cb20fc0680c1a54d545eae4f31248c2cdb972faf3ab5ce0ae43327b10" Feb 25 09:49:34 crc kubenswrapper[4978]: I0225 09:49:34.852791 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bwjnj"] Feb 25 09:49:35 crc kubenswrapper[4978]: I0225 09:49:35.348571 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" path="/var/lib/kubelet/pods/b052d38e-931f-4eb1-85ea-142ef3da9881/volumes" Feb 25 09:49:39 crc kubenswrapper[4978]: I0225 09:49:39.917576 4978 scope.go:117] "RemoveContainer" containerID="4c8626fa4fcab0ab57df1558940121fe1f9b7e43ee9753873ad98e2baf118846" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.169044 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533550-vpr9z"] Feb 25 09:50:00 crc kubenswrapper[4978]: E0225 09:50:00.170469 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerName="extract-content" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.170487 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerName="extract-content" Feb 25 09:50:00 crc kubenswrapper[4978]: E0225 09:50:00.170506 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerName="registry-server" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.170515 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerName="registry-server" Feb 25 09:50:00 crc kubenswrapper[4978]: E0225 09:50:00.170549 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerName="extract-utilities" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.170559 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerName="extract-utilities" Feb 25 09:50:00 crc kubenswrapper[4978]: E0225 09:50:00.170576 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b6de1c3-bc6d-464d-99a5-13643a735738" containerName="swift-ring-rebalance" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.170584 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b6de1c3-bc6d-464d-99a5-13643a735738" containerName="swift-ring-rebalance" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.170863 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b6de1c3-bc6d-464d-99a5-13643a735738" containerName="swift-ring-rebalance" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.170885 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b052d38e-931f-4eb1-85ea-142ef3da9881" containerName="registry-server" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.171838 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533550-vpr9z" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.176736 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.176968 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.180881 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.201126 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533550-vpr9z"] Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.344308 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj2nm\" (UniqueName: \"kubernetes.io/projected/15f58023-7480-4ce8-8484-67ce19a3bdd8-kube-api-access-sj2nm\") pod \"auto-csr-approver-29533550-vpr9z\" (UID: \"15f58023-7480-4ce8-8484-67ce19a3bdd8\") " pod="openshift-infra/auto-csr-approver-29533550-vpr9z" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.446458 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj2nm\" (UniqueName: \"kubernetes.io/projected/15f58023-7480-4ce8-8484-67ce19a3bdd8-kube-api-access-sj2nm\") pod \"auto-csr-approver-29533550-vpr9z\" (UID: \"15f58023-7480-4ce8-8484-67ce19a3bdd8\") " pod="openshift-infra/auto-csr-approver-29533550-vpr9z" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.471817 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj2nm\" (UniqueName: \"kubernetes.io/projected/15f58023-7480-4ce8-8484-67ce19a3bdd8-kube-api-access-sj2nm\") pod \"auto-csr-approver-29533550-vpr9z\" (UID: \"15f58023-7480-4ce8-8484-67ce19a3bdd8\") " pod="openshift-infra/auto-csr-approver-29533550-vpr9z" Feb 25 09:50:00 crc kubenswrapper[4978]: I0225 09:50:00.503149 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533550-vpr9z" Feb 25 09:50:01 crc kubenswrapper[4978]: I0225 09:50:01.212066 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533550-vpr9z"] Feb 25 09:50:02 crc kubenswrapper[4978]: I0225 09:50:02.104264 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533550-vpr9z" event={"ID":"15f58023-7480-4ce8-8484-67ce19a3bdd8","Type":"ContainerStarted","Data":"b6bfe3919ef9f94304b49ba77fba9568fc656dc3e16735e96f378449d8a5a82f"} Feb 25 09:50:03 crc kubenswrapper[4978]: I0225 09:50:03.133325 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533550-vpr9z" event={"ID":"15f58023-7480-4ce8-8484-67ce19a3bdd8","Type":"ContainerStarted","Data":"e21d70c9ad15d6a2c5d7babfca5a17d068e82991d82ab7e2ac726716713b91fd"} Feb 25 09:50:03 crc kubenswrapper[4978]: I0225 09:50:03.156957 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533550-vpr9z" podStartSLOduration=2.103562674 podStartE2EDuration="3.156941323s" podCreationTimestamp="2026-02-25 09:50:00 +0000 UTC" firstStartedPulling="2026-02-25 09:50:01.226256523 +0000 UTC m=+11094.665512982" lastFinishedPulling="2026-02-25 09:50:02.279635172 +0000 UTC m=+11095.718891631" observedRunningTime="2026-02-25 09:50:03.150029577 +0000 UTC m=+11096.589286056" watchObservedRunningTime="2026-02-25 09:50:03.156941323 +0000 UTC m=+11096.596197782" Feb 25 09:50:04 crc kubenswrapper[4978]: I0225 09:50:04.148201 4978 generic.go:334] "Generic (PLEG): container finished" podID="15f58023-7480-4ce8-8484-67ce19a3bdd8" containerID="e21d70c9ad15d6a2c5d7babfca5a17d068e82991d82ab7e2ac726716713b91fd" exitCode=0 Feb 25 09:50:04 crc kubenswrapper[4978]: I0225 09:50:04.148276 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533550-vpr9z" event={"ID":"15f58023-7480-4ce8-8484-67ce19a3bdd8","Type":"ContainerDied","Data":"e21d70c9ad15d6a2c5d7babfca5a17d068e82991d82ab7e2ac726716713b91fd"} Feb 25 09:50:06 crc kubenswrapper[4978]: I0225 09:50:06.052020 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533550-vpr9z" Feb 25 09:50:06 crc kubenswrapper[4978]: I0225 09:50:06.169792 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533550-vpr9z" event={"ID":"15f58023-7480-4ce8-8484-67ce19a3bdd8","Type":"ContainerDied","Data":"b6bfe3919ef9f94304b49ba77fba9568fc656dc3e16735e96f378449d8a5a82f"} Feb 25 09:50:06 crc kubenswrapper[4978]: I0225 09:50:06.169842 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6bfe3919ef9f94304b49ba77fba9568fc656dc3e16735e96f378449d8a5a82f" Feb 25 09:50:06 crc kubenswrapper[4978]: I0225 09:50:06.170202 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533550-vpr9z" Feb 25 09:50:06 crc kubenswrapper[4978]: I0225 09:50:06.180606 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj2nm\" (UniqueName: \"kubernetes.io/projected/15f58023-7480-4ce8-8484-67ce19a3bdd8-kube-api-access-sj2nm\") pod \"15f58023-7480-4ce8-8484-67ce19a3bdd8\" (UID: \"15f58023-7480-4ce8-8484-67ce19a3bdd8\") " Feb 25 09:50:06 crc kubenswrapper[4978]: I0225 09:50:06.200661 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15f58023-7480-4ce8-8484-67ce19a3bdd8-kube-api-access-sj2nm" (OuterVolumeSpecName: "kube-api-access-sj2nm") pod "15f58023-7480-4ce8-8484-67ce19a3bdd8" (UID: "15f58023-7480-4ce8-8484-67ce19a3bdd8"). InnerVolumeSpecName "kube-api-access-sj2nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:50:06 crc kubenswrapper[4978]: I0225 09:50:06.243844 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533544-bjw6s"] Feb 25 09:50:06 crc kubenswrapper[4978]: I0225 09:50:06.253097 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533544-bjw6s"] Feb 25 09:50:06 crc kubenswrapper[4978]: I0225 09:50:06.283019 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj2nm\" (UniqueName: \"kubernetes.io/projected/15f58023-7480-4ce8-8484-67ce19a3bdd8-kube-api-access-sj2nm\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.348478 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="677ce558-3f07-465e-b6a2-97126eaea074" path="/var/lib/kubelet/pods/677ce558-3f07-465e-b6a2-97126eaea074/volumes" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.434323 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6qjw5"] Feb 25 09:50:07 crc kubenswrapper[4978]: E0225 09:50:07.434929 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f58023-7480-4ce8-8484-67ce19a3bdd8" containerName="oc" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.434956 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f58023-7480-4ce8-8484-67ce19a3bdd8" containerName="oc" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.435251 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="15f58023-7480-4ce8-8484-67ce19a3bdd8" containerName="oc" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.437631 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.493477 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6qjw5"] Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.613909 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-catalog-content\") pod \"certified-operators-6qjw5\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.614028 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdxcn\" (UniqueName: \"kubernetes.io/projected/969040a8-e02f-4b30-a579-db2185d5734d-kube-api-access-wdxcn\") pod \"certified-operators-6qjw5\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.614115 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-utilities\") pod \"certified-operators-6qjw5\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.716405 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-catalog-content\") pod \"certified-operators-6qjw5\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.716518 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdxcn\" (UniqueName: \"kubernetes.io/projected/969040a8-e02f-4b30-a579-db2185d5734d-kube-api-access-wdxcn\") pod \"certified-operators-6qjw5\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.716584 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-utilities\") pod \"certified-operators-6qjw5\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.717106 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-utilities\") pod \"certified-operators-6qjw5\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.717112 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-catalog-content\") pod \"certified-operators-6qjw5\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.733971 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdxcn\" (UniqueName: \"kubernetes.io/projected/969040a8-e02f-4b30-a579-db2185d5734d-kube-api-access-wdxcn\") pod \"certified-operators-6qjw5\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:07 crc kubenswrapper[4978]: I0225 09:50:07.768484 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:08 crc kubenswrapper[4978]: I0225 09:50:08.384567 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6qjw5"] Feb 25 09:50:09 crc kubenswrapper[4978]: I0225 09:50:09.199931 4978 generic.go:334] "Generic (PLEG): container finished" podID="969040a8-e02f-4b30-a579-db2185d5734d" containerID="28d0c5c859ce593fdb9f9d1d0633804dc1d61a049ecb727639adae54c298711e" exitCode=0 Feb 25 09:50:09 crc kubenswrapper[4978]: I0225 09:50:09.199990 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qjw5" event={"ID":"969040a8-e02f-4b30-a579-db2185d5734d","Type":"ContainerDied","Data":"28d0c5c859ce593fdb9f9d1d0633804dc1d61a049ecb727639adae54c298711e"} Feb 25 09:50:09 crc kubenswrapper[4978]: I0225 09:50:09.200483 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qjw5" event={"ID":"969040a8-e02f-4b30-a579-db2185d5734d","Type":"ContainerStarted","Data":"5d26afefe13212ac7189c4b0238a37c910b7eb4f2e807ec78f6b502dd2135712"} Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.038108 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-5dp4p"] Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.041977 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.045445 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.045493 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.053752 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-5dp4p"] Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.212773 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-swiftconf\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.212859 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t696g\" (UniqueName: \"kubernetes.io/projected/ab975f15-eabc-4226-a32c-de4509e987db-kube-api-access-t696g\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.213385 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-dispersionconf\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.213506 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab975f15-eabc-4226-a32c-de4509e987db-etc-swift\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.214060 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.214253 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-ring-data-devices\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.214422 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-scripts\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.316187 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.316250 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-ring-data-devices\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.316289 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-scripts\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.316324 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-swiftconf\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.316348 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t696g\" (UniqueName: \"kubernetes.io/projected/ab975f15-eabc-4226-a32c-de4509e987db-kube-api-access-t696g\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.316420 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-dispersionconf\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.316441 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab975f15-eabc-4226-a32c-de4509e987db-etc-swift\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.316917 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab975f15-eabc-4226-a32c-de4509e987db-etc-swift\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.318156 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-ring-data-devices\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.318190 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-scripts\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.327537 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.328131 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-swiftconf\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.329227 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-dispersionconf\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.334289 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t696g\" (UniqueName: \"kubernetes.io/projected/ab975f15-eabc-4226-a32c-de4509e987db-kube-api-access-t696g\") pod \"swift-ring-rebalance-debug-5dp4p\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:12 crc kubenswrapper[4978]: I0225 09:50:12.391181 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:14 crc kubenswrapper[4978]: I0225 09:50:14.512611 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-5dp4p"] Feb 25 09:50:14 crc kubenswrapper[4978]: W0225 09:50:14.529787 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab975f15_eabc_4226_a32c_de4509e987db.slice/crio-97149e530900f70148d36039c36f7d9276479ec3e9662bfe2e1c3029bafc6cb0 WatchSource:0}: Error finding container 97149e530900f70148d36039c36f7d9276479ec3e9662bfe2e1c3029bafc6cb0: Status 404 returned error can't find the container with id 97149e530900f70148d36039c36f7d9276479ec3e9662bfe2e1c3029bafc6cb0 Feb 25 09:50:15 crc kubenswrapper[4978]: I0225 09:50:15.276731 4978 generic.go:334] "Generic (PLEG): container finished" podID="969040a8-e02f-4b30-a579-db2185d5734d" containerID="91cfbb0c04743a40f7bf750a5c1bcec6b8dd0f811d9f390bee40b1fcb05235bf" exitCode=0 Feb 25 09:50:15 crc kubenswrapper[4978]: I0225 09:50:15.276808 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qjw5" event={"ID":"969040a8-e02f-4b30-a579-db2185d5734d","Type":"ContainerDied","Data":"91cfbb0c04743a40f7bf750a5c1bcec6b8dd0f811d9f390bee40b1fcb05235bf"} Feb 25 09:50:15 crc kubenswrapper[4978]: I0225 09:50:15.298064 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-5dp4p" event={"ID":"ab975f15-eabc-4226-a32c-de4509e987db","Type":"ContainerStarted","Data":"97149e530900f70148d36039c36f7d9276479ec3e9662bfe2e1c3029bafc6cb0"} Feb 25 09:50:16 crc kubenswrapper[4978]: I0225 09:50:16.313359 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qjw5" event={"ID":"969040a8-e02f-4b30-a579-db2185d5734d","Type":"ContainerStarted","Data":"3dea3d3c0d9e14a8c98cf34b101bdeb6cc5f7535af8d47913aa5aac799d38aed"} Feb 25 09:50:16 crc kubenswrapper[4978]: I0225 09:50:16.315508 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-5dp4p" event={"ID":"ab975f15-eabc-4226-a32c-de4509e987db","Type":"ContainerStarted","Data":"cdb458d4af18b1af89a9e2a6effebeee0d26c1e61b8bf8dd4b061d0717959f76"} Feb 25 09:50:16 crc kubenswrapper[4978]: I0225 09:50:16.353036 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6qjw5" podStartSLOduration=2.667259645 podStartE2EDuration="9.353009536s" podCreationTimestamp="2026-02-25 09:50:07 +0000 UTC" firstStartedPulling="2026-02-25 09:50:09.201721209 +0000 UTC m=+11102.640977668" lastFinishedPulling="2026-02-25 09:50:15.88747106 +0000 UTC m=+11109.326727559" observedRunningTime="2026-02-25 09:50:16.337637798 +0000 UTC m=+11109.776894297" watchObservedRunningTime="2026-02-25 09:50:16.353009536 +0000 UTC m=+11109.792266005" Feb 25 09:50:16 crc kubenswrapper[4978]: I0225 09:50:16.375344 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-5dp4p" podStartSLOduration=5.375327101 podStartE2EDuration="5.375327101s" podCreationTimestamp="2026-02-25 09:50:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:50:16.370948625 +0000 UTC m=+11109.810205104" watchObservedRunningTime="2026-02-25 09:50:16.375327101 +0000 UTC m=+11109.814583560" Feb 25 09:50:17 crc kubenswrapper[4978]: I0225 09:50:17.769977 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:17 crc kubenswrapper[4978]: I0225 09:50:17.771483 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:18 crc kubenswrapper[4978]: I0225 09:50:18.830988 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-6qjw5" podUID="969040a8-e02f-4b30-a579-db2185d5734d" containerName="registry-server" probeResult="failure" output=< Feb 25 09:50:18 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 09:50:18 crc kubenswrapper[4978]: > Feb 25 09:50:24 crc kubenswrapper[4978]: I0225 09:50:24.412168 4978 generic.go:334] "Generic (PLEG): container finished" podID="ab975f15-eabc-4226-a32c-de4509e987db" containerID="cdb458d4af18b1af89a9e2a6effebeee0d26c1e61b8bf8dd4b061d0717959f76" exitCode=0 Feb 25 09:50:24 crc kubenswrapper[4978]: I0225 09:50:24.412328 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-5dp4p" event={"ID":"ab975f15-eabc-4226-a32c-de4509e987db","Type":"ContainerDied","Data":"cdb458d4af18b1af89a9e2a6effebeee0d26c1e61b8bf8dd4b061d0717959f76"} Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.644637 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.704360 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-5dp4p"] Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.714867 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-5dp4p"] Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.756273 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-swiftconf\") pod \"ab975f15-eabc-4226-a32c-de4509e987db\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.756358 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab975f15-eabc-4226-a32c-de4509e987db-etc-swift\") pod \"ab975f15-eabc-4226-a32c-de4509e987db\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.756437 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-ring-data-devices\") pod \"ab975f15-eabc-4226-a32c-de4509e987db\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.756512 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-scripts\") pod \"ab975f15-eabc-4226-a32c-de4509e987db\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.756586 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-dispersionconf\") pod \"ab975f15-eabc-4226-a32c-de4509e987db\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.757024 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ab975f15-eabc-4226-a32c-de4509e987db" (UID: "ab975f15-eabc-4226-a32c-de4509e987db"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.757330 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab975f15-eabc-4226-a32c-de4509e987db-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ab975f15-eabc-4226-a32c-de4509e987db" (UID: "ab975f15-eabc-4226-a32c-de4509e987db"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.757341 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t696g\" (UniqueName: \"kubernetes.io/projected/ab975f15-eabc-4226-a32c-de4509e987db-kube-api-access-t696g\") pod \"ab975f15-eabc-4226-a32c-de4509e987db\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.757532 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-combined-ca-bundle\") pod \"ab975f15-eabc-4226-a32c-de4509e987db\" (UID: \"ab975f15-eabc-4226-a32c-de4509e987db\") " Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.758475 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab975f15-eabc-4226-a32c-de4509e987db-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.758703 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.783225 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab975f15-eabc-4226-a32c-de4509e987db-kube-api-access-t696g" (OuterVolumeSpecName: "kube-api-access-t696g") pod "ab975f15-eabc-4226-a32c-de4509e987db" (UID: "ab975f15-eabc-4226-a32c-de4509e987db"). InnerVolumeSpecName "kube-api-access-t696g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.788120 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-scripts" (OuterVolumeSpecName: "scripts") pod "ab975f15-eabc-4226-a32c-de4509e987db" (UID: "ab975f15-eabc-4226-a32c-de4509e987db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.802662 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab975f15-eabc-4226-a32c-de4509e987db" (UID: "ab975f15-eabc-4226-a32c-de4509e987db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.802952 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ab975f15-eabc-4226-a32c-de4509e987db" (UID: "ab975f15-eabc-4226-a32c-de4509e987db"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.807858 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ab975f15-eabc-4226-a32c-de4509e987db" (UID: "ab975f15-eabc-4226-a32c-de4509e987db"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.860570 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t696g\" (UniqueName: \"kubernetes.io/projected/ab975f15-eabc-4226-a32c-de4509e987db-kube-api-access-t696g\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.860615 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.860629 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.860639 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab975f15-eabc-4226-a32c-de4509e987db-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:26 crc kubenswrapper[4978]: I0225 09:50:26.860651 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab975f15-eabc-4226-a32c-de4509e987db-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:27 crc kubenswrapper[4978]: I0225 09:50:27.344160 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab975f15-eabc-4226-a32c-de4509e987db" path="/var/lib/kubelet/pods/ab975f15-eabc-4226-a32c-de4509e987db/volumes" Feb 25 09:50:27 crc kubenswrapper[4978]: I0225 09:50:27.447714 4978 scope.go:117] "RemoveContainer" containerID="cdb458d4af18b1af89a9e2a6effebeee0d26c1e61b8bf8dd4b061d0717959f76" Feb 25 09:50:27 crc kubenswrapper[4978]: I0225 09:50:27.447764 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-5dp4p" Feb 25 09:50:27 crc kubenswrapper[4978]: I0225 09:50:27.833967 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:27 crc kubenswrapper[4978]: I0225 09:50:27.900086 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 09:50:30 crc kubenswrapper[4978]: I0225 09:50:30.109072 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6qjw5"] Feb 25 09:50:30 crc kubenswrapper[4978]: I0225 09:50:30.480684 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p9gbz"] Feb 25 09:50:30 crc kubenswrapper[4978]: I0225 09:50:30.480946 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p9gbz" podUID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerName="registry-server" containerID="cri-o://a5824fbad2e9141759c8ce6c7565783807ce9f28c844443d3d3fe1b52533423e" gracePeriod=2 Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.505593 4978 generic.go:334] "Generic (PLEG): container finished" podID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerID="a5824fbad2e9141759c8ce6c7565783807ce9f28c844443d3d3fe1b52533423e" exitCode=0 Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.505693 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9gbz" event={"ID":"376ae4f8-8441-46a8-ab12-07d3f9c729d9","Type":"ContainerDied","Data":"a5824fbad2e9141759c8ce6c7565783807ce9f28c844443d3d3fe1b52533423e"} Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.677123 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.774536 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-catalog-content\") pod \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.774797 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-utilities\") pod \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.774859 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxpz4\" (UniqueName: \"kubernetes.io/projected/376ae4f8-8441-46a8-ab12-07d3f9c729d9-kube-api-access-wxpz4\") pod \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\" (UID: \"376ae4f8-8441-46a8-ab12-07d3f9c729d9\") " Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.776722 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-utilities" (OuterVolumeSpecName: "utilities") pod "376ae4f8-8441-46a8-ab12-07d3f9c729d9" (UID: "376ae4f8-8441-46a8-ab12-07d3f9c729d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.781700 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/376ae4f8-8441-46a8-ab12-07d3f9c729d9-kube-api-access-wxpz4" (OuterVolumeSpecName: "kube-api-access-wxpz4") pod "376ae4f8-8441-46a8-ab12-07d3f9c729d9" (UID: "376ae4f8-8441-46a8-ab12-07d3f9c729d9"). InnerVolumeSpecName "kube-api-access-wxpz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.869046 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "376ae4f8-8441-46a8-ab12-07d3f9c729d9" (UID: "376ae4f8-8441-46a8-ab12-07d3f9c729d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.877529 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxpz4\" (UniqueName: \"kubernetes.io/projected/376ae4f8-8441-46a8-ab12-07d3f9c729d9-kube-api-access-wxpz4\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.877575 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:31 crc kubenswrapper[4978]: I0225 09:50:31.877621 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/376ae4f8-8441-46a8-ab12-07d3f9c729d9-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:32 crc kubenswrapper[4978]: I0225 09:50:32.526900 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p9gbz" event={"ID":"376ae4f8-8441-46a8-ab12-07d3f9c729d9","Type":"ContainerDied","Data":"62a9b13fbcfb83dd4141d27fb137a93e25a27aec9e0388d2a9c7d562be295a5a"} Feb 25 09:50:32 crc kubenswrapper[4978]: I0225 09:50:32.526983 4978 scope.go:117] "RemoveContainer" containerID="a5824fbad2e9141759c8ce6c7565783807ce9f28c844443d3d3fe1b52533423e" Feb 25 09:50:32 crc kubenswrapper[4978]: I0225 09:50:32.527193 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p9gbz" Feb 25 09:50:32 crc kubenswrapper[4978]: I0225 09:50:32.571574 4978 scope.go:117] "RemoveContainer" containerID="483743f33f036db29860af77324099774dafa4d24b6780d28576261c2a1843ed" Feb 25 09:50:32 crc kubenswrapper[4978]: I0225 09:50:32.604446 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p9gbz"] Feb 25 09:50:32 crc kubenswrapper[4978]: I0225 09:50:32.616031 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p9gbz"] Feb 25 09:50:32 crc kubenswrapper[4978]: I0225 09:50:32.618609 4978 scope.go:117] "RemoveContainer" containerID="651089a42eddc2c0ef7fb45e03a8c62470d00477605851084e76764c43ac5288" Feb 25 09:50:33 crc kubenswrapper[4978]: I0225 09:50:33.348879 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" path="/var/lib/kubelet/pods/376ae4f8-8441-46a8-ab12-07d3f9c729d9/volumes" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.293696 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lrfnt"] Feb 25 09:50:37 crc kubenswrapper[4978]: E0225 09:50:37.295617 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerName="registry-server" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.295730 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerName="registry-server" Feb 25 09:50:37 crc kubenswrapper[4978]: E0225 09:50:37.295842 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerName="extract-utilities" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.295918 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerName="extract-utilities" Feb 25 09:50:37 crc kubenswrapper[4978]: E0225 09:50:37.296020 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerName="extract-content" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.296113 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerName="extract-content" Feb 25 09:50:37 crc kubenswrapper[4978]: E0225 09:50:37.296209 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab975f15-eabc-4226-a32c-de4509e987db" containerName="swift-ring-rebalance" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.296290 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab975f15-eabc-4226-a32c-de4509e987db" containerName="swift-ring-rebalance" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.296649 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab975f15-eabc-4226-a32c-de4509e987db" containerName="swift-ring-rebalance" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.296750 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="376ae4f8-8441-46a8-ab12-07d3f9c729d9" containerName="registry-server" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.298816 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.318293 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lrfnt"] Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.354095 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-catalog-content\") pod \"community-operators-lrfnt\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.354204 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45rd2\" (UniqueName: \"kubernetes.io/projected/17f93342-1b88-4143-98ea-afcfe9bfc5fc-kube-api-access-45rd2\") pod \"community-operators-lrfnt\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.354248 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-utilities\") pod \"community-operators-lrfnt\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.456031 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-utilities\") pod \"community-operators-lrfnt\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.456165 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-catalog-content\") pod \"community-operators-lrfnt\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.456270 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45rd2\" (UniqueName: \"kubernetes.io/projected/17f93342-1b88-4143-98ea-afcfe9bfc5fc-kube-api-access-45rd2\") pod \"community-operators-lrfnt\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.456644 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-catalog-content\") pod \"community-operators-lrfnt\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.456639 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-utilities\") pod \"community-operators-lrfnt\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.476426 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45rd2\" (UniqueName: \"kubernetes.io/projected/17f93342-1b88-4143-98ea-afcfe9bfc5fc-kube-api-access-45rd2\") pod \"community-operators-lrfnt\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:37 crc kubenswrapper[4978]: I0225 09:50:37.627639 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:38 crc kubenswrapper[4978]: I0225 09:50:38.376605 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lrfnt"] Feb 25 09:50:38 crc kubenswrapper[4978]: I0225 09:50:38.606906 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrfnt" event={"ID":"17f93342-1b88-4143-98ea-afcfe9bfc5fc","Type":"ContainerStarted","Data":"400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538"} Feb 25 09:50:38 crc kubenswrapper[4978]: I0225 09:50:38.606965 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrfnt" event={"ID":"17f93342-1b88-4143-98ea-afcfe9bfc5fc","Type":"ContainerStarted","Data":"70e901cf103dc2cef15ba61614a810cde467b9ebde87e504deec3275cbbb3dab"} Feb 25 09:50:39 crc kubenswrapper[4978]: I0225 09:50:39.621348 4978 generic.go:334] "Generic (PLEG): container finished" podID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerID="400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538" exitCode=0 Feb 25 09:50:39 crc kubenswrapper[4978]: I0225 09:50:39.621498 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrfnt" event={"ID":"17f93342-1b88-4143-98ea-afcfe9bfc5fc","Type":"ContainerDied","Data":"400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538"} Feb 25 09:50:40 crc kubenswrapper[4978]: I0225 09:50:40.101488 4978 scope.go:117] "RemoveContainer" containerID="9e8674c1a831eddf11614a8d71a99ffcd693a78212eca2b09d49dea60d79075c" Feb 25 09:50:40 crc kubenswrapper[4978]: I0225 09:50:40.171474 4978 scope.go:117] "RemoveContainer" containerID="776daad305ecf8c4c0101a3a22dd8314e950bf51e4dc4967f3092ef6dc9d8237" Feb 25 09:50:40 crc kubenswrapper[4978]: I0225 09:50:40.634507 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrfnt" event={"ID":"17f93342-1b88-4143-98ea-afcfe9bfc5fc","Type":"ContainerStarted","Data":"c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca"} Feb 25 09:50:42 crc kubenswrapper[4978]: I0225 09:50:42.657219 4978 generic.go:334] "Generic (PLEG): container finished" podID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerID="c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca" exitCode=0 Feb 25 09:50:42 crc kubenswrapper[4978]: I0225 09:50:42.657522 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrfnt" event={"ID":"17f93342-1b88-4143-98ea-afcfe9bfc5fc","Type":"ContainerDied","Data":"c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca"} Feb 25 09:50:43 crc kubenswrapper[4978]: I0225 09:50:43.679323 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrfnt" event={"ID":"17f93342-1b88-4143-98ea-afcfe9bfc5fc","Type":"ContainerStarted","Data":"e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96"} Feb 25 09:50:43 crc kubenswrapper[4978]: I0225 09:50:43.709357 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lrfnt" podStartSLOduration=3.231148712 podStartE2EDuration="6.709332777s" podCreationTimestamp="2026-02-25 09:50:37 +0000 UTC" firstStartedPulling="2026-02-25 09:50:39.624267876 +0000 UTC m=+11133.063524345" lastFinishedPulling="2026-02-25 09:50:43.102451951 +0000 UTC m=+11136.541708410" observedRunningTime="2026-02-25 09:50:43.701828843 +0000 UTC m=+11137.141085302" watchObservedRunningTime="2026-02-25 09:50:43.709332777 +0000 UTC m=+11137.148589236" Feb 25 09:50:47 crc kubenswrapper[4978]: I0225 09:50:47.628310 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:47 crc kubenswrapper[4978]: I0225 09:50:47.629075 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:47 crc kubenswrapper[4978]: I0225 09:50:47.688016 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:57 crc kubenswrapper[4978]: I0225 09:50:57.700122 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:57 crc kubenswrapper[4978]: I0225 09:50:57.749142 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lrfnt"] Feb 25 09:50:57 crc kubenswrapper[4978]: I0225 09:50:57.867346 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lrfnt" podUID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerName="registry-server" containerID="cri-o://e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96" gracePeriod=2 Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.747504 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.854214 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-utilities\") pod \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.854386 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-catalog-content\") pod \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.854491 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45rd2\" (UniqueName: \"kubernetes.io/projected/17f93342-1b88-4143-98ea-afcfe9bfc5fc-kube-api-access-45rd2\") pod \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\" (UID: \"17f93342-1b88-4143-98ea-afcfe9bfc5fc\") " Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.855838 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-utilities" (OuterVolumeSpecName: "utilities") pod "17f93342-1b88-4143-98ea-afcfe9bfc5fc" (UID: "17f93342-1b88-4143-98ea-afcfe9bfc5fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.861219 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17f93342-1b88-4143-98ea-afcfe9bfc5fc-kube-api-access-45rd2" (OuterVolumeSpecName: "kube-api-access-45rd2") pod "17f93342-1b88-4143-98ea-afcfe9bfc5fc" (UID: "17f93342-1b88-4143-98ea-afcfe9bfc5fc"). InnerVolumeSpecName "kube-api-access-45rd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.878290 4978 generic.go:334] "Generic (PLEG): container finished" podID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerID="e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96" exitCode=0 Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.878334 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrfnt" event={"ID":"17f93342-1b88-4143-98ea-afcfe9bfc5fc","Type":"ContainerDied","Data":"e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96"} Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.878360 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lrfnt" event={"ID":"17f93342-1b88-4143-98ea-afcfe9bfc5fc","Type":"ContainerDied","Data":"70e901cf103dc2cef15ba61614a810cde467b9ebde87e504deec3275cbbb3dab"} Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.878396 4978 scope.go:117] "RemoveContainer" containerID="e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96" Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.878394 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lrfnt" Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.924984 4978 scope.go:117] "RemoveContainer" containerID="c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca" Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.931257 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17f93342-1b88-4143-98ea-afcfe9bfc5fc" (UID: "17f93342-1b88-4143-98ea-afcfe9bfc5fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.952674 4978 scope.go:117] "RemoveContainer" containerID="400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538" Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.957926 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.957964 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f93342-1b88-4143-98ea-afcfe9bfc5fc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:58 crc kubenswrapper[4978]: I0225 09:50:58.957979 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45rd2\" (UniqueName: \"kubernetes.io/projected/17f93342-1b88-4143-98ea-afcfe9bfc5fc-kube-api-access-45rd2\") on node \"crc\" DevicePath \"\"" Feb 25 09:50:59 crc kubenswrapper[4978]: I0225 09:50:59.019137 4978 scope.go:117] "RemoveContainer" containerID="e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96" Feb 25 09:50:59 crc kubenswrapper[4978]: E0225 09:50:59.020172 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96\": container with ID starting with e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96 not found: ID does not exist" containerID="e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96" Feb 25 09:50:59 crc kubenswrapper[4978]: I0225 09:50:59.020244 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96"} err="failed to get container status \"e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96\": rpc error: code = NotFound desc = could not find container \"e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96\": container with ID starting with e82bafbb56f80de09d12caf7e8868516b16140f61ed44100a182339c4560bc96 not found: ID does not exist" Feb 25 09:50:59 crc kubenswrapper[4978]: I0225 09:50:59.020271 4978 scope.go:117] "RemoveContainer" containerID="c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca" Feb 25 09:50:59 crc kubenswrapper[4978]: E0225 09:50:59.020655 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca\": container with ID starting with c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca not found: ID does not exist" containerID="c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca" Feb 25 09:50:59 crc kubenswrapper[4978]: I0225 09:50:59.020713 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca"} err="failed to get container status \"c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca\": rpc error: code = NotFound desc = could not find container \"c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca\": container with ID starting with c17be6d7326433c24d1db6c2222c0befac889138c737167105674e70c81194ca not found: ID does not exist" Feb 25 09:50:59 crc kubenswrapper[4978]: I0225 09:50:59.020749 4978 scope.go:117] "RemoveContainer" containerID="400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538" Feb 25 09:50:59 crc kubenswrapper[4978]: E0225 09:50:59.021031 4978 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538\": container with ID starting with 400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538 not found: ID does not exist" containerID="400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538" Feb 25 09:50:59 crc kubenswrapper[4978]: I0225 09:50:59.021064 4978 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538"} err="failed to get container status \"400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538\": rpc error: code = NotFound desc = could not find container \"400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538\": container with ID starting with 400e2335308f4a35946e13ebe26796f10ceabbf6cb539f448736268438e7c538 not found: ID does not exist" Feb 25 09:50:59 crc kubenswrapper[4978]: I0225 09:50:59.234765 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lrfnt"] Feb 25 09:50:59 crc kubenswrapper[4978]: I0225 09:50:59.249798 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lrfnt"] Feb 25 09:50:59 crc kubenswrapper[4978]: I0225 09:50:59.342545 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" path="/var/lib/kubelet/pods/17f93342-1b88-4143-98ea-afcfe9bfc5fc/volumes" Feb 25 09:51:01 crc kubenswrapper[4978]: I0225 09:51:01.010733 4978 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-66cf8f6986-prvmn" podUID="a0d7bad3-d58a-4225-8390-e2785235d412" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.862824 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-nj2p8"] Feb 25 09:51:26 crc kubenswrapper[4978]: E0225 09:51:26.863755 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerName="registry-server" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.863770 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerName="registry-server" Feb 25 09:51:26 crc kubenswrapper[4978]: E0225 09:51:26.863793 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerName="extract-utilities" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.863799 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerName="extract-utilities" Feb 25 09:51:26 crc kubenswrapper[4978]: E0225 09:51:26.863827 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerName="extract-content" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.863833 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerName="extract-content" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.864064 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f93342-1b88-4143-98ea-afcfe9bfc5fc" containerName="registry-server" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.864791 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.867105 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.872473 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.876435 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-nj2p8"] Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.964081 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-scripts\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.964154 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e06582ca-1b37-4469-975a-8cfcd9b42235-etc-swift\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.964195 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-dispersionconf\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.964301 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jtph\" (UniqueName: \"kubernetes.io/projected/e06582ca-1b37-4469-975a-8cfcd9b42235-kube-api-access-5jtph\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.964429 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.964475 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-swiftconf\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:26 crc kubenswrapper[4978]: I0225 09:51:26.964607 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-ring-data-devices\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.066242 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.066353 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-swiftconf\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.066576 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-ring-data-devices\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.066634 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-scripts\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.066689 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e06582ca-1b37-4469-975a-8cfcd9b42235-etc-swift\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.066722 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-dispersionconf\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.066817 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jtph\" (UniqueName: \"kubernetes.io/projected/e06582ca-1b37-4469-975a-8cfcd9b42235-kube-api-access-5jtph\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.067166 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e06582ca-1b37-4469-975a-8cfcd9b42235-etc-swift\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.067455 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-scripts\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.067979 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-ring-data-devices\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.073683 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-dispersionconf\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.073785 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.075517 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-swiftconf\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.090101 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jtph\" (UniqueName: \"kubernetes.io/projected/e06582ca-1b37-4469-975a-8cfcd9b42235-kube-api-access-5jtph\") pod \"swift-ring-rebalance-debug-nj2p8\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.216433 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:27 crc kubenswrapper[4978]: I0225 09:51:27.917912 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-nj2p8"] Feb 25 09:51:27 crc kubenswrapper[4978]: W0225 09:51:27.920820 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode06582ca_1b37_4469_975a_8cfcd9b42235.slice/crio-c63385877ea4d7be05482c21fc7f4289c73636e3b8630a0a9f0055de4cda4a06 WatchSource:0}: Error finding container c63385877ea4d7be05482c21fc7f4289c73636e3b8630a0a9f0055de4cda4a06: Status 404 returned error can't find the container with id c63385877ea4d7be05482c21fc7f4289c73636e3b8630a0a9f0055de4cda4a06 Feb 25 09:51:28 crc kubenswrapper[4978]: I0225 09:51:28.186240 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nj2p8" event={"ID":"e06582ca-1b37-4469-975a-8cfcd9b42235","Type":"ContainerStarted","Data":"5f5db15b2027e2308d7f06fe447661fe864a61e14a7cd7633c425bebf26ebae7"} Feb 25 09:51:28 crc kubenswrapper[4978]: I0225 09:51:28.186282 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nj2p8" event={"ID":"e06582ca-1b37-4469-975a-8cfcd9b42235","Type":"ContainerStarted","Data":"c63385877ea4d7be05482c21fc7f4289c73636e3b8630a0a9f0055de4cda4a06"} Feb 25 09:51:28 crc kubenswrapper[4978]: I0225 09:51:28.203617 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-nj2p8" podStartSLOduration=2.203600307 podStartE2EDuration="2.203600307s" podCreationTimestamp="2026-02-25 09:51:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:51:28.201940456 +0000 UTC m=+11181.641196935" watchObservedRunningTime="2026-02-25 09:51:28.203600307 +0000 UTC m=+11181.642856766" Feb 25 09:51:38 crc kubenswrapper[4978]: I0225 09:51:38.293127 4978 generic.go:334] "Generic (PLEG): container finished" podID="e06582ca-1b37-4469-975a-8cfcd9b42235" containerID="5f5db15b2027e2308d7f06fe447661fe864a61e14a7cd7633c425bebf26ebae7" exitCode=0 Feb 25 09:51:38 crc kubenswrapper[4978]: I0225 09:51:38.293211 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nj2p8" event={"ID":"e06582ca-1b37-4469-975a-8cfcd9b42235","Type":"ContainerDied","Data":"5f5db15b2027e2308d7f06fe447661fe864a61e14a7cd7633c425bebf26ebae7"} Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.322474 4978 scope.go:117] "RemoveContainer" containerID="35bb28f592ad20de882ef44283459b18da4aa95778cdb97f669674f957e9791d" Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.840481 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.880433 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-nj2p8"] Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.892836 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-nj2p8"] Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.963848 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-scripts\") pod \"e06582ca-1b37-4469-975a-8cfcd9b42235\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.963979 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-ring-data-devices\") pod \"e06582ca-1b37-4469-975a-8cfcd9b42235\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.964055 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-dispersionconf\") pod \"e06582ca-1b37-4469-975a-8cfcd9b42235\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.964075 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jtph\" (UniqueName: \"kubernetes.io/projected/e06582ca-1b37-4469-975a-8cfcd9b42235-kube-api-access-5jtph\") pod \"e06582ca-1b37-4469-975a-8cfcd9b42235\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.964115 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e06582ca-1b37-4469-975a-8cfcd9b42235-etc-swift\") pod \"e06582ca-1b37-4469-975a-8cfcd9b42235\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.964156 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-swiftconf\") pod \"e06582ca-1b37-4469-975a-8cfcd9b42235\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.964179 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-combined-ca-bundle\") pod \"e06582ca-1b37-4469-975a-8cfcd9b42235\" (UID: \"e06582ca-1b37-4469-975a-8cfcd9b42235\") " Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.967940 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e06582ca-1b37-4469-975a-8cfcd9b42235" (UID: "e06582ca-1b37-4469-975a-8cfcd9b42235"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.968683 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e06582ca-1b37-4469-975a-8cfcd9b42235-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e06582ca-1b37-4469-975a-8cfcd9b42235" (UID: "e06582ca-1b37-4469-975a-8cfcd9b42235"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:51:40 crc kubenswrapper[4978]: I0225 09:51:40.987173 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e06582ca-1b37-4469-975a-8cfcd9b42235-kube-api-access-5jtph" (OuterVolumeSpecName: "kube-api-access-5jtph") pod "e06582ca-1b37-4469-975a-8cfcd9b42235" (UID: "e06582ca-1b37-4469-975a-8cfcd9b42235"). InnerVolumeSpecName "kube-api-access-5jtph". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.010944 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e06582ca-1b37-4469-975a-8cfcd9b42235" (UID: "e06582ca-1b37-4469-975a-8cfcd9b42235"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.014284 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e06582ca-1b37-4469-975a-8cfcd9b42235" (UID: "e06582ca-1b37-4469-975a-8cfcd9b42235"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.026966 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-scripts" (OuterVolumeSpecName: "scripts") pod "e06582ca-1b37-4469-975a-8cfcd9b42235" (UID: "e06582ca-1b37-4469-975a-8cfcd9b42235"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.027161 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e06582ca-1b37-4469-975a-8cfcd9b42235" (UID: "e06582ca-1b37-4469-975a-8cfcd9b42235"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.073718 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.073755 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e06582ca-1b37-4469-975a-8cfcd9b42235-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.073766 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.073776 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jtph\" (UniqueName: \"kubernetes.io/projected/e06582ca-1b37-4469-975a-8cfcd9b42235-kube-api-access-5jtph\") on node \"crc\" DevicePath \"\"" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.073787 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e06582ca-1b37-4469-975a-8cfcd9b42235-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.073797 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.073805 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e06582ca-1b37-4469-975a-8cfcd9b42235-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.332408 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nj2p8" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.339885 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e06582ca-1b37-4469-975a-8cfcd9b42235" path="/var/lib/kubelet/pods/e06582ca-1b37-4469-975a-8cfcd9b42235/volumes" Feb 25 09:51:41 crc kubenswrapper[4978]: I0225 09:51:41.341165 4978 scope.go:117] "RemoveContainer" containerID="5f5db15b2027e2308d7f06fe447661fe864a61e14a7cd7633c425bebf26ebae7" Feb 25 09:51:46 crc kubenswrapper[4978]: I0225 09:51:46.540970 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:51:46 crc kubenswrapper[4978]: I0225 09:51:46.541489 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.177251 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533552-tt5kt"] Feb 25 09:52:00 crc kubenswrapper[4978]: E0225 09:52:00.178693 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e06582ca-1b37-4469-975a-8cfcd9b42235" containerName="swift-ring-rebalance" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.178720 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e06582ca-1b37-4469-975a-8cfcd9b42235" containerName="swift-ring-rebalance" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.179112 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e06582ca-1b37-4469-975a-8cfcd9b42235" containerName="swift-ring-rebalance" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.180487 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533552-tt5kt" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.184537 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.184570 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.186815 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.194537 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533552-tt5kt"] Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.359940 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv78v\" (UniqueName: \"kubernetes.io/projected/f37933d9-a190-48ad-80d5-fab7402c3389-kube-api-access-hv78v\") pod \"auto-csr-approver-29533552-tt5kt\" (UID: \"f37933d9-a190-48ad-80d5-fab7402c3389\") " pod="openshift-infra/auto-csr-approver-29533552-tt5kt" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.462952 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv78v\" (UniqueName: \"kubernetes.io/projected/f37933d9-a190-48ad-80d5-fab7402c3389-kube-api-access-hv78v\") pod \"auto-csr-approver-29533552-tt5kt\" (UID: \"f37933d9-a190-48ad-80d5-fab7402c3389\") " pod="openshift-infra/auto-csr-approver-29533552-tt5kt" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.484125 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv78v\" (UniqueName: \"kubernetes.io/projected/f37933d9-a190-48ad-80d5-fab7402c3389-kube-api-access-hv78v\") pod \"auto-csr-approver-29533552-tt5kt\" (UID: \"f37933d9-a190-48ad-80d5-fab7402c3389\") " pod="openshift-infra/auto-csr-approver-29533552-tt5kt" Feb 25 09:52:00 crc kubenswrapper[4978]: I0225 09:52:00.500557 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533552-tt5kt" Feb 25 09:52:01 crc kubenswrapper[4978]: I0225 09:52:01.286588 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533552-tt5kt"] Feb 25 09:52:01 crc kubenswrapper[4978]: I0225 09:52:01.304399 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:52:01 crc kubenswrapper[4978]: I0225 09:52:01.547503 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533552-tt5kt" event={"ID":"f37933d9-a190-48ad-80d5-fab7402c3389","Type":"ContainerStarted","Data":"c9a51fda2cdc7c04bc7a67481468cd2116b721f73c8f40899eca49c8f841b077"} Feb 25 09:52:02 crc kubenswrapper[4978]: I0225 09:52:02.559861 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533552-tt5kt" event={"ID":"f37933d9-a190-48ad-80d5-fab7402c3389","Type":"ContainerStarted","Data":"17e3edb35384e83572ddce7044eab08fd9c260d16cefafee081996b8ad606e1c"} Feb 25 09:52:02 crc kubenswrapper[4978]: I0225 09:52:02.576335 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533552-tt5kt" podStartSLOduration=1.755788482 podStartE2EDuration="2.576313401s" podCreationTimestamp="2026-02-25 09:52:00 +0000 UTC" firstStartedPulling="2026-02-25 09:52:01.304093818 +0000 UTC m=+11214.743350277" lastFinishedPulling="2026-02-25 09:52:02.124618687 +0000 UTC m=+11215.563875196" observedRunningTime="2026-02-25 09:52:02.573353189 +0000 UTC m=+11216.012609648" watchObservedRunningTime="2026-02-25 09:52:02.576313401 +0000 UTC m=+11216.015569860" Feb 25 09:52:03 crc kubenswrapper[4978]: I0225 09:52:03.573255 4978 generic.go:334] "Generic (PLEG): container finished" podID="f37933d9-a190-48ad-80d5-fab7402c3389" containerID="17e3edb35384e83572ddce7044eab08fd9c260d16cefafee081996b8ad606e1c" exitCode=0 Feb 25 09:52:03 crc kubenswrapper[4978]: I0225 09:52:03.573314 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533552-tt5kt" event={"ID":"f37933d9-a190-48ad-80d5-fab7402c3389","Type":"ContainerDied","Data":"17e3edb35384e83572ddce7044eab08fd9c260d16cefafee081996b8ad606e1c"} Feb 25 09:52:05 crc kubenswrapper[4978]: I0225 09:52:05.598664 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533552-tt5kt" event={"ID":"f37933d9-a190-48ad-80d5-fab7402c3389","Type":"ContainerDied","Data":"c9a51fda2cdc7c04bc7a67481468cd2116b721f73c8f40899eca49c8f841b077"} Feb 25 09:52:05 crc kubenswrapper[4978]: I0225 09:52:05.599024 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9a51fda2cdc7c04bc7a67481468cd2116b721f73c8f40899eca49c8f841b077" Feb 25 09:52:05 crc kubenswrapper[4978]: I0225 09:52:05.667524 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533552-tt5kt" Feb 25 09:52:05 crc kubenswrapper[4978]: I0225 09:52:05.831593 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv78v\" (UniqueName: \"kubernetes.io/projected/f37933d9-a190-48ad-80d5-fab7402c3389-kube-api-access-hv78v\") pod \"f37933d9-a190-48ad-80d5-fab7402c3389\" (UID: \"f37933d9-a190-48ad-80d5-fab7402c3389\") " Feb 25 09:52:05 crc kubenswrapper[4978]: I0225 09:52:05.841108 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f37933d9-a190-48ad-80d5-fab7402c3389-kube-api-access-hv78v" (OuterVolumeSpecName: "kube-api-access-hv78v") pod "f37933d9-a190-48ad-80d5-fab7402c3389" (UID: "f37933d9-a190-48ad-80d5-fab7402c3389"). InnerVolumeSpecName "kube-api-access-hv78v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:52:05 crc kubenswrapper[4978]: I0225 09:52:05.936052 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv78v\" (UniqueName: \"kubernetes.io/projected/f37933d9-a190-48ad-80d5-fab7402c3389-kube-api-access-hv78v\") on node \"crc\" DevicePath \"\"" Feb 25 09:52:06 crc kubenswrapper[4978]: I0225 09:52:06.608238 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533552-tt5kt" Feb 25 09:52:06 crc kubenswrapper[4978]: I0225 09:52:06.754036 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533546-6d24g"] Feb 25 09:52:06 crc kubenswrapper[4978]: I0225 09:52:06.770458 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533546-6d24g"] Feb 25 09:52:07 crc kubenswrapper[4978]: I0225 09:52:07.341252 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="341213eb-7544-4b05-883b-f4925bc39f03" path="/var/lib/kubelet/pods/341213eb-7544-4b05-883b-f4925bc39f03/volumes" Feb 25 09:52:16 crc kubenswrapper[4978]: I0225 09:52:16.540760 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:52:16 crc kubenswrapper[4978]: I0225 09:52:16.541419 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:52:40 crc kubenswrapper[4978]: I0225 09:52:40.416730 4978 scope.go:117] "RemoveContainer" containerID="c9a54f53cc8cab85d14c843ac080e82b9a21b029c5684a7a36e7328aca364e63" Feb 25 09:52:40 crc kubenswrapper[4978]: I0225 09:52:40.462629 4978 scope.go:117] "RemoveContainer" containerID="7ddeb26c059979325ce291599ee23a29779593d64790c534c75a9444acfdb540" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.170540 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-nm64w"] Feb 25 09:52:41 crc kubenswrapper[4978]: E0225 09:52:41.171441 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f37933d9-a190-48ad-80d5-fab7402c3389" containerName="oc" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.171564 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f37933d9-a190-48ad-80d5-fab7402c3389" containerName="oc" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.171944 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f37933d9-a190-48ad-80d5-fab7402c3389" containerName="oc" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.173023 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.176494 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.176994 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.223882 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-nm64w"] Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.254821 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-swiftconf\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.255495 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.255656 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab864e52-767d-4347-bf59-8056b2c044d3-etc-swift\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.255947 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-dispersionconf\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.256089 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzkgp\" (UniqueName: \"kubernetes.io/projected/ab864e52-767d-4347-bf59-8056b2c044d3-kube-api-access-bzkgp\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.256347 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-ring-data-devices\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.256530 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-scripts\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.358936 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-swiftconf\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.358994 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.359254 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab864e52-767d-4347-bf59-8056b2c044d3-etc-swift\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.359425 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-dispersionconf\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.359470 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzkgp\" (UniqueName: \"kubernetes.io/projected/ab864e52-767d-4347-bf59-8056b2c044d3-kube-api-access-bzkgp\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.359497 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-ring-data-devices\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.359525 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-scripts\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.360505 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-scripts\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.361516 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-ring-data-devices\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.361784 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab864e52-767d-4347-bf59-8056b2c044d3-etc-swift\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.365031 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-dispersionconf\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.365155 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-swiftconf\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.365781 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.380952 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzkgp\" (UniqueName: \"kubernetes.io/projected/ab864e52-767d-4347-bf59-8056b2c044d3-kube-api-access-bzkgp\") pod \"swift-ring-rebalance-debug-nm64w\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:41 crc kubenswrapper[4978]: I0225 09:52:41.512167 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:42 crc kubenswrapper[4978]: W0225 09:52:42.086562 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab864e52_767d_4347_bf59_8056b2c044d3.slice/crio-8fdfde299659750c48c9258c4547349c0c9055729ba0332f79684291b39615f5 WatchSource:0}: Error finding container 8fdfde299659750c48c9258c4547349c0c9055729ba0332f79684291b39615f5: Status 404 returned error can't find the container with id 8fdfde299659750c48c9258c4547349c0c9055729ba0332f79684291b39615f5 Feb 25 09:52:42 crc kubenswrapper[4978]: I0225 09:52:42.088994 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-nm64w"] Feb 25 09:52:43 crc kubenswrapper[4978]: I0225 09:52:43.091278 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nm64w" event={"ID":"ab864e52-767d-4347-bf59-8056b2c044d3","Type":"ContainerStarted","Data":"0d0f302c12d987c491e53525872f46bb3ce6a6fade5c8d6c01a615822bb7d396"} Feb 25 09:52:43 crc kubenswrapper[4978]: I0225 09:52:43.091630 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nm64w" event={"ID":"ab864e52-767d-4347-bf59-8056b2c044d3","Type":"ContainerStarted","Data":"8fdfde299659750c48c9258c4547349c0c9055729ba0332f79684291b39615f5"} Feb 25 09:52:46 crc kubenswrapper[4978]: I0225 09:52:46.540940 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 09:52:46 crc kubenswrapper[4978]: I0225 09:52:46.541478 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 09:52:46 crc kubenswrapper[4978]: I0225 09:52:46.541558 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 09:52:46 crc kubenswrapper[4978]: I0225 09:52:46.542710 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 09:52:46 crc kubenswrapper[4978]: I0225 09:52:46.542845 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" gracePeriod=600 Feb 25 09:52:46 crc kubenswrapper[4978]: E0225 09:52:46.671857 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:52:47 crc kubenswrapper[4978]: I0225 09:52:47.134114 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" exitCode=0 Feb 25 09:52:47 crc kubenswrapper[4978]: I0225 09:52:47.134206 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c"} Feb 25 09:52:47 crc kubenswrapper[4978]: I0225 09:52:47.134567 4978 scope.go:117] "RemoveContainer" containerID="deecfe5b697e078cff808a3eda1a38a3bf48b1db8ad7591055a352645c854593" Feb 25 09:52:47 crc kubenswrapper[4978]: I0225 09:52:47.135462 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:52:47 crc kubenswrapper[4978]: E0225 09:52:47.135916 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:52:47 crc kubenswrapper[4978]: I0225 09:52:47.163332 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-nm64w" podStartSLOduration=6.163304215 podStartE2EDuration="6.163304215s" podCreationTimestamp="2026-02-25 09:52:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:52:43.112552432 +0000 UTC m=+11256.551808901" watchObservedRunningTime="2026-02-25 09:52:47.163304215 +0000 UTC m=+11260.602560684" Feb 25 09:52:53 crc kubenswrapper[4978]: I0225 09:52:53.202675 4978 generic.go:334] "Generic (PLEG): container finished" podID="ab864e52-767d-4347-bf59-8056b2c044d3" containerID="0d0f302c12d987c491e53525872f46bb3ce6a6fade5c8d6c01a615822bb7d396" exitCode=0 Feb 25 09:52:53 crc kubenswrapper[4978]: I0225 09:52:53.202766 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nm64w" event={"ID":"ab864e52-767d-4347-bf59-8056b2c044d3","Type":"ContainerDied","Data":"0d0f302c12d987c491e53525872f46bb3ce6a6fade5c8d6c01a615822bb7d396"} Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.630032 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.699293 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-nm64w"] Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.713988 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-nm64w"] Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.718587 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-combined-ca-bundle\") pod \"ab864e52-767d-4347-bf59-8056b2c044d3\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.718814 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-ring-data-devices\") pod \"ab864e52-767d-4347-bf59-8056b2c044d3\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.718902 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-swiftconf\") pod \"ab864e52-767d-4347-bf59-8056b2c044d3\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.719045 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzkgp\" (UniqueName: \"kubernetes.io/projected/ab864e52-767d-4347-bf59-8056b2c044d3-kube-api-access-bzkgp\") pod \"ab864e52-767d-4347-bf59-8056b2c044d3\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.719961 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ab864e52-767d-4347-bf59-8056b2c044d3" (UID: "ab864e52-767d-4347-bf59-8056b2c044d3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.720065 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-scripts\") pod \"ab864e52-767d-4347-bf59-8056b2c044d3\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.720354 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab864e52-767d-4347-bf59-8056b2c044d3-etc-swift\") pod \"ab864e52-767d-4347-bf59-8056b2c044d3\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.720518 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-dispersionconf\") pod \"ab864e52-767d-4347-bf59-8056b2c044d3\" (UID: \"ab864e52-767d-4347-bf59-8056b2c044d3\") " Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.721824 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab864e52-767d-4347-bf59-8056b2c044d3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ab864e52-767d-4347-bf59-8056b2c044d3" (UID: "ab864e52-767d-4347-bf59-8056b2c044d3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.722275 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab864e52-767d-4347-bf59-8056b2c044d3-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.722422 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.724905 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab864e52-767d-4347-bf59-8056b2c044d3-kube-api-access-bzkgp" (OuterVolumeSpecName: "kube-api-access-bzkgp") pod "ab864e52-767d-4347-bf59-8056b2c044d3" (UID: "ab864e52-767d-4347-bf59-8056b2c044d3"). InnerVolumeSpecName "kube-api-access-bzkgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.750341 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ab864e52-767d-4347-bf59-8056b2c044d3" (UID: "ab864e52-767d-4347-bf59-8056b2c044d3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.756690 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-scripts" (OuterVolumeSpecName: "scripts") pod "ab864e52-767d-4347-bf59-8056b2c044d3" (UID: "ab864e52-767d-4347-bf59-8056b2c044d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.760199 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ab864e52-767d-4347-bf59-8056b2c044d3" (UID: "ab864e52-767d-4347-bf59-8056b2c044d3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.774092 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab864e52-767d-4347-bf59-8056b2c044d3" (UID: "ab864e52-767d-4347-bf59-8056b2c044d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.826160 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.826209 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.826229 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzkgp\" (UniqueName: \"kubernetes.io/projected/ab864e52-767d-4347-bf59-8056b2c044d3-kube-api-access-bzkgp\") on node \"crc\" DevicePath \"\"" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.826247 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab864e52-767d-4347-bf59-8056b2c044d3-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:52:55 crc kubenswrapper[4978]: I0225 09:52:55.826264 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab864e52-767d-4347-bf59-8056b2c044d3-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.013357 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-2wd5v"] Feb 25 09:52:56 crc kubenswrapper[4978]: E0225 09:52:56.013873 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab864e52-767d-4347-bf59-8056b2c044d3" containerName="swift-ring-rebalance" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.013890 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab864e52-767d-4347-bf59-8056b2c044d3" containerName="swift-ring-rebalance" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.014115 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab864e52-767d-4347-bf59-8056b2c044d3" containerName="swift-ring-rebalance" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.014949 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.025961 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-2wd5v"] Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.132850 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-scripts\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.132949 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-ring-data-devices\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.132979 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-dispersionconf\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.133055 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-swiftconf\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.133235 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.133267 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn7nv\" (UniqueName: \"kubernetes.io/projected/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-kube-api-access-wn7nv\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.133330 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-etc-swift\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.235795 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-etc-swift\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.236194 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-etc-swift\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.236490 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-scripts\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.236738 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-ring-data-devices\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.236891 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-dispersionconf\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.237075 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-swiftconf\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.237265 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-scripts\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.237562 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.237748 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn7nv\" (UniqueName: \"kubernetes.io/projected/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-kube-api-access-wn7nv\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.237884 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-ring-data-devices\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.243347 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-dispersionconf\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.243560 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.243807 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-swiftconf\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.256417 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn7nv\" (UniqueName: \"kubernetes.io/projected/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-kube-api-access-wn7nv\") pod \"swift-ring-rebalance-debug-2wd5v\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.260103 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fdfde299659750c48c9258c4547349c0c9055729ba0332f79684291b39615f5" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.260188 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nm64w" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.357547 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:52:56 crc kubenswrapper[4978]: I0225 09:52:56.909684 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-2wd5v"] Feb 25 09:52:57 crc kubenswrapper[4978]: I0225 09:52:57.272908 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2wd5v" event={"ID":"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689","Type":"ContainerStarted","Data":"a0e9e3643bdad4623540131afa3971177df97708423c9556125343381a5a8dcc"} Feb 25 09:52:57 crc kubenswrapper[4978]: I0225 09:52:57.273245 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2wd5v" event={"ID":"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689","Type":"ContainerStarted","Data":"d8d3ecaec99f305ce20a23798143a9e98479ea728ea00eab1ccd914b444fb4d7"} Feb 25 09:52:57 crc kubenswrapper[4978]: I0225 09:52:57.296260 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-2wd5v" podStartSLOduration=2.296242269 podStartE2EDuration="2.296242269s" podCreationTimestamp="2026-02-25 09:52:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:52:57.288712424 +0000 UTC m=+11270.727968903" watchObservedRunningTime="2026-02-25 09:52:57.296242269 +0000 UTC m=+11270.735498728" Feb 25 09:52:57 crc kubenswrapper[4978]: I0225 09:52:57.341039 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab864e52-767d-4347-bf59-8056b2c044d3" path="/var/lib/kubelet/pods/ab864e52-767d-4347-bf59-8056b2c044d3/volumes" Feb 25 09:52:58 crc kubenswrapper[4978]: I0225 09:52:58.327828 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:52:58 crc kubenswrapper[4978]: E0225 09:52:58.328565 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:53:02 crc kubenswrapper[4978]: I0225 09:53:02.330569 4978 generic.go:334] "Generic (PLEG): container finished" podID="de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" containerID="a0e9e3643bdad4623540131afa3971177df97708423c9556125343381a5a8dcc" exitCode=0 Feb 25 09:53:02 crc kubenswrapper[4978]: I0225 09:53:02.331180 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2wd5v" event={"ID":"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689","Type":"ContainerDied","Data":"a0e9e3643bdad4623540131afa3971177df97708423c9556125343381a5a8dcc"} Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.313102 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.358201 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2wd5v" event={"ID":"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689","Type":"ContainerDied","Data":"d8d3ecaec99f305ce20a23798143a9e98479ea728ea00eab1ccd914b444fb4d7"} Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.358257 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8d3ecaec99f305ce20a23798143a9e98479ea728ea00eab1ccd914b444fb4d7" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.358326 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2wd5v" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.358912 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-2wd5v"] Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.371548 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-2wd5v"] Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.469131 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn7nv\" (UniqueName: \"kubernetes.io/projected/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-kube-api-access-wn7nv\") pod \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.469227 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-scripts\") pod \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.469264 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-ring-data-devices\") pod \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.469301 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-swiftconf\") pod \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.469357 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-combined-ca-bundle\") pod \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.469493 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-dispersionconf\") pod \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.469566 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-etc-swift\") pod \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\" (UID: \"de3f04ea-f8bd-47c5-9a66-1b06b4a4f689\") " Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.473334 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" (UID: "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.476032 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" (UID: "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.497517 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-kube-api-access-wn7nv" (OuterVolumeSpecName: "kube-api-access-wn7nv") pod "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" (UID: "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689"). InnerVolumeSpecName "kube-api-access-wn7nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.503246 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-scripts" (OuterVolumeSpecName: "scripts") pod "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" (UID: "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.508347 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" (UID: "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.512479 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" (UID: "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.530690 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" (UID: "de3f04ea-f8bd-47c5-9a66-1b06b4a4f689"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.572460 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.572508 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.572521 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.572530 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.572539 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn7nv\" (UniqueName: \"kubernetes.io/projected/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-kube-api-access-wn7nv\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.572549 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:04 crc kubenswrapper[4978]: I0225 09:53:04.572557 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:05 crc kubenswrapper[4978]: I0225 09:53:05.339863 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" path="/var/lib/kubelet/pods/de3f04ea-f8bd-47c5-9a66-1b06b4a4f689/volumes" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.853940 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-bgw7x"] Feb 25 09:53:07 crc kubenswrapper[4978]: E0225 09:53:07.854977 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" containerName="swift-ring-rebalance" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.854991 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" containerName="swift-ring-rebalance" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.855255 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="de3f04ea-f8bd-47c5-9a66-1b06b4a4f689" containerName="swift-ring-rebalance" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.857180 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.863511 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.863718 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.868744 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-bgw7x"] Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.951540 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a4886958-e594-4aca-994f-81f3b7b6f00c-etc-swift\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.951587 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5z6h\" (UniqueName: \"kubernetes.io/projected/a4886958-e594-4aca-994f-81f3b7b6f00c-kube-api-access-s5z6h\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.951650 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-dispersionconf\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.951675 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-swiftconf\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.951763 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-scripts\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.951795 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-ring-data-devices\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:07 crc kubenswrapper[4978]: I0225 09:53:07.951819 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.053031 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a4886958-e594-4aca-994f-81f3b7b6f00c-etc-swift\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.053114 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5z6h\" (UniqueName: \"kubernetes.io/projected/a4886958-e594-4aca-994f-81f3b7b6f00c-kube-api-access-s5z6h\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.053164 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-dispersionconf\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.053191 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-swiftconf\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.053302 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-scripts\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.053337 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-ring-data-devices\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.053361 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.054518 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a4886958-e594-4aca-994f-81f3b7b6f00c-etc-swift\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.057857 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-scripts\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.058239 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-ring-data-devices\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.062205 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-swiftconf\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.067472 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-dispersionconf\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.071563 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5z6h\" (UniqueName: \"kubernetes.io/projected/a4886958-e594-4aca-994f-81f3b7b6f00c-kube-api-access-s5z6h\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.075629 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-bgw7x\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.176809 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:08 crc kubenswrapper[4978]: I0225 09:53:08.874672 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-bgw7x"] Feb 25 09:53:09 crc kubenswrapper[4978]: I0225 09:53:09.410647 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-bgw7x" event={"ID":"a4886958-e594-4aca-994f-81f3b7b6f00c","Type":"ContainerStarted","Data":"f8e815fc6956a8c72429c60ae2147c2f111ad2e79a28ddc80fc34ec455a40756"} Feb 25 09:53:09 crc kubenswrapper[4978]: I0225 09:53:09.410935 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-bgw7x" event={"ID":"a4886958-e594-4aca-994f-81f3b7b6f00c","Type":"ContainerStarted","Data":"fe730b22f3d30ba6dc871b700d208ab58d1cf2f30aa0f47eb0901d51deae0d54"} Feb 25 09:53:09 crc kubenswrapper[4978]: I0225 09:53:09.435965 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-bgw7x" podStartSLOduration=2.435946189 podStartE2EDuration="2.435946189s" podCreationTimestamp="2026-02-25 09:53:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:53:09.425764092 +0000 UTC m=+11282.865020551" watchObservedRunningTime="2026-02-25 09:53:09.435946189 +0000 UTC m=+11282.875202648" Feb 25 09:53:13 crc kubenswrapper[4978]: I0225 09:53:13.330402 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:53:13 crc kubenswrapper[4978]: E0225 09:53:13.331303 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:53:19 crc kubenswrapper[4978]: I0225 09:53:19.507478 4978 generic.go:334] "Generic (PLEG): container finished" podID="a4886958-e594-4aca-994f-81f3b7b6f00c" containerID="f8e815fc6956a8c72429c60ae2147c2f111ad2e79a28ddc80fc34ec455a40756" exitCode=0 Feb 25 09:53:19 crc kubenswrapper[4978]: I0225 09:53:19.507519 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-bgw7x" event={"ID":"a4886958-e594-4aca-994f-81f3b7b6f00c","Type":"ContainerDied","Data":"f8e815fc6956a8c72429c60ae2147c2f111ad2e79a28ddc80fc34ec455a40756"} Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.536459 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-bgw7x" event={"ID":"a4886958-e594-4aca-994f-81f3b7b6f00c","Type":"ContainerDied","Data":"fe730b22f3d30ba6dc871b700d208ab58d1cf2f30aa0f47eb0901d51deae0d54"} Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.537680 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe730b22f3d30ba6dc871b700d208ab58d1cf2f30aa0f47eb0901d51deae0d54" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.632112 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.682774 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-swiftconf\") pod \"a4886958-e594-4aca-994f-81f3b7b6f00c\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.683619 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-ring-data-devices\") pod \"a4886958-e594-4aca-994f-81f3b7b6f00c\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.683757 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-combined-ca-bundle\") pod \"a4886958-e594-4aca-994f-81f3b7b6f00c\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.683913 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a4886958-e594-4aca-994f-81f3b7b6f00c-etc-swift\") pod \"a4886958-e594-4aca-994f-81f3b7b6f00c\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.684010 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-scripts\") pod \"a4886958-e594-4aca-994f-81f3b7b6f00c\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.684194 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-dispersionconf\") pod \"a4886958-e594-4aca-994f-81f3b7b6f00c\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.684298 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5z6h\" (UniqueName: \"kubernetes.io/projected/a4886958-e594-4aca-994f-81f3b7b6f00c-kube-api-access-s5z6h\") pod \"a4886958-e594-4aca-994f-81f3b7b6f00c\" (UID: \"a4886958-e594-4aca-994f-81f3b7b6f00c\") " Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.689002 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-bgw7x"] Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.689542 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a4886958-e594-4aca-994f-81f3b7b6f00c" (UID: "a4886958-e594-4aca-994f-81f3b7b6f00c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.693050 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4886958-e594-4aca-994f-81f3b7b6f00c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a4886958-e594-4aca-994f-81f3b7b6f00c" (UID: "a4886958-e594-4aca-994f-81f3b7b6f00c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.695795 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4886958-e594-4aca-994f-81f3b7b6f00c-kube-api-access-s5z6h" (OuterVolumeSpecName: "kube-api-access-s5z6h") pod "a4886958-e594-4aca-994f-81f3b7b6f00c" (UID: "a4886958-e594-4aca-994f-81f3b7b6f00c"). InnerVolumeSpecName "kube-api-access-s5z6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.708157 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-bgw7x"] Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.720783 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-scripts" (OuterVolumeSpecName: "scripts") pod "a4886958-e594-4aca-994f-81f3b7b6f00c" (UID: "a4886958-e594-4aca-994f-81f3b7b6f00c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.730559 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4886958-e594-4aca-994f-81f3b7b6f00c" (UID: "a4886958-e594-4aca-994f-81f3b7b6f00c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.732045 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a4886958-e594-4aca-994f-81f3b7b6f00c" (UID: "a4886958-e594-4aca-994f-81f3b7b6f00c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.739694 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a4886958-e594-4aca-994f-81f3b7b6f00c" (UID: "a4886958-e594-4aca-994f-81f3b7b6f00c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.787715 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.787760 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5z6h\" (UniqueName: \"kubernetes.io/projected/a4886958-e594-4aca-994f-81f3b7b6f00c-kube-api-access-s5z6h\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.787823 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.787840 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.787858 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4886958-e594-4aca-994f-81f3b7b6f00c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.787875 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a4886958-e594-4aca-994f-81f3b7b6f00c-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:21 crc kubenswrapper[4978]: I0225 09:53:21.787892 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4886958-e594-4aca-994f-81f3b7b6f00c-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:53:22 crc kubenswrapper[4978]: I0225 09:53:22.547799 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-bgw7x" Feb 25 09:53:23 crc kubenswrapper[4978]: I0225 09:53:23.343481 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4886958-e594-4aca-994f-81f3b7b6f00c" path="/var/lib/kubelet/pods/a4886958-e594-4aca-994f-81f3b7b6f00c/volumes" Feb 25 09:53:24 crc kubenswrapper[4978]: I0225 09:53:24.329271 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:53:24 crc kubenswrapper[4978]: E0225 09:53:24.329892 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:53:37 crc kubenswrapper[4978]: I0225 09:53:37.350901 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:53:37 crc kubenswrapper[4978]: E0225 09:53:37.352007 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:53:52 crc kubenswrapper[4978]: I0225 09:53:52.328672 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:53:52 crc kubenswrapper[4978]: E0225 09:53:52.329937 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.152913 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533554-2g77f"] Feb 25 09:54:00 crc kubenswrapper[4978]: E0225 09:54:00.153873 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4886958-e594-4aca-994f-81f3b7b6f00c" containerName="swift-ring-rebalance" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.153887 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4886958-e594-4aca-994f-81f3b7b6f00c" containerName="swift-ring-rebalance" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.154116 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4886958-e594-4aca-994f-81f3b7b6f00c" containerName="swift-ring-rebalance" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.155175 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533554-2g77f" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.158350 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.158582 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.158978 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.168823 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533554-2g77f"] Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.352935 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mjg5\" (UniqueName: \"kubernetes.io/projected/5d399ffb-1ec7-4b5f-a547-99a9c0b81e09-kube-api-access-7mjg5\") pod \"auto-csr-approver-29533554-2g77f\" (UID: \"5d399ffb-1ec7-4b5f-a547-99a9c0b81e09\") " pod="openshift-infra/auto-csr-approver-29533554-2g77f" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.455361 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mjg5\" (UniqueName: \"kubernetes.io/projected/5d399ffb-1ec7-4b5f-a547-99a9c0b81e09-kube-api-access-7mjg5\") pod \"auto-csr-approver-29533554-2g77f\" (UID: \"5d399ffb-1ec7-4b5f-a547-99a9c0b81e09\") " pod="openshift-infra/auto-csr-approver-29533554-2g77f" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.484830 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mjg5\" (UniqueName: \"kubernetes.io/projected/5d399ffb-1ec7-4b5f-a547-99a9c0b81e09-kube-api-access-7mjg5\") pod \"auto-csr-approver-29533554-2g77f\" (UID: \"5d399ffb-1ec7-4b5f-a547-99a9c0b81e09\") " pod="openshift-infra/auto-csr-approver-29533554-2g77f" Feb 25 09:54:00 crc kubenswrapper[4978]: I0225 09:54:00.503381 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533554-2g77f" Feb 25 09:54:01 crc kubenswrapper[4978]: I0225 09:54:01.181300 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533554-2g77f"] Feb 25 09:54:02 crc kubenswrapper[4978]: I0225 09:54:02.035958 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533554-2g77f" event={"ID":"5d399ffb-1ec7-4b5f-a547-99a9c0b81e09","Type":"ContainerStarted","Data":"ff555cd93cf99d05aa0f5987650452b29fabb80e548cc483419eb66d0a72deb6"} Feb 25 09:54:03 crc kubenswrapper[4978]: I0225 09:54:03.050247 4978 generic.go:334] "Generic (PLEG): container finished" podID="5d399ffb-1ec7-4b5f-a547-99a9c0b81e09" containerID="fd9a38a61a73f52412f0cf6ce127062e73c93e15ffa38550c9a8fb0111d09840" exitCode=0 Feb 25 09:54:03 crc kubenswrapper[4978]: I0225 09:54:03.050320 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533554-2g77f" event={"ID":"5d399ffb-1ec7-4b5f-a547-99a9c0b81e09","Type":"ContainerDied","Data":"fd9a38a61a73f52412f0cf6ce127062e73c93e15ffa38550c9a8fb0111d09840"} Feb 25 09:54:04 crc kubenswrapper[4978]: I0225 09:54:04.972018 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533554-2g77f" Feb 25 09:54:05 crc kubenswrapper[4978]: I0225 09:54:05.051316 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mjg5\" (UniqueName: \"kubernetes.io/projected/5d399ffb-1ec7-4b5f-a547-99a9c0b81e09-kube-api-access-7mjg5\") pod \"5d399ffb-1ec7-4b5f-a547-99a9c0b81e09\" (UID: \"5d399ffb-1ec7-4b5f-a547-99a9c0b81e09\") " Feb 25 09:54:05 crc kubenswrapper[4978]: I0225 09:54:05.059095 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d399ffb-1ec7-4b5f-a547-99a9c0b81e09-kube-api-access-7mjg5" (OuterVolumeSpecName: "kube-api-access-7mjg5") pod "5d399ffb-1ec7-4b5f-a547-99a9c0b81e09" (UID: "5d399ffb-1ec7-4b5f-a547-99a9c0b81e09"). InnerVolumeSpecName "kube-api-access-7mjg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:54:05 crc kubenswrapper[4978]: I0225 09:54:05.077698 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533554-2g77f" event={"ID":"5d399ffb-1ec7-4b5f-a547-99a9c0b81e09","Type":"ContainerDied","Data":"ff555cd93cf99d05aa0f5987650452b29fabb80e548cc483419eb66d0a72deb6"} Feb 25 09:54:05 crc kubenswrapper[4978]: I0225 09:54:05.077741 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff555cd93cf99d05aa0f5987650452b29fabb80e548cc483419eb66d0a72deb6" Feb 25 09:54:05 crc kubenswrapper[4978]: I0225 09:54:05.077771 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533554-2g77f" Feb 25 09:54:05 crc kubenswrapper[4978]: I0225 09:54:05.154222 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mjg5\" (UniqueName: \"kubernetes.io/projected/5d399ffb-1ec7-4b5f-a547-99a9c0b81e09-kube-api-access-7mjg5\") on node \"crc\" DevicePath \"\"" Feb 25 09:54:05 crc kubenswrapper[4978]: I0225 09:54:05.327675 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:54:05 crc kubenswrapper[4978]: E0225 09:54:05.328178 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:54:06 crc kubenswrapper[4978]: I0225 09:54:06.051662 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533548-4khrw"] Feb 25 09:54:06 crc kubenswrapper[4978]: I0225 09:54:06.064877 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533548-4khrw"] Feb 25 09:54:07 crc kubenswrapper[4978]: I0225 09:54:07.343768 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd80fd9a-eb0d-4555-8822-a34bb6085184" path="/var/lib/kubelet/pods/bd80fd9a-eb0d-4555-8822-a34bb6085184/volumes" Feb 25 09:54:19 crc kubenswrapper[4978]: I0225 09:54:19.327532 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:54:19 crc kubenswrapper[4978]: E0225 09:54:19.328133 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.886638 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-2lzp8"] Feb 25 09:54:21 crc kubenswrapper[4978]: E0225 09:54:21.887699 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d399ffb-1ec7-4b5f-a547-99a9c0b81e09" containerName="oc" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.887715 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d399ffb-1ec7-4b5f-a547-99a9c0b81e09" containerName="oc" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.887907 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d399ffb-1ec7-4b5f-a547-99a9c0b81e09" containerName="oc" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.888631 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.896488 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.896780 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.955997 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-2lzp8"] Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.968253 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-scripts\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.968349 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-swiftconf\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.968371 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.968421 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-etc-swift\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.968444 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-dispersionconf\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.968508 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xpsc\" (UniqueName: \"kubernetes.io/projected/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-kube-api-access-5xpsc\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:21 crc kubenswrapper[4978]: I0225 09:54:21.968716 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-ring-data-devices\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.071533 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-swiftconf\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.071606 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.071662 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-etc-swift\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.071714 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-dispersionconf\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.071876 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xpsc\" (UniqueName: \"kubernetes.io/projected/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-kube-api-access-5xpsc\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.071981 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-ring-data-devices\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.072070 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-scripts\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.072281 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-etc-swift\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.072768 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-ring-data-devices\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.073292 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-scripts\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.078289 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-dispersionconf\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.080264 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-swiftconf\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.084263 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.096695 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xpsc\" (UniqueName: \"kubernetes.io/projected/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-kube-api-access-5xpsc\") pod \"swift-ring-rebalance-debug-2lzp8\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.235133 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:22 crc kubenswrapper[4978]: I0225 09:54:22.739000 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-2lzp8"] Feb 25 09:54:23 crc kubenswrapper[4978]: I0225 09:54:23.306864 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2lzp8" event={"ID":"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b","Type":"ContainerStarted","Data":"0068108cea530d5e9576bb600995e2e590d7822b5923cf3a2c80cb203aa2ad9f"} Feb 25 09:54:23 crc kubenswrapper[4978]: I0225 09:54:23.307156 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2lzp8" event={"ID":"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b","Type":"ContainerStarted","Data":"3dbc90d48d9a77129825eb0d1b9976d9eb4eaffae7bc3f8f4886019a45b637c5"} Feb 25 09:54:23 crc kubenswrapper[4978]: I0225 09:54:23.354287 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-2lzp8" podStartSLOduration=2.354260398 podStartE2EDuration="2.354260398s" podCreationTimestamp="2026-02-25 09:54:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:54:23.327760252 +0000 UTC m=+11356.767016731" watchObservedRunningTime="2026-02-25 09:54:23.354260398 +0000 UTC m=+11356.793516867" Feb 25 09:54:32 crc kubenswrapper[4978]: I0225 09:54:32.329121 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:54:32 crc kubenswrapper[4978]: E0225 09:54:32.329957 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:54:33 crc kubenswrapper[4978]: I0225 09:54:33.445153 4978 generic.go:334] "Generic (PLEG): container finished" podID="dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" containerID="0068108cea530d5e9576bb600995e2e590d7822b5923cf3a2c80cb203aa2ad9f" exitCode=0 Feb 25 09:54:33 crc kubenswrapper[4978]: I0225 09:54:33.445795 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-2lzp8" event={"ID":"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b","Type":"ContainerDied","Data":"0068108cea530d5e9576bb600995e2e590d7822b5923cf3a2c80cb203aa2ad9f"} Feb 25 09:54:35 crc kubenswrapper[4978]: I0225 09:54:35.958922 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.000021 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-2lzp8"] Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.011748 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-2lzp8"] Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.139006 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xpsc\" (UniqueName: \"kubernetes.io/projected/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-kube-api-access-5xpsc\") pod \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.139063 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-ring-data-devices\") pod \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.139092 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-scripts\") pod \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.139139 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-dispersionconf\") pod \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.139184 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-swiftconf\") pod \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.139238 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-combined-ca-bundle\") pod \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.139257 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-etc-swift\") pod \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\" (UID: \"dd95450f-aa5d-4bb9-9f8a-540dbe525d3b\") " Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.140683 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" (UID: "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.140914 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" (UID: "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.161260 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-kube-api-access-5xpsc" (OuterVolumeSpecName: "kube-api-access-5xpsc") pod "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" (UID: "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b"). InnerVolumeSpecName "kube-api-access-5xpsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.175654 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-scripts" (OuterVolumeSpecName: "scripts") pod "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" (UID: "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.184175 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" (UID: "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.184544 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" (UID: "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.186451 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" (UID: "dd95450f-aa5d-4bb9-9f8a-540dbe525d3b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.242064 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xpsc\" (UniqueName: \"kubernetes.io/projected/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-kube-api-access-5xpsc\") on node \"crc\" DevicePath \"\"" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.242096 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.242108 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.242117 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.242128 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.242136 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.242144 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.475476 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dbc90d48d9a77129825eb0d1b9976d9eb4eaffae7bc3f8f4886019a45b637c5" Feb 25 09:54:36 crc kubenswrapper[4978]: I0225 09:54:36.475619 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-2lzp8" Feb 25 09:54:37 crc kubenswrapper[4978]: I0225 09:54:37.349192 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" path="/var/lib/kubelet/pods/dd95450f-aa5d-4bb9-9f8a-540dbe525d3b/volumes" Feb 25 09:54:40 crc kubenswrapper[4978]: I0225 09:54:40.656083 4978 scope.go:117] "RemoveContainer" containerID="65a1631b6d75383c8585ac91c4bfaf91a59ff77ae5576d8ae577dd3a69e5d296" Feb 25 09:54:40 crc kubenswrapper[4978]: I0225 09:54:40.707025 4978 scope.go:117] "RemoveContainer" containerID="3381cf5e13bb15501678cb41908eb08881530c85c268e5df66b8afc6c9bb5933" Feb 25 09:54:43 crc kubenswrapper[4978]: I0225 09:54:43.328139 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:54:43 crc kubenswrapper[4978]: E0225 09:54:43.328693 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:54:55 crc kubenswrapper[4978]: I0225 09:54:55.329255 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:54:55 crc kubenswrapper[4978]: E0225 09:54:55.329957 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:55:08 crc kubenswrapper[4978]: I0225 09:55:08.328124 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:55:08 crc kubenswrapper[4978]: E0225 09:55:08.329123 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:55:20 crc kubenswrapper[4978]: I0225 09:55:20.328077 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:55:20 crc kubenswrapper[4978]: E0225 09:55:20.328947 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:55:33 crc kubenswrapper[4978]: I0225 09:55:33.327789 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:55:33 crc kubenswrapper[4978]: E0225 09:55:33.328531 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.196065 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-nz4nv"] Feb 25 09:55:36 crc kubenswrapper[4978]: E0225 09:55:36.197268 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" containerName="swift-ring-rebalance" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.197337 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" containerName="swift-ring-rebalance" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.197650 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd95450f-aa5d-4bb9-9f8a-540dbe525d3b" containerName="swift-ring-rebalance" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.198586 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.201336 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.201343 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.227230 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-nz4nv"] Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.287329 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-scripts\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.287414 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-swiftconf\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.287468 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-dispersionconf\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.287559 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/59858eae-bbec-4995-a788-f34a32cb85b4-etc-swift\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.287583 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjhmb\" (UniqueName: \"kubernetes.io/projected/59858eae-bbec-4995-a788-f34a32cb85b4-kube-api-access-kjhmb\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.287624 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-ring-data-devices\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.287646 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.389508 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-ring-data-devices\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.389754 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.389850 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-scripts\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.389943 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-swiftconf\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.390056 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-dispersionconf\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.390227 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/59858eae-bbec-4995-a788-f34a32cb85b4-etc-swift\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.390317 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjhmb\" (UniqueName: \"kubernetes.io/projected/59858eae-bbec-4995-a788-f34a32cb85b4-kube-api-access-kjhmb\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.390324 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-ring-data-devices\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.390602 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-scripts\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.391111 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/59858eae-bbec-4995-a788-f34a32cb85b4-etc-swift\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.395121 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-dispersionconf\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.395938 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.410696 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-swiftconf\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.411929 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjhmb\" (UniqueName: \"kubernetes.io/projected/59858eae-bbec-4995-a788-f34a32cb85b4-kube-api-access-kjhmb\") pod \"swift-ring-rebalance-debug-nz4nv\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:36 crc kubenswrapper[4978]: I0225 09:55:36.527327 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:37 crc kubenswrapper[4978]: I0225 09:55:37.268670 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-nz4nv"] Feb 25 09:55:38 crc kubenswrapper[4978]: I0225 09:55:38.039181 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nz4nv" event={"ID":"59858eae-bbec-4995-a788-f34a32cb85b4","Type":"ContainerStarted","Data":"f3899c480d8bf57e7f37390f900f1a897efe250956bf1ee27c8bce18aa3a1969"} Feb 25 09:55:38 crc kubenswrapper[4978]: I0225 09:55:38.039242 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nz4nv" event={"ID":"59858eae-bbec-4995-a788-f34a32cb85b4","Type":"ContainerStarted","Data":"8ac576dda34d40a86835306ee114429cd8193f2d215e3bcc85cd589ece5c1096"} Feb 25 09:55:38 crc kubenswrapper[4978]: I0225 09:55:38.067340 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-nz4nv" podStartSLOduration=2.067320457 podStartE2EDuration="2.067320457s" podCreationTimestamp="2026-02-25 09:55:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:55:38.062557068 +0000 UTC m=+11431.501813547" watchObservedRunningTime="2026-02-25 09:55:38.067320457 +0000 UTC m=+11431.506576916" Feb 25 09:55:40 crc kubenswrapper[4978]: I0225 09:55:40.840022 4978 scope.go:117] "RemoveContainer" containerID="21c38642b544dc4523d811bb78efe5121ec40f769c78080834a8abb8a19d38fe" Feb 25 09:55:46 crc kubenswrapper[4978]: I0225 09:55:46.327598 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:55:46 crc kubenswrapper[4978]: E0225 09:55:46.328311 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:55:47 crc kubenswrapper[4978]: I0225 09:55:47.166042 4978 generic.go:334] "Generic (PLEG): container finished" podID="59858eae-bbec-4995-a788-f34a32cb85b4" containerID="f3899c480d8bf57e7f37390f900f1a897efe250956bf1ee27c8bce18aa3a1969" exitCode=0 Feb 25 09:55:47 crc kubenswrapper[4978]: I0225 09:55:47.166117 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-nz4nv" event={"ID":"59858eae-bbec-4995-a788-f34a32cb85b4","Type":"ContainerDied","Data":"f3899c480d8bf57e7f37390f900f1a897efe250956bf1ee27c8bce18aa3a1969"} Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.661322 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.716241 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/59858eae-bbec-4995-a788-f34a32cb85b4-etc-swift\") pod \"59858eae-bbec-4995-a788-f34a32cb85b4\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.716675 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-combined-ca-bundle\") pod \"59858eae-bbec-4995-a788-f34a32cb85b4\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.716747 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-swiftconf\") pod \"59858eae-bbec-4995-a788-f34a32cb85b4\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.716838 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjhmb\" (UniqueName: \"kubernetes.io/projected/59858eae-bbec-4995-a788-f34a32cb85b4-kube-api-access-kjhmb\") pod \"59858eae-bbec-4995-a788-f34a32cb85b4\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.716923 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-dispersionconf\") pod \"59858eae-bbec-4995-a788-f34a32cb85b4\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.717003 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-ring-data-devices\") pod \"59858eae-bbec-4995-a788-f34a32cb85b4\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.717034 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59858eae-bbec-4995-a788-f34a32cb85b4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "59858eae-bbec-4995-a788-f34a32cb85b4" (UID: "59858eae-bbec-4995-a788-f34a32cb85b4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.717071 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-scripts\") pod \"59858eae-bbec-4995-a788-f34a32cb85b4\" (UID: \"59858eae-bbec-4995-a788-f34a32cb85b4\") " Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.717958 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/59858eae-bbec-4995-a788-f34a32cb85b4-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.726028 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59858eae-bbec-4995-a788-f34a32cb85b4-kube-api-access-kjhmb" (OuterVolumeSpecName: "kube-api-access-kjhmb") pod "59858eae-bbec-4995-a788-f34a32cb85b4" (UID: "59858eae-bbec-4995-a788-f34a32cb85b4"). InnerVolumeSpecName "kube-api-access-kjhmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.727616 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-nz4nv"] Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.727679 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "59858eae-bbec-4995-a788-f34a32cb85b4" (UID: "59858eae-bbec-4995-a788-f34a32cb85b4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.743971 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-nz4nv"] Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.766114 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "59858eae-bbec-4995-a788-f34a32cb85b4" (UID: "59858eae-bbec-4995-a788-f34a32cb85b4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.766441 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-scripts" (OuterVolumeSpecName: "scripts") pod "59858eae-bbec-4995-a788-f34a32cb85b4" (UID: "59858eae-bbec-4995-a788-f34a32cb85b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.766716 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59858eae-bbec-4995-a788-f34a32cb85b4" (UID: "59858eae-bbec-4995-a788-f34a32cb85b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.780335 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "59858eae-bbec-4995-a788-f34a32cb85b4" (UID: "59858eae-bbec-4995-a788-f34a32cb85b4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.820115 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.820159 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59858eae-bbec-4995-a788-f34a32cb85b4-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.820172 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.820182 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.820195 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjhmb\" (UniqueName: \"kubernetes.io/projected/59858eae-bbec-4995-a788-f34a32cb85b4-kube-api-access-kjhmb\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:49 crc kubenswrapper[4978]: I0225 09:55:49.820209 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/59858eae-bbec-4995-a788-f34a32cb85b4-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.094156 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-dffpb"] Feb 25 09:55:50 crc kubenswrapper[4978]: E0225 09:55:50.094617 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59858eae-bbec-4995-a788-f34a32cb85b4" containerName="swift-ring-rebalance" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.094634 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="59858eae-bbec-4995-a788-f34a32cb85b4" containerName="swift-ring-rebalance" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.094877 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="59858eae-bbec-4995-a788-f34a32cb85b4" containerName="swift-ring-rebalance" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.095824 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.121624 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-dffpb"] Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.198814 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ac576dda34d40a86835306ee114429cd8193f2d215e3bcc85cd589ece5c1096" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.198904 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-nz4nv" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.228257 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-scripts\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.228322 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cfxr\" (UniqueName: \"kubernetes.io/projected/da632d92-a8c3-4d9b-b62a-5b61a24a178b-kube-api-access-5cfxr\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.228348 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-dispersionconf\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.228711 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.228772 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/da632d92-a8c3-4d9b-b62a-5b61a24a178b-etc-swift\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.228808 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-ring-data-devices\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.228874 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-swiftconf\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.330987 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-scripts\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.331039 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cfxr\" (UniqueName: \"kubernetes.io/projected/da632d92-a8c3-4d9b-b62a-5b61a24a178b-kube-api-access-5cfxr\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.331062 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-dispersionconf\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.331140 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.331188 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/da632d92-a8c3-4d9b-b62a-5b61a24a178b-etc-swift\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.331226 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-ring-data-devices\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.331259 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-swiftconf\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.331730 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/da632d92-a8c3-4d9b-b62a-5b61a24a178b-etc-swift\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.331806 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-scripts\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.332041 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-ring-data-devices\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.334950 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-swiftconf\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.335447 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.336353 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-dispersionconf\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.346505 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cfxr\" (UniqueName: \"kubernetes.io/projected/da632d92-a8c3-4d9b-b62a-5b61a24a178b-kube-api-access-5cfxr\") pod \"swift-ring-rebalance-debug-dffpb\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:50 crc kubenswrapper[4978]: I0225 09:55:50.427943 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:51 crc kubenswrapper[4978]: I0225 09:55:51.156142 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-dffpb"] Feb 25 09:55:51 crc kubenswrapper[4978]: I0225 09:55:51.221017 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-dffpb" event={"ID":"da632d92-a8c3-4d9b-b62a-5b61a24a178b","Type":"ContainerStarted","Data":"b8113bdb4654c25c35d4efd2ae8a5150a186e1ef6e659ef0c3f76059ac78fc5a"} Feb 25 09:55:51 crc kubenswrapper[4978]: I0225 09:55:51.340897 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59858eae-bbec-4995-a788-f34a32cb85b4" path="/var/lib/kubelet/pods/59858eae-bbec-4995-a788-f34a32cb85b4/volumes" Feb 25 09:55:52 crc kubenswrapper[4978]: I0225 09:55:52.233808 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-dffpb" event={"ID":"da632d92-a8c3-4d9b-b62a-5b61a24a178b","Type":"ContainerStarted","Data":"faea6c3c7912ced8f428bed325fb94c31feffcc20e42d035f69ab5d01c702edf"} Feb 25 09:55:52 crc kubenswrapper[4978]: I0225 09:55:52.266406 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-dffpb" podStartSLOduration=2.26636128 podStartE2EDuration="2.26636128s" podCreationTimestamp="2026-02-25 09:55:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:55:52.253347104 +0000 UTC m=+11445.692603573" watchObservedRunningTime="2026-02-25 09:55:52.26636128 +0000 UTC m=+11445.705617769" Feb 25 09:55:56 crc kubenswrapper[4978]: I0225 09:55:56.284889 4978 generic.go:334] "Generic (PLEG): container finished" podID="da632d92-a8c3-4d9b-b62a-5b61a24a178b" containerID="faea6c3c7912ced8f428bed325fb94c31feffcc20e42d035f69ab5d01c702edf" exitCode=0 Feb 25 09:55:56 crc kubenswrapper[4978]: I0225 09:55:56.285021 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-dffpb" event={"ID":"da632d92-a8c3-4d9b-b62a-5b61a24a178b","Type":"ContainerDied","Data":"faea6c3c7912ced8f428bed325fb94c31feffcc20e42d035f69ab5d01c702edf"} Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.568194 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.618978 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-combined-ca-bundle\") pod \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.619192 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/da632d92-a8c3-4d9b-b62a-5b61a24a178b-etc-swift\") pod \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.619287 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-ring-data-devices\") pod \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.619400 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cfxr\" (UniqueName: \"kubernetes.io/projected/da632d92-a8c3-4d9b-b62a-5b61a24a178b-kube-api-access-5cfxr\") pod \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.619479 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-swiftconf\") pod \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.619536 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-scripts\") pod \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.619719 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-dispersionconf\") pod \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\" (UID: \"da632d92-a8c3-4d9b-b62a-5b61a24a178b\") " Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.619838 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-dffpb"] Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.620306 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "da632d92-a8c3-4d9b-b62a-5b61a24a178b" (UID: "da632d92-a8c3-4d9b-b62a-5b61a24a178b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.620515 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.621007 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da632d92-a8c3-4d9b-b62a-5b61a24a178b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "da632d92-a8c3-4d9b-b62a-5b61a24a178b" (UID: "da632d92-a8c3-4d9b-b62a-5b61a24a178b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.626829 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da632d92-a8c3-4d9b-b62a-5b61a24a178b-kube-api-access-5cfxr" (OuterVolumeSpecName: "kube-api-access-5cfxr") pod "da632d92-a8c3-4d9b-b62a-5b61a24a178b" (UID: "da632d92-a8c3-4d9b-b62a-5b61a24a178b"). InnerVolumeSpecName "kube-api-access-5cfxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.633457 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-dffpb"] Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.678728 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "da632d92-a8c3-4d9b-b62a-5b61a24a178b" (UID: "da632d92-a8c3-4d9b-b62a-5b61a24a178b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.683033 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "da632d92-a8c3-4d9b-b62a-5b61a24a178b" (UID: "da632d92-a8c3-4d9b-b62a-5b61a24a178b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.691346 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da632d92-a8c3-4d9b-b62a-5b61a24a178b" (UID: "da632d92-a8c3-4d9b-b62a-5b61a24a178b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.700283 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-scripts" (OuterVolumeSpecName: "scripts") pod "da632d92-a8c3-4d9b-b62a-5b61a24a178b" (UID: "da632d92-a8c3-4d9b-b62a-5b61a24a178b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.722524 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/da632d92-a8c3-4d9b-b62a-5b61a24a178b-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.722558 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cfxr\" (UniqueName: \"kubernetes.io/projected/da632d92-a8c3-4d9b-b62a-5b61a24a178b-kube-api-access-5cfxr\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.722570 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.722578 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da632d92-a8c3-4d9b-b62a-5b61a24a178b-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.722587 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:58 crc kubenswrapper[4978]: I0225 09:55:58.722594 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da632d92-a8c3-4d9b-b62a-5b61a24a178b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:55:59 crc kubenswrapper[4978]: I0225 09:55:59.319675 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8113bdb4654c25c35d4efd2ae8a5150a186e1ef6e659ef0c3f76059ac78fc5a" Feb 25 09:55:59 crc kubenswrapper[4978]: I0225 09:55:59.319757 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-dffpb" Feb 25 09:55:59 crc kubenswrapper[4978]: I0225 09:55:59.340730 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da632d92-a8c3-4d9b-b62a-5b61a24a178b" path="/var/lib/kubelet/pods/da632d92-a8c3-4d9b-b62a-5b61a24a178b/volumes" Feb 25 09:55:59 crc kubenswrapper[4978]: E0225 09:55:59.435766 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda632d92_a8c3_4d9b_b62a_5b61a24a178b.slice\": RecentStats: unable to find data in memory cache]" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.154844 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533556-hn6bn"] Feb 25 09:56:00 crc kubenswrapper[4978]: E0225 09:56:00.155474 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da632d92-a8c3-4d9b-b62a-5b61a24a178b" containerName="swift-ring-rebalance" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.155497 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="da632d92-a8c3-4d9b-b62a-5b61a24a178b" containerName="swift-ring-rebalance" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.155767 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="da632d92-a8c3-4d9b-b62a-5b61a24a178b" containerName="swift-ring-rebalance" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.156713 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533556-hn6bn" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.163453 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.163679 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.163864 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.166280 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533556-hn6bn"] Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.259612 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bc49\" (UniqueName: \"kubernetes.io/projected/54e7776d-9897-4420-ae00-7422083846ee-kube-api-access-2bc49\") pod \"auto-csr-approver-29533556-hn6bn\" (UID: \"54e7776d-9897-4420-ae00-7422083846ee\") " pod="openshift-infra/auto-csr-approver-29533556-hn6bn" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.361451 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bc49\" (UniqueName: \"kubernetes.io/projected/54e7776d-9897-4420-ae00-7422083846ee-kube-api-access-2bc49\") pod \"auto-csr-approver-29533556-hn6bn\" (UID: \"54e7776d-9897-4420-ae00-7422083846ee\") " pod="openshift-infra/auto-csr-approver-29533556-hn6bn" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.383689 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bc49\" (UniqueName: \"kubernetes.io/projected/54e7776d-9897-4420-ae00-7422083846ee-kube-api-access-2bc49\") pod \"auto-csr-approver-29533556-hn6bn\" (UID: \"54e7776d-9897-4420-ae00-7422083846ee\") " pod="openshift-infra/auto-csr-approver-29533556-hn6bn" Feb 25 09:56:00 crc kubenswrapper[4978]: I0225 09:56:00.492381 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533556-hn6bn" Feb 25 09:56:01 crc kubenswrapper[4978]: I0225 09:56:01.248322 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533556-hn6bn"] Feb 25 09:56:01 crc kubenswrapper[4978]: I0225 09:56:01.327885 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:56:01 crc kubenswrapper[4978]: E0225 09:56:01.328225 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:56:01 crc kubenswrapper[4978]: I0225 09:56:01.358920 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533556-hn6bn" event={"ID":"54e7776d-9897-4420-ae00-7422083846ee","Type":"ContainerStarted","Data":"d186a66341e18dd8e72073e5fa2703cbcf1fcf2539a20b2941e2d79035bc71e1"} Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.631652 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-hhtg4"] Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.633472 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.635416 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.635423 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.642350 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-hhtg4"] Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.709868 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-swiftconf\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.709909 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-scripts\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.709972 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zscw4\" (UniqueName: \"kubernetes.io/projected/e087e471-cc1e-49c9-b3d5-474904548861-kube-api-access-zscw4\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.709997 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e087e471-cc1e-49c9-b3d5-474904548861-etc-swift\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.710081 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.710109 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-ring-data-devices\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.710145 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-dispersionconf\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.811722 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-swiftconf\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.811766 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-scripts\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.811808 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zscw4\" (UniqueName: \"kubernetes.io/projected/e087e471-cc1e-49c9-b3d5-474904548861-kube-api-access-zscw4\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.811833 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e087e471-cc1e-49c9-b3d5-474904548861-etc-swift\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.811939 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.811968 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-ring-data-devices\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.812005 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-dispersionconf\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.813215 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e087e471-cc1e-49c9-b3d5-474904548861-etc-swift\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.813804 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-ring-data-devices\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.813954 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-scripts\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.818629 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.818843 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-swiftconf\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.820907 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-dispersionconf\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.828899 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zscw4\" (UniqueName: \"kubernetes.io/projected/e087e471-cc1e-49c9-b3d5-474904548861-kube-api-access-zscw4\") pod \"swift-ring-rebalance-debug-hhtg4\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:02 crc kubenswrapper[4978]: I0225 09:56:02.959837 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:03 crc kubenswrapper[4978]: I0225 09:56:03.366081 4978 generic.go:334] "Generic (PLEG): container finished" podID="54e7776d-9897-4420-ae00-7422083846ee" containerID="c9ebc3b74847af6d32f8155c200063a475c00abc84efd1082e4c85a6e76b83ae" exitCode=0 Feb 25 09:56:03 crc kubenswrapper[4978]: I0225 09:56:03.366148 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533556-hn6bn" event={"ID":"54e7776d-9897-4420-ae00-7422083846ee","Type":"ContainerDied","Data":"c9ebc3b74847af6d32f8155c200063a475c00abc84efd1082e4c85a6e76b83ae"} Feb 25 09:56:03 crc kubenswrapper[4978]: I0225 09:56:03.659816 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-hhtg4"] Feb 25 09:56:04 crc kubenswrapper[4978]: I0225 09:56:04.380409 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-hhtg4" event={"ID":"e087e471-cc1e-49c9-b3d5-474904548861","Type":"ContainerStarted","Data":"ac56bf969c651b9dbd17ea7204ff9a270cb5765f5b0535e968e418a93605e394"} Feb 25 09:56:04 crc kubenswrapper[4978]: I0225 09:56:04.380776 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-hhtg4" event={"ID":"e087e471-cc1e-49c9-b3d5-474904548861","Type":"ContainerStarted","Data":"8a692f94ef3ddca28e81bf5f413aade4a7108706d00826f91e3296e471015781"} Feb 25 09:56:04 crc kubenswrapper[4978]: I0225 09:56:04.412553 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-hhtg4" podStartSLOduration=2.412529806 podStartE2EDuration="2.412529806s" podCreationTimestamp="2026-02-25 09:56:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 09:56:04.401027878 +0000 UTC m=+11457.840284357" watchObservedRunningTime="2026-02-25 09:56:04.412529806 +0000 UTC m=+11457.851786265" Feb 25 09:56:05 crc kubenswrapper[4978]: I0225 09:56:05.541833 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533556-hn6bn" Feb 25 09:56:05 crc kubenswrapper[4978]: I0225 09:56:05.683459 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bc49\" (UniqueName: \"kubernetes.io/projected/54e7776d-9897-4420-ae00-7422083846ee-kube-api-access-2bc49\") pod \"54e7776d-9897-4420-ae00-7422083846ee\" (UID: \"54e7776d-9897-4420-ae00-7422083846ee\") " Feb 25 09:56:05 crc kubenswrapper[4978]: I0225 09:56:05.688917 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54e7776d-9897-4420-ae00-7422083846ee-kube-api-access-2bc49" (OuterVolumeSpecName: "kube-api-access-2bc49") pod "54e7776d-9897-4420-ae00-7422083846ee" (UID: "54e7776d-9897-4420-ae00-7422083846ee"). InnerVolumeSpecName "kube-api-access-2bc49". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:56:05 crc kubenswrapper[4978]: I0225 09:56:05.785510 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bc49\" (UniqueName: \"kubernetes.io/projected/54e7776d-9897-4420-ae00-7422083846ee-kube-api-access-2bc49\") on node \"crc\" DevicePath \"\"" Feb 25 09:56:06 crc kubenswrapper[4978]: I0225 09:56:06.403310 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533556-hn6bn" event={"ID":"54e7776d-9897-4420-ae00-7422083846ee","Type":"ContainerDied","Data":"d186a66341e18dd8e72073e5fa2703cbcf1fcf2539a20b2941e2d79035bc71e1"} Feb 25 09:56:06 crc kubenswrapper[4978]: I0225 09:56:06.403384 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533556-hn6bn" Feb 25 09:56:06 crc kubenswrapper[4978]: I0225 09:56:06.403393 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d186a66341e18dd8e72073e5fa2703cbcf1fcf2539a20b2941e2d79035bc71e1" Feb 25 09:56:06 crc kubenswrapper[4978]: I0225 09:56:06.612804 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533550-vpr9z"] Feb 25 09:56:06 crc kubenswrapper[4978]: I0225 09:56:06.623928 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533550-vpr9z"] Feb 25 09:56:07 crc kubenswrapper[4978]: I0225 09:56:07.350482 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15f58023-7480-4ce8-8484-67ce19a3bdd8" path="/var/lib/kubelet/pods/15f58023-7480-4ce8-8484-67ce19a3bdd8/volumes" Feb 25 09:56:13 crc kubenswrapper[4978]: I0225 09:56:13.472602 4978 generic.go:334] "Generic (PLEG): container finished" podID="e087e471-cc1e-49c9-b3d5-474904548861" containerID="ac56bf969c651b9dbd17ea7204ff9a270cb5765f5b0535e968e418a93605e394" exitCode=0 Feb 25 09:56:13 crc kubenswrapper[4978]: I0225 09:56:13.472674 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-hhtg4" event={"ID":"e087e471-cc1e-49c9-b3d5-474904548861","Type":"ContainerDied","Data":"ac56bf969c651b9dbd17ea7204ff9a270cb5765f5b0535e968e418a93605e394"} Feb 25 09:56:14 crc kubenswrapper[4978]: I0225 09:56:14.327576 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:56:14 crc kubenswrapper[4978]: E0225 09:56:14.327974 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.053650 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.088452 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-hhtg4"] Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.106530 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-hhtg4"] Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.203506 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-ring-data-devices\") pod \"e087e471-cc1e-49c9-b3d5-474904548861\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.203566 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e087e471-cc1e-49c9-b3d5-474904548861-etc-swift\") pod \"e087e471-cc1e-49c9-b3d5-474904548861\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.203602 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-dispersionconf\") pod \"e087e471-cc1e-49c9-b3d5-474904548861\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.203624 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-scripts\") pod \"e087e471-cc1e-49c9-b3d5-474904548861\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.203928 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-combined-ca-bundle\") pod \"e087e471-cc1e-49c9-b3d5-474904548861\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.203985 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-swiftconf\") pod \"e087e471-cc1e-49c9-b3d5-474904548861\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.204026 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zscw4\" (UniqueName: \"kubernetes.io/projected/e087e471-cc1e-49c9-b3d5-474904548861-kube-api-access-zscw4\") pod \"e087e471-cc1e-49c9-b3d5-474904548861\" (UID: \"e087e471-cc1e-49c9-b3d5-474904548861\") " Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.204830 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e087e471-cc1e-49c9-b3d5-474904548861-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e087e471-cc1e-49c9-b3d5-474904548861" (UID: "e087e471-cc1e-49c9-b3d5-474904548861"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.205001 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e087e471-cc1e-49c9-b3d5-474904548861" (UID: "e087e471-cc1e-49c9-b3d5-474904548861"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.211103 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e087e471-cc1e-49c9-b3d5-474904548861-kube-api-access-zscw4" (OuterVolumeSpecName: "kube-api-access-zscw4") pod "e087e471-cc1e-49c9-b3d5-474904548861" (UID: "e087e471-cc1e-49c9-b3d5-474904548861"). InnerVolumeSpecName "kube-api-access-zscw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.237919 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e087e471-cc1e-49c9-b3d5-474904548861" (UID: "e087e471-cc1e-49c9-b3d5-474904548861"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.242016 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e087e471-cc1e-49c9-b3d5-474904548861" (UID: "e087e471-cc1e-49c9-b3d5-474904548861"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.244587 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e087e471-cc1e-49c9-b3d5-474904548861" (UID: "e087e471-cc1e-49c9-b3d5-474904548861"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.250313 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-scripts" (OuterVolumeSpecName: "scripts") pod "e087e471-cc1e-49c9-b3d5-474904548861" (UID: "e087e471-cc1e-49c9-b3d5-474904548861"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.306820 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.306854 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.306866 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zscw4\" (UniqueName: \"kubernetes.io/projected/e087e471-cc1e-49c9-b3d5-474904548861-kube-api-access-zscw4\") on node \"crc\" DevicePath \"\"" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.306875 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.306884 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e087e471-cc1e-49c9-b3d5-474904548861-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.306891 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e087e471-cc1e-49c9-b3d5-474904548861-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.306900 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e087e471-cc1e-49c9-b3d5-474904548861-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.533328 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a692f94ef3ddca28e81bf5f413aade4a7108706d00826f91e3296e471015781" Feb 25 09:56:16 crc kubenswrapper[4978]: I0225 09:56:16.533423 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-hhtg4" Feb 25 09:56:17 crc kubenswrapper[4978]: I0225 09:56:17.338534 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e087e471-cc1e-49c9-b3d5-474904548861" path="/var/lib/kubelet/pods/e087e471-cc1e-49c9-b3d5-474904548861/volumes" Feb 25 09:56:26 crc kubenswrapper[4978]: I0225 09:56:26.328317 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:56:26 crc kubenswrapper[4978]: E0225 09:56:26.329505 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:56:40 crc kubenswrapper[4978]: I0225 09:56:40.328256 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:56:40 crc kubenswrapper[4978]: E0225 09:56:40.330183 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:56:40 crc kubenswrapper[4978]: I0225 09:56:40.940234 4978 scope.go:117] "RemoveContainer" containerID="e21d70c9ad15d6a2c5d7babfca5a17d068e82991d82ab7e2ac726716713b91fd" Feb 25 09:56:54 crc kubenswrapper[4978]: I0225 09:56:54.328650 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:56:54 crc kubenswrapper[4978]: E0225 09:56:54.329973 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:57:05 crc kubenswrapper[4978]: I0225 09:57:05.331337 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:57:05 crc kubenswrapper[4978]: E0225 09:57:05.332200 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:57:20 crc kubenswrapper[4978]: I0225 09:57:20.327975 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:57:20 crc kubenswrapper[4978]: E0225 09:57:20.328851 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:57:33 crc kubenswrapper[4978]: I0225 09:57:33.328553 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:57:33 crc kubenswrapper[4978]: E0225 09:57:33.329507 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 09:57:47 crc kubenswrapper[4978]: I0225 09:57:47.340946 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 09:57:48 crc kubenswrapper[4978]: I0225 09:57:48.563032 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"ddedb687f1478b20f3544dbc0c4281804392d9541579ad2e88d046540e639049"} Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.150775 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533558-vwg6p"] Feb 25 09:58:00 crc kubenswrapper[4978]: E0225 09:58:00.151704 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e087e471-cc1e-49c9-b3d5-474904548861" containerName="swift-ring-rebalance" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.151717 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="e087e471-cc1e-49c9-b3d5-474904548861" containerName="swift-ring-rebalance" Feb 25 09:58:00 crc kubenswrapper[4978]: E0225 09:58:00.151740 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54e7776d-9897-4420-ae00-7422083846ee" containerName="oc" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.151746 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="54e7776d-9897-4420-ae00-7422083846ee" containerName="oc" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.151939 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="54e7776d-9897-4420-ae00-7422083846ee" containerName="oc" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.151963 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="e087e471-cc1e-49c9-b3d5-474904548861" containerName="swift-ring-rebalance" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.152752 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533558-vwg6p" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.154777 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.155428 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.155435 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.162588 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533558-vwg6p"] Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.281469 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pqg2\" (UniqueName: \"kubernetes.io/projected/2e1095ae-2b2d-480d-8a71-e056849fa6b5-kube-api-access-6pqg2\") pod \"auto-csr-approver-29533558-vwg6p\" (UID: \"2e1095ae-2b2d-480d-8a71-e056849fa6b5\") " pod="openshift-infra/auto-csr-approver-29533558-vwg6p" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.383466 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pqg2\" (UniqueName: \"kubernetes.io/projected/2e1095ae-2b2d-480d-8a71-e056849fa6b5-kube-api-access-6pqg2\") pod \"auto-csr-approver-29533558-vwg6p\" (UID: \"2e1095ae-2b2d-480d-8a71-e056849fa6b5\") " pod="openshift-infra/auto-csr-approver-29533558-vwg6p" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.415623 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pqg2\" (UniqueName: \"kubernetes.io/projected/2e1095ae-2b2d-480d-8a71-e056849fa6b5-kube-api-access-6pqg2\") pod \"auto-csr-approver-29533558-vwg6p\" (UID: \"2e1095ae-2b2d-480d-8a71-e056849fa6b5\") " pod="openshift-infra/auto-csr-approver-29533558-vwg6p" Feb 25 09:58:00 crc kubenswrapper[4978]: I0225 09:58:00.469611 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533558-vwg6p" Feb 25 09:58:01 crc kubenswrapper[4978]: I0225 09:58:01.380966 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533558-vwg6p"] Feb 25 09:58:01 crc kubenswrapper[4978]: I0225 09:58:01.385014 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 09:58:01 crc kubenswrapper[4978]: I0225 09:58:01.717845 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533558-vwg6p" event={"ID":"2e1095ae-2b2d-480d-8a71-e056849fa6b5","Type":"ContainerStarted","Data":"8c0c55eaacff815890fcc12af5e1e27bfc8f5b9dc935940381cfde63a43072b8"} Feb 25 09:58:02 crc kubenswrapper[4978]: I0225 09:58:02.728012 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533558-vwg6p" event={"ID":"2e1095ae-2b2d-480d-8a71-e056849fa6b5","Type":"ContainerStarted","Data":"6533d65023411a5f7bd38c24d6762bf5885bd9e3b9f8a7b9fa39cd976c279378"} Feb 25 09:58:02 crc kubenswrapper[4978]: I0225 09:58:02.750311 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533558-vwg6p" podStartSLOduration=1.7772256899999999 podStartE2EDuration="2.75029002s" podCreationTimestamp="2026-02-25 09:58:00 +0000 UTC" firstStartedPulling="2026-02-25 09:58:01.384812766 +0000 UTC m=+11574.824069215" lastFinishedPulling="2026-02-25 09:58:02.357877066 +0000 UTC m=+11575.797133545" observedRunningTime="2026-02-25 09:58:02.748791383 +0000 UTC m=+11576.188047852" watchObservedRunningTime="2026-02-25 09:58:02.75029002 +0000 UTC m=+11576.189546479" Feb 25 09:58:03 crc kubenswrapper[4978]: I0225 09:58:03.744864 4978 generic.go:334] "Generic (PLEG): container finished" podID="2e1095ae-2b2d-480d-8a71-e056849fa6b5" containerID="6533d65023411a5f7bd38c24d6762bf5885bd9e3b9f8a7b9fa39cd976c279378" exitCode=0 Feb 25 09:58:03 crc kubenswrapper[4978]: I0225 09:58:03.745146 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533558-vwg6p" event={"ID":"2e1095ae-2b2d-480d-8a71-e056849fa6b5","Type":"ContainerDied","Data":"6533d65023411a5f7bd38c24d6762bf5885bd9e3b9f8a7b9fa39cd976c279378"} Feb 25 09:58:06 crc kubenswrapper[4978]: I0225 09:58:06.329332 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533558-vwg6p" Feb 25 09:58:06 crc kubenswrapper[4978]: I0225 09:58:06.411661 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pqg2\" (UniqueName: \"kubernetes.io/projected/2e1095ae-2b2d-480d-8a71-e056849fa6b5-kube-api-access-6pqg2\") pod \"2e1095ae-2b2d-480d-8a71-e056849fa6b5\" (UID: \"2e1095ae-2b2d-480d-8a71-e056849fa6b5\") " Feb 25 09:58:06 crc kubenswrapper[4978]: I0225 09:58:06.420001 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e1095ae-2b2d-480d-8a71-e056849fa6b5-kube-api-access-6pqg2" (OuterVolumeSpecName: "kube-api-access-6pqg2") pod "2e1095ae-2b2d-480d-8a71-e056849fa6b5" (UID: "2e1095ae-2b2d-480d-8a71-e056849fa6b5"). InnerVolumeSpecName "kube-api-access-6pqg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 09:58:06 crc kubenswrapper[4978]: I0225 09:58:06.514879 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pqg2\" (UniqueName: \"kubernetes.io/projected/2e1095ae-2b2d-480d-8a71-e056849fa6b5-kube-api-access-6pqg2\") on node \"crc\" DevicePath \"\"" Feb 25 09:58:06 crc kubenswrapper[4978]: I0225 09:58:06.788053 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533558-vwg6p" event={"ID":"2e1095ae-2b2d-480d-8a71-e056849fa6b5","Type":"ContainerDied","Data":"8c0c55eaacff815890fcc12af5e1e27bfc8f5b9dc935940381cfde63a43072b8"} Feb 25 09:58:06 crc kubenswrapper[4978]: I0225 09:58:06.788088 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533558-vwg6p" Feb 25 09:58:06 crc kubenswrapper[4978]: I0225 09:58:06.788091 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c0c55eaacff815890fcc12af5e1e27bfc8f5b9dc935940381cfde63a43072b8" Feb 25 09:58:07 crc kubenswrapper[4978]: I0225 09:58:07.416360 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533552-tt5kt"] Feb 25 09:58:07 crc kubenswrapper[4978]: I0225 09:58:07.432525 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533552-tt5kt"] Feb 25 09:58:09 crc kubenswrapper[4978]: I0225 09:58:09.338486 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f37933d9-a190-48ad-80d5-fab7402c3389" path="/var/lib/kubelet/pods/f37933d9-a190-48ad-80d5-fab7402c3389/volumes" Feb 25 09:58:41 crc kubenswrapper[4978]: I0225 09:58:41.073688 4978 scope.go:117] "RemoveContainer" containerID="17e3edb35384e83572ddce7044eab08fd9c260d16cefafee081996b8ad606e1c" Feb 25 09:59:41 crc kubenswrapper[4978]: I0225 09:59:41.164393 4978 scope.go:117] "RemoveContainer" containerID="f8e815fc6956a8c72429c60ae2147c2f111ad2e79a28ddc80fc34ec455a40756" Feb 25 09:59:41 crc kubenswrapper[4978]: I0225 09:59:41.204153 4978 scope.go:117] "RemoveContainer" containerID="a0e9e3643bdad4623540131afa3971177df97708423c9556125343381a5a8dcc" Feb 25 09:59:41 crc kubenswrapper[4978]: I0225 09:59:41.293393 4978 scope.go:117] "RemoveContainer" containerID="0d0f302c12d987c491e53525872f46bb3ce6a6fade5c8d6c01a615822bb7d396" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.106142 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-debug-wxz4w"] Feb 25 09:59:59 crc kubenswrapper[4978]: E0225 09:59:59.107585 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1095ae-2b2d-480d-8a71-e056849fa6b5" containerName="oc" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.107600 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1095ae-2b2d-480d-8a71-e056849fa6b5" containerName="oc" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.107832 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1095ae-2b2d-480d-8a71-e056849fa6b5" containerName="oc" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.109139 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.111173 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.112825 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.129833 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-wxz4w"] Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.201518 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-scripts\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.201635 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-ring-data-devices\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.201775 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.201927 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxsrj\" (UniqueName: \"kubernetes.io/projected/a809020b-dd80-4507-9aae-e88c2784d96c-kube-api-access-lxsrj\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.201987 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-swiftconf\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.202167 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-dispersionconf\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.202199 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a809020b-dd80-4507-9aae-e88c2784d96c-etc-swift\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.304265 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-dispersionconf\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.304576 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a809020b-dd80-4507-9aae-e88c2784d96c-etc-swift\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.304674 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-scripts\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.304814 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-ring-data-devices\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.304920 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.304994 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a809020b-dd80-4507-9aae-e88c2784d96c-etc-swift\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.305103 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxsrj\" (UniqueName: \"kubernetes.io/projected/a809020b-dd80-4507-9aae-e88c2784d96c-kube-api-access-lxsrj\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.305188 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-swiftconf\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.305417 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-scripts\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.305440 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-ring-data-devices\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.309830 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-dispersionconf\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.309877 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-swiftconf\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.321631 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-combined-ca-bundle\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.325504 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxsrj\" (UniqueName: \"kubernetes.io/projected/a809020b-dd80-4507-9aae-e88c2784d96c-kube-api-access-lxsrj\") pod \"swift-ring-rebalance-debug-wxz4w\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 09:59:59 crc kubenswrapper[4978]: I0225 09:59:59.441179 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.158439 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-debug-wxz4w"] Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.177474 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533560-464p8"] Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.178925 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533560-464p8" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.181224 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.181426 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.181600 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.195127 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9"] Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.196751 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.198553 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.199527 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.225832 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533560-464p8"] Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.251784 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9"] Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.323609 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrpjm\" (UniqueName: \"kubernetes.io/projected/4bd2ef26-a6e6-4787-b8e9-96841e04ae52-kube-api-access-rrpjm\") pod \"auto-csr-approver-29533560-464p8\" (UID: \"4bd2ef26-a6e6-4787-b8e9-96841e04ae52\") " pod="openshift-infra/auto-csr-approver-29533560-464p8" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.323688 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd264555-1c2d-4563-967b-a453dd24b078-config-volume\") pod \"collect-profiles-29533560-5ndm9\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.323711 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9s89\" (UniqueName: \"kubernetes.io/projected/bd264555-1c2d-4563-967b-a453dd24b078-kube-api-access-m9s89\") pod \"collect-profiles-29533560-5ndm9\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.323883 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd264555-1c2d-4563-967b-a453dd24b078-secret-volume\") pod \"collect-profiles-29533560-5ndm9\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.426381 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrpjm\" (UniqueName: \"kubernetes.io/projected/4bd2ef26-a6e6-4787-b8e9-96841e04ae52-kube-api-access-rrpjm\") pod \"auto-csr-approver-29533560-464p8\" (UID: \"4bd2ef26-a6e6-4787-b8e9-96841e04ae52\") " pod="openshift-infra/auto-csr-approver-29533560-464p8" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.426534 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd264555-1c2d-4563-967b-a453dd24b078-config-volume\") pod \"collect-profiles-29533560-5ndm9\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.426566 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9s89\" (UniqueName: \"kubernetes.io/projected/bd264555-1c2d-4563-967b-a453dd24b078-kube-api-access-m9s89\") pod \"collect-profiles-29533560-5ndm9\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.426596 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd264555-1c2d-4563-967b-a453dd24b078-secret-volume\") pod \"collect-profiles-29533560-5ndm9\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.427926 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd264555-1c2d-4563-967b-a453dd24b078-config-volume\") pod \"collect-profiles-29533560-5ndm9\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.431274 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd264555-1c2d-4563-967b-a453dd24b078-secret-volume\") pod \"collect-profiles-29533560-5ndm9\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.447834 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9s89\" (UniqueName: \"kubernetes.io/projected/bd264555-1c2d-4563-967b-a453dd24b078-kube-api-access-m9s89\") pod \"collect-profiles-29533560-5ndm9\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.449495 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrpjm\" (UniqueName: \"kubernetes.io/projected/4bd2ef26-a6e6-4787-b8e9-96841e04ae52-kube-api-access-rrpjm\") pod \"auto-csr-approver-29533560-464p8\" (UID: \"4bd2ef26-a6e6-4787-b8e9-96841e04ae52\") " pod="openshift-infra/auto-csr-approver-29533560-464p8" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.539263 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533560-464p8" Feb 25 10:00:00 crc kubenswrapper[4978]: I0225 10:00:00.549703 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:01 crc kubenswrapper[4978]: I0225 10:00:01.180700 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-wxz4w" event={"ID":"a809020b-dd80-4507-9aae-e88c2784d96c","Type":"ContainerStarted","Data":"14de3b94951640888fd04512eeb7a4fd8d2049646328284d8507fe9c04ae56a6"} Feb 25 10:00:01 crc kubenswrapper[4978]: I0225 10:00:01.181020 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-wxz4w" event={"ID":"a809020b-dd80-4507-9aae-e88c2784d96c","Type":"ContainerStarted","Data":"2d0f49991163390cb6f2e563338ad80e673bf821b267496a5e8e29eeecee32cb"} Feb 25 10:00:01 crc kubenswrapper[4978]: I0225 10:00:01.207362 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-debug-wxz4w" podStartSLOduration=2.207329309 podStartE2EDuration="2.207329309s" podCreationTimestamp="2026-02-25 09:59:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 10:00:01.196089599 +0000 UTC m=+11694.635346068" watchObservedRunningTime="2026-02-25 10:00:01.207329309 +0000 UTC m=+11694.646585808" Feb 25 10:00:01 crc kubenswrapper[4978]: I0225 10:00:01.319518 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9"] Feb 25 10:00:01 crc kubenswrapper[4978]: W0225 10:00:01.320267 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd264555_1c2d_4563_967b_a453dd24b078.slice/crio-a62628aea14d746fa237036c61f5f85b8434c9607e50938fc8a68a0258eed446 WatchSource:0}: Error finding container a62628aea14d746fa237036c61f5f85b8434c9607e50938fc8a68a0258eed446: Status 404 returned error can't find the container with id a62628aea14d746fa237036c61f5f85b8434c9607e50938fc8a68a0258eed446 Feb 25 10:00:01 crc kubenswrapper[4978]: I0225 10:00:01.738426 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533560-464p8"] Feb 25 10:00:01 crc kubenswrapper[4978]: W0225 10:00:01.745439 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bd2ef26_a6e6_4787_b8e9_96841e04ae52.slice/crio-fe65fc7b625ecda98aafb9b777dffc88c4ef23b7e9e25008f369bda8792f4643 WatchSource:0}: Error finding container fe65fc7b625ecda98aafb9b777dffc88c4ef23b7e9e25008f369bda8792f4643: Status 404 returned error can't find the container with id fe65fc7b625ecda98aafb9b777dffc88c4ef23b7e9e25008f369bda8792f4643 Feb 25 10:00:02 crc kubenswrapper[4978]: I0225 10:00:02.197103 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533560-464p8" event={"ID":"4bd2ef26-a6e6-4787-b8e9-96841e04ae52","Type":"ContainerStarted","Data":"fe65fc7b625ecda98aafb9b777dffc88c4ef23b7e9e25008f369bda8792f4643"} Feb 25 10:00:02 crc kubenswrapper[4978]: I0225 10:00:02.199115 4978 generic.go:334] "Generic (PLEG): container finished" podID="bd264555-1c2d-4563-967b-a453dd24b078" containerID="844ff351a871d3480ced99a6eafd082c661e8055328eda2b5c5048973b3e4522" exitCode=0 Feb 25 10:00:02 crc kubenswrapper[4978]: I0225 10:00:02.199720 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" event={"ID":"bd264555-1c2d-4563-967b-a453dd24b078","Type":"ContainerDied","Data":"844ff351a871d3480ced99a6eafd082c661e8055328eda2b5c5048973b3e4522"} Feb 25 10:00:02 crc kubenswrapper[4978]: I0225 10:00:02.199746 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" event={"ID":"bd264555-1c2d-4563-967b-a453dd24b078","Type":"ContainerStarted","Data":"a62628aea14d746fa237036c61f5f85b8434c9607e50938fc8a68a0258eed446"} Feb 25 10:00:04 crc kubenswrapper[4978]: I0225 10:00:04.608026 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:04 crc kubenswrapper[4978]: I0225 10:00:04.718143 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9s89\" (UniqueName: \"kubernetes.io/projected/bd264555-1c2d-4563-967b-a453dd24b078-kube-api-access-m9s89\") pod \"bd264555-1c2d-4563-967b-a453dd24b078\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " Feb 25 10:00:04 crc kubenswrapper[4978]: I0225 10:00:04.718252 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd264555-1c2d-4563-967b-a453dd24b078-secret-volume\") pod \"bd264555-1c2d-4563-967b-a453dd24b078\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " Feb 25 10:00:04 crc kubenswrapper[4978]: I0225 10:00:04.718378 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd264555-1c2d-4563-967b-a453dd24b078-config-volume\") pod \"bd264555-1c2d-4563-967b-a453dd24b078\" (UID: \"bd264555-1c2d-4563-967b-a453dd24b078\") " Feb 25 10:00:04 crc kubenswrapper[4978]: I0225 10:00:04.719148 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd264555-1c2d-4563-967b-a453dd24b078-config-volume" (OuterVolumeSpecName: "config-volume") pod "bd264555-1c2d-4563-967b-a453dd24b078" (UID: "bd264555-1c2d-4563-967b-a453dd24b078"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 10:00:04 crc kubenswrapper[4978]: I0225 10:00:04.727021 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd264555-1c2d-4563-967b-a453dd24b078-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bd264555-1c2d-4563-967b-a453dd24b078" (UID: "bd264555-1c2d-4563-967b-a453dd24b078"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:00:04 crc kubenswrapper[4978]: I0225 10:00:04.727172 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd264555-1c2d-4563-967b-a453dd24b078-kube-api-access-m9s89" (OuterVolumeSpecName: "kube-api-access-m9s89") pod "bd264555-1c2d-4563-967b-a453dd24b078" (UID: "bd264555-1c2d-4563-967b-a453dd24b078"). InnerVolumeSpecName "kube-api-access-m9s89". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:00:04 crc kubenswrapper[4978]: I0225 10:00:04.821280 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bd264555-1c2d-4563-967b-a453dd24b078-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:04 crc kubenswrapper[4978]: I0225 10:00:04.821341 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bd264555-1c2d-4563-967b-a453dd24b078-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:04 crc kubenswrapper[4978]: I0225 10:00:04.821356 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9s89\" (UniqueName: \"kubernetes.io/projected/bd264555-1c2d-4563-967b-a453dd24b078-kube-api-access-m9s89\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:05 crc kubenswrapper[4978]: I0225 10:00:05.228800 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533560-464p8" event={"ID":"4bd2ef26-a6e6-4787-b8e9-96841e04ae52","Type":"ContainerStarted","Data":"f5a38eca26fd819b52414a41986f336e8b7b8ba03e039bdd58f1d4a015bfd00e"} Feb 25 10:00:05 crc kubenswrapper[4978]: I0225 10:00:05.233206 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" event={"ID":"bd264555-1c2d-4563-967b-a453dd24b078","Type":"ContainerDied","Data":"a62628aea14d746fa237036c61f5f85b8434c9607e50938fc8a68a0258eed446"} Feb 25 10:00:05 crc kubenswrapper[4978]: I0225 10:00:05.233283 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a62628aea14d746fa237036c61f5f85b8434c9607e50938fc8a68a0258eed446" Feb 25 10:00:05 crc kubenswrapper[4978]: I0225 10:00:05.233230 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533560-5ndm9" Feb 25 10:00:05 crc kubenswrapper[4978]: I0225 10:00:05.260328 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533560-464p8" podStartSLOduration=2.3053722629999998 podStartE2EDuration="5.260307697s" podCreationTimestamp="2026-02-25 10:00:00 +0000 UTC" firstStartedPulling="2026-02-25 10:00:01.747400672 +0000 UTC m=+11695.186657141" lastFinishedPulling="2026-02-25 10:00:04.702336116 +0000 UTC m=+11698.141592575" observedRunningTime="2026-02-25 10:00:05.252068331 +0000 UTC m=+11698.691324790" watchObservedRunningTime="2026-02-25 10:00:05.260307697 +0000 UTC m=+11698.699564156" Feb 25 10:00:05 crc kubenswrapper[4978]: I0225 10:00:05.680544 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv"] Feb 25 10:00:05 crc kubenswrapper[4978]: I0225 10:00:05.694545 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533515-tfbtv"] Feb 25 10:00:06 crc kubenswrapper[4978]: I0225 10:00:06.243587 4978 generic.go:334] "Generic (PLEG): container finished" podID="4bd2ef26-a6e6-4787-b8e9-96841e04ae52" containerID="f5a38eca26fd819b52414a41986f336e8b7b8ba03e039bdd58f1d4a015bfd00e" exitCode=0 Feb 25 10:00:06 crc kubenswrapper[4978]: I0225 10:00:06.243683 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533560-464p8" event={"ID":"4bd2ef26-a6e6-4787-b8e9-96841e04ae52","Type":"ContainerDied","Data":"f5a38eca26fd819b52414a41986f336e8b7b8ba03e039bdd58f1d4a015bfd00e"} Feb 25 10:00:07 crc kubenswrapper[4978]: I0225 10:00:07.253960 4978 generic.go:334] "Generic (PLEG): container finished" podID="a809020b-dd80-4507-9aae-e88c2784d96c" containerID="14de3b94951640888fd04512eeb7a4fd8d2049646328284d8507fe9c04ae56a6" exitCode=0 Feb 25 10:00:07 crc kubenswrapper[4978]: I0225 10:00:07.254247 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-wxz4w" event={"ID":"a809020b-dd80-4507-9aae-e88c2784d96c","Type":"ContainerDied","Data":"14de3b94951640888fd04512eeb7a4fd8d2049646328284d8507fe9c04ae56a6"} Feb 25 10:00:07 crc kubenswrapper[4978]: I0225 10:00:07.345987 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a03ab396-abf8-41a4-a30d-b4698113a688" path="/var/lib/kubelet/pods/a03ab396-abf8-41a4-a30d-b4698113a688/volumes" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.107281 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533560-464p8" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.214328 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrpjm\" (UniqueName: \"kubernetes.io/projected/4bd2ef26-a6e6-4787-b8e9-96841e04ae52-kube-api-access-rrpjm\") pod \"4bd2ef26-a6e6-4787-b8e9-96841e04ae52\" (UID: \"4bd2ef26-a6e6-4787-b8e9-96841e04ae52\") " Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.220039 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bd2ef26-a6e6-4787-b8e9-96841e04ae52-kube-api-access-rrpjm" (OuterVolumeSpecName: "kube-api-access-rrpjm") pod "4bd2ef26-a6e6-4787-b8e9-96841e04ae52" (UID: "4bd2ef26-a6e6-4787-b8e9-96841e04ae52"). InnerVolumeSpecName "kube-api-access-rrpjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.308526 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533560-464p8" event={"ID":"4bd2ef26-a6e6-4787-b8e9-96841e04ae52","Type":"ContainerDied","Data":"fe65fc7b625ecda98aafb9b777dffc88c4ef23b7e9e25008f369bda8792f4643"} Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.308589 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe65fc7b625ecda98aafb9b777dffc88c4ef23b7e9e25008f369bda8792f4643" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.308558 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533560-464p8" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.310843 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-debug-wxz4w" event={"ID":"a809020b-dd80-4507-9aae-e88c2784d96c","Type":"ContainerDied","Data":"2d0f49991163390cb6f2e563338ad80e673bf821b267496a5e8e29eeecee32cb"} Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.310888 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d0f49991163390cb6f2e563338ad80e673bf821b267496a5e8e29eeecee32cb" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.316488 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrpjm\" (UniqueName: \"kubernetes.io/projected/4bd2ef26-a6e6-4787-b8e9-96841e04ae52-kube-api-access-rrpjm\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.332764 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.386122 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-debug-wxz4w"] Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.401415 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-debug-wxz4w"] Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.417823 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-combined-ca-bundle\") pod \"a809020b-dd80-4507-9aae-e88c2784d96c\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.417940 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-scripts\") pod \"a809020b-dd80-4507-9aae-e88c2784d96c\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.418106 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-swiftconf\") pod \"a809020b-dd80-4507-9aae-e88c2784d96c\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.418150 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-ring-data-devices\") pod \"a809020b-dd80-4507-9aae-e88c2784d96c\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.418189 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxsrj\" (UniqueName: \"kubernetes.io/projected/a809020b-dd80-4507-9aae-e88c2784d96c-kube-api-access-lxsrj\") pod \"a809020b-dd80-4507-9aae-e88c2784d96c\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.418233 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a809020b-dd80-4507-9aae-e88c2784d96c-etc-swift\") pod \"a809020b-dd80-4507-9aae-e88c2784d96c\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.418259 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-dispersionconf\") pod \"a809020b-dd80-4507-9aae-e88c2784d96c\" (UID: \"a809020b-dd80-4507-9aae-e88c2784d96c\") " Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.419078 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a809020b-dd80-4507-9aae-e88c2784d96c" (UID: "a809020b-dd80-4507-9aae-e88c2784d96c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.420209 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a809020b-dd80-4507-9aae-e88c2784d96c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a809020b-dd80-4507-9aae-e88c2784d96c" (UID: "a809020b-dd80-4507-9aae-e88c2784d96c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.422376 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a809020b-dd80-4507-9aae-e88c2784d96c-kube-api-access-lxsrj" (OuterVolumeSpecName: "kube-api-access-lxsrj") pod "a809020b-dd80-4507-9aae-e88c2784d96c" (UID: "a809020b-dd80-4507-9aae-e88c2784d96c"). InnerVolumeSpecName "kube-api-access-lxsrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.453751 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a809020b-dd80-4507-9aae-e88c2784d96c" (UID: "a809020b-dd80-4507-9aae-e88c2784d96c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.471731 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a809020b-dd80-4507-9aae-e88c2784d96c" (UID: "a809020b-dd80-4507-9aae-e88c2784d96c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.471798 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-scripts" (OuterVolumeSpecName: "scripts") pod "a809020b-dd80-4507-9aae-e88c2784d96c" (UID: "a809020b-dd80-4507-9aae-e88c2784d96c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.491841 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a809020b-dd80-4507-9aae-e88c2784d96c" (UID: "a809020b-dd80-4507-9aae-e88c2784d96c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.521023 4978 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.521050 4978 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.521061 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxsrj\" (UniqueName: \"kubernetes.io/projected/a809020b-dd80-4507-9aae-e88c2784d96c-kube-api-access-lxsrj\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.521069 4978 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a809020b-dd80-4507-9aae-e88c2784d96c-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.521077 4978 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.521087 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a809020b-dd80-4507-9aae-e88c2784d96c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:09 crc kubenswrapper[4978]: I0225 10:00:09.521094 4978 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a809020b-dd80-4507-9aae-e88c2784d96c-scripts\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:10 crc kubenswrapper[4978]: I0225 10:00:10.173138 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533554-2g77f"] Feb 25 10:00:10 crc kubenswrapper[4978]: I0225 10:00:10.183584 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533554-2g77f"] Feb 25 10:00:10 crc kubenswrapper[4978]: I0225 10:00:10.320727 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-debug-wxz4w" Feb 25 10:00:11 crc kubenswrapper[4978]: I0225 10:00:11.339907 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d399ffb-1ec7-4b5f-a547-99a9c0b81e09" path="/var/lib/kubelet/pods/5d399ffb-1ec7-4b5f-a547-99a9c0b81e09/volumes" Feb 25 10:00:11 crc kubenswrapper[4978]: I0225 10:00:11.342414 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a809020b-dd80-4507-9aae-e88c2784d96c" path="/var/lib/kubelet/pods/a809020b-dd80-4507-9aae-e88c2784d96c/volumes" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.062530 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zpwpb"] Feb 25 10:00:13 crc kubenswrapper[4978]: E0225 10:00:13.063197 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bd2ef26-a6e6-4787-b8e9-96841e04ae52" containerName="oc" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.063208 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bd2ef26-a6e6-4787-b8e9-96841e04ae52" containerName="oc" Feb 25 10:00:13 crc kubenswrapper[4978]: E0225 10:00:13.063236 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a809020b-dd80-4507-9aae-e88c2784d96c" containerName="swift-ring-rebalance" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.063242 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a809020b-dd80-4507-9aae-e88c2784d96c" containerName="swift-ring-rebalance" Feb 25 10:00:13 crc kubenswrapper[4978]: E0225 10:00:13.063255 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd264555-1c2d-4563-967b-a453dd24b078" containerName="collect-profiles" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.063260 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd264555-1c2d-4563-967b-a453dd24b078" containerName="collect-profiles" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.063495 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bd2ef26-a6e6-4787-b8e9-96841e04ae52" containerName="oc" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.063518 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a809020b-dd80-4507-9aae-e88c2784d96c" containerName="swift-ring-rebalance" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.063538 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd264555-1c2d-4563-967b-a453dd24b078" containerName="collect-profiles" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.064963 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.094404 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zpwpb"] Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.108202 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-utilities\") pod \"certified-operators-zpwpb\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.108280 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crnc7\" (UniqueName: \"kubernetes.io/projected/743688fc-e633-4097-8e56-d395510600e5-kube-api-access-crnc7\") pod \"certified-operators-zpwpb\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.108321 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-catalog-content\") pod \"certified-operators-zpwpb\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.209968 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-catalog-content\") pod \"certified-operators-zpwpb\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.210109 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-utilities\") pod \"certified-operators-zpwpb\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.210716 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-utilities\") pod \"certified-operators-zpwpb\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.210770 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-catalog-content\") pod \"certified-operators-zpwpb\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.210778 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crnc7\" (UniqueName: \"kubernetes.io/projected/743688fc-e633-4097-8e56-d395510600e5-kube-api-access-crnc7\") pod \"certified-operators-zpwpb\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.233226 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crnc7\" (UniqueName: \"kubernetes.io/projected/743688fc-e633-4097-8e56-d395510600e5-kube-api-access-crnc7\") pod \"certified-operators-zpwpb\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:13 crc kubenswrapper[4978]: I0225 10:00:13.437008 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:14 crc kubenswrapper[4978]: I0225 10:00:14.437148 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zpwpb"] Feb 25 10:00:15 crc kubenswrapper[4978]: I0225 10:00:15.376625 4978 generic.go:334] "Generic (PLEG): container finished" podID="743688fc-e633-4097-8e56-d395510600e5" containerID="1fb3f03cd980e206d66d2a53389d656fe64fb4d465a658765a75c564ef2fe2d7" exitCode=0 Feb 25 10:00:15 crc kubenswrapper[4978]: I0225 10:00:15.376730 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpwpb" event={"ID":"743688fc-e633-4097-8e56-d395510600e5","Type":"ContainerDied","Data":"1fb3f03cd980e206d66d2a53389d656fe64fb4d465a658765a75c564ef2fe2d7"} Feb 25 10:00:15 crc kubenswrapper[4978]: I0225 10:00:15.377003 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpwpb" event={"ID":"743688fc-e633-4097-8e56-d395510600e5","Type":"ContainerStarted","Data":"9e6447545e2a8d6daa7ec583254568ec9bb536ac4413dae1daf8bc9e060b8945"} Feb 25 10:00:16 crc kubenswrapper[4978]: I0225 10:00:16.391202 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpwpb" event={"ID":"743688fc-e633-4097-8e56-d395510600e5","Type":"ContainerStarted","Data":"39bb74fea68a18f2c36d136fe34cf2aaf3a3f7dd272f3dd8fead59da17431ece"} Feb 25 10:00:16 crc kubenswrapper[4978]: I0225 10:00:16.539819 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:00:16 crc kubenswrapper[4978]: I0225 10:00:16.539875 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:00:16 crc kubenswrapper[4978]: E0225 10:00:16.683641 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod743688fc_e633_4097_8e56_d395510600e5.slice/crio-conmon-39bb74fea68a18f2c36d136fe34cf2aaf3a3f7dd272f3dd8fead59da17431ece.scope\": RecentStats: unable to find data in memory cache]" Feb 25 10:00:17 crc kubenswrapper[4978]: I0225 10:00:17.412780 4978 generic.go:334] "Generic (PLEG): container finished" podID="743688fc-e633-4097-8e56-d395510600e5" containerID="39bb74fea68a18f2c36d136fe34cf2aaf3a3f7dd272f3dd8fead59da17431ece" exitCode=0 Feb 25 10:00:17 crc kubenswrapper[4978]: I0225 10:00:17.412859 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpwpb" event={"ID":"743688fc-e633-4097-8e56-d395510600e5","Type":"ContainerDied","Data":"39bb74fea68a18f2c36d136fe34cf2aaf3a3f7dd272f3dd8fead59da17431ece"} Feb 25 10:00:18 crc kubenswrapper[4978]: I0225 10:00:18.439334 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpwpb" event={"ID":"743688fc-e633-4097-8e56-d395510600e5","Type":"ContainerStarted","Data":"5e58835baf50a6b076ed665afb0858b9a83d592caf210752d1e07f3782aa0f45"} Feb 25 10:00:18 crc kubenswrapper[4978]: I0225 10:00:18.471845 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zpwpb" podStartSLOduration=3.057822485 podStartE2EDuration="5.471818009s" podCreationTimestamp="2026-02-25 10:00:13 +0000 UTC" firstStartedPulling="2026-02-25 10:00:15.379163885 +0000 UTC m=+11708.818420354" lastFinishedPulling="2026-02-25 10:00:17.793159379 +0000 UTC m=+11711.232415878" observedRunningTime="2026-02-25 10:00:18.464876893 +0000 UTC m=+11711.904133432" watchObservedRunningTime="2026-02-25 10:00:18.471818009 +0000 UTC m=+11711.911074508" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.012734 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.016184 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.020058 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fm969" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.022110 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.022777 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.032919 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.033279 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.054003 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.054052 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxr2m\" (UniqueName: \"kubernetes.io/projected/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-kube-api-access-fxr2m\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.054098 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.054128 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.054155 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.054213 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-config-data\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.054237 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.054305 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.054411 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.155987 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.156075 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.156103 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxr2m\" (UniqueName: \"kubernetes.io/projected/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-kube-api-access-fxr2m\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.156131 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.156151 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.156171 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.156199 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-config-data\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.156216 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.156268 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.156776 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.158358 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.161153 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.161703 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-config-data\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.161724 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.163568 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.169403 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.174896 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxr2m\" (UniqueName: \"kubernetes.io/projected/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-kube-api-access-fxr2m\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.178467 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.233232 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.394034 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.438037 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.438329 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:23 crc kubenswrapper[4978]: I0225 10:00:23.531244 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:24 crc kubenswrapper[4978]: I0225 10:00:24.115697 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 25 10:00:24 crc kubenswrapper[4978]: I0225 10:00:24.535747 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd","Type":"ContainerStarted","Data":"09461336f7e6ab390807996a0637070f0954dbc79bf0245695b4a6000fce7e24"} Feb 25 10:00:24 crc kubenswrapper[4978]: I0225 10:00:24.610704 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:24 crc kubenswrapper[4978]: I0225 10:00:24.677681 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zpwpb"] Feb 25 10:00:26 crc kubenswrapper[4978]: I0225 10:00:26.556613 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zpwpb" podUID="743688fc-e633-4097-8e56-d395510600e5" containerName="registry-server" containerID="cri-o://5e58835baf50a6b076ed665afb0858b9a83d592caf210752d1e07f3782aa0f45" gracePeriod=2 Feb 25 10:00:27 crc kubenswrapper[4978]: I0225 10:00:27.574247 4978 generic.go:334] "Generic (PLEG): container finished" podID="743688fc-e633-4097-8e56-d395510600e5" containerID="5e58835baf50a6b076ed665afb0858b9a83d592caf210752d1e07f3782aa0f45" exitCode=0 Feb 25 10:00:27 crc kubenswrapper[4978]: I0225 10:00:27.574322 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpwpb" event={"ID":"743688fc-e633-4097-8e56-d395510600e5","Type":"ContainerDied","Data":"5e58835baf50a6b076ed665afb0858b9a83d592caf210752d1e07f3782aa0f45"} Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.060287 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.074263 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crnc7\" (UniqueName: \"kubernetes.io/projected/743688fc-e633-4097-8e56-d395510600e5-kube-api-access-crnc7\") pod \"743688fc-e633-4097-8e56-d395510600e5\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.074707 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-catalog-content\") pod \"743688fc-e633-4097-8e56-d395510600e5\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.074853 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-utilities\") pod \"743688fc-e633-4097-8e56-d395510600e5\" (UID: \"743688fc-e633-4097-8e56-d395510600e5\") " Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.076295 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-utilities" (OuterVolumeSpecName: "utilities") pod "743688fc-e633-4097-8e56-d395510600e5" (UID: "743688fc-e633-4097-8e56-d395510600e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.114015 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/743688fc-e633-4097-8e56-d395510600e5-kube-api-access-crnc7" (OuterVolumeSpecName: "kube-api-access-crnc7") pod "743688fc-e633-4097-8e56-d395510600e5" (UID: "743688fc-e633-4097-8e56-d395510600e5"). InnerVolumeSpecName "kube-api-access-crnc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.171758 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "743688fc-e633-4097-8e56-d395510600e5" (UID: "743688fc-e633-4097-8e56-d395510600e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.177427 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crnc7\" (UniqueName: \"kubernetes.io/projected/743688fc-e633-4097-8e56-d395510600e5-kube-api-access-crnc7\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.177461 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.177472 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743688fc-e633-4097-8e56-d395510600e5-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.587260 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpwpb" event={"ID":"743688fc-e633-4097-8e56-d395510600e5","Type":"ContainerDied","Data":"9e6447545e2a8d6daa7ec583254568ec9bb536ac4413dae1daf8bc9e060b8945"} Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.587309 4978 scope.go:117] "RemoveContainer" containerID="5e58835baf50a6b076ed665afb0858b9a83d592caf210752d1e07f3782aa0f45" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.587376 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpwpb" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.625857 4978 scope.go:117] "RemoveContainer" containerID="39bb74fea68a18f2c36d136fe34cf2aaf3a3f7dd272f3dd8fead59da17431ece" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.637539 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zpwpb"] Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.647609 4978 scope.go:117] "RemoveContainer" containerID="1fb3f03cd980e206d66d2a53389d656fe64fb4d465a658765a75c564ef2fe2d7" Feb 25 10:00:28 crc kubenswrapper[4978]: I0225 10:00:28.647773 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zpwpb"] Feb 25 10:00:29 crc kubenswrapper[4978]: I0225 10:00:29.346281 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="743688fc-e633-4097-8e56-d395510600e5" path="/var/lib/kubelet/pods/743688fc-e633-4097-8e56-d395510600e5/volumes" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.148145 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d7qwh"] Feb 25 10:00:41 crc kubenswrapper[4978]: E0225 10:00:41.149849 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743688fc-e633-4097-8e56-d395510600e5" containerName="registry-server" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.149868 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="743688fc-e633-4097-8e56-d395510600e5" containerName="registry-server" Feb 25 10:00:41 crc kubenswrapper[4978]: E0225 10:00:41.149913 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743688fc-e633-4097-8e56-d395510600e5" containerName="extract-utilities" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.149923 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="743688fc-e633-4097-8e56-d395510600e5" containerName="extract-utilities" Feb 25 10:00:41 crc kubenswrapper[4978]: E0225 10:00:41.149941 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743688fc-e633-4097-8e56-d395510600e5" containerName="extract-content" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.149950 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="743688fc-e633-4097-8e56-d395510600e5" containerName="extract-content" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.150286 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="743688fc-e633-4097-8e56-d395510600e5" containerName="registry-server" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.152474 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.163946 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d7qwh"] Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.186875 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-catalog-content\") pod \"redhat-marketplace-d7qwh\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.186937 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnrpg\" (UniqueName: \"kubernetes.io/projected/02b2c7fa-5935-470e-b349-63993285f7c4-kube-api-access-lnrpg\") pod \"redhat-marketplace-d7qwh\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.187046 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-utilities\") pod \"redhat-marketplace-d7qwh\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.288793 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-catalog-content\") pod \"redhat-marketplace-d7qwh\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.288839 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnrpg\" (UniqueName: \"kubernetes.io/projected/02b2c7fa-5935-470e-b349-63993285f7c4-kube-api-access-lnrpg\") pod \"redhat-marketplace-d7qwh\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.288877 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-utilities\") pod \"redhat-marketplace-d7qwh\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.289393 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-utilities\") pod \"redhat-marketplace-d7qwh\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.289439 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-catalog-content\") pod \"redhat-marketplace-d7qwh\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.312058 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnrpg\" (UniqueName: \"kubernetes.io/projected/02b2c7fa-5935-470e-b349-63993285f7c4-kube-api-access-lnrpg\") pod \"redhat-marketplace-d7qwh\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.443170 4978 scope.go:117] "RemoveContainer" containerID="0068108cea530d5e9576bb600995e2e590d7822b5923cf3a2c80cb203aa2ad9f" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.486156 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.533164 4978 scope.go:117] "RemoveContainer" containerID="fd9a38a61a73f52412f0cf6ce127062e73c93e15ffa38550c9a8fb0111d09840" Feb 25 10:00:41 crc kubenswrapper[4978]: I0225 10:00:41.647487 4978 scope.go:117] "RemoveContainer" containerID="f02b504744df9f5d676874c9f2715fe0811c7c258ddc559d2e4aba9b09dd7e0b" Feb 25 10:00:42 crc kubenswrapper[4978]: I0225 10:00:42.364693 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d7qwh"] Feb 25 10:00:42 crc kubenswrapper[4978]: W0225 10:00:42.377890 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02b2c7fa_5935_470e_b349_63993285f7c4.slice/crio-9ad9dcd5663e25ed66a25e8cb3f12123f8b2bd179e7a4b31e58e5481f6dbddcd WatchSource:0}: Error finding container 9ad9dcd5663e25ed66a25e8cb3f12123f8b2bd179e7a4b31e58e5481f6dbddcd: Status 404 returned error can't find the container with id 9ad9dcd5663e25ed66a25e8cb3f12123f8b2bd179e7a4b31e58e5481f6dbddcd Feb 25 10:00:42 crc kubenswrapper[4978]: I0225 10:00:42.776247 4978 generic.go:334] "Generic (PLEG): container finished" podID="02b2c7fa-5935-470e-b349-63993285f7c4" containerID="1308f379631b6d2212c95ba3277d27e2ae3c7619554aa2fe308b702030c0ceaa" exitCode=0 Feb 25 10:00:42 crc kubenswrapper[4978]: I0225 10:00:42.776502 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d7qwh" event={"ID":"02b2c7fa-5935-470e-b349-63993285f7c4","Type":"ContainerDied","Data":"1308f379631b6d2212c95ba3277d27e2ae3c7619554aa2fe308b702030c0ceaa"} Feb 25 10:00:42 crc kubenswrapper[4978]: I0225 10:00:42.776587 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d7qwh" event={"ID":"02b2c7fa-5935-470e-b349-63993285f7c4","Type":"ContainerStarted","Data":"9ad9dcd5663e25ed66a25e8cb3f12123f8b2bd179e7a4b31e58e5481f6dbddcd"} Feb 25 10:00:44 crc kubenswrapper[4978]: I0225 10:00:44.802328 4978 generic.go:334] "Generic (PLEG): container finished" podID="02b2c7fa-5935-470e-b349-63993285f7c4" containerID="129f62157fc102e7bee5da8bfa065510cbc44436f7de641aa33971f8f109b3d9" exitCode=0 Feb 25 10:00:44 crc kubenswrapper[4978]: I0225 10:00:44.802457 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d7qwh" event={"ID":"02b2c7fa-5935-470e-b349-63993285f7c4","Type":"ContainerDied","Data":"129f62157fc102e7bee5da8bfa065510cbc44436f7de641aa33971f8f109b3d9"} Feb 25 10:00:45 crc kubenswrapper[4978]: I0225 10:00:45.818458 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d7qwh" event={"ID":"02b2c7fa-5935-470e-b349-63993285f7c4","Type":"ContainerStarted","Data":"f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90"} Feb 25 10:00:45 crc kubenswrapper[4978]: I0225 10:00:45.843027 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d7qwh" podStartSLOduration=2.268778324 podStartE2EDuration="4.84300747s" podCreationTimestamp="2026-02-25 10:00:41 +0000 UTC" firstStartedPulling="2026-02-25 10:00:42.778234084 +0000 UTC m=+11736.217490543" lastFinishedPulling="2026-02-25 10:00:45.35246323 +0000 UTC m=+11738.791719689" observedRunningTime="2026-02-25 10:00:45.839542792 +0000 UTC m=+11739.278799251" watchObservedRunningTime="2026-02-25 10:00:45.84300747 +0000 UTC m=+11739.282263929" Feb 25 10:00:46 crc kubenswrapper[4978]: I0225 10:00:46.540518 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:00:46 crc kubenswrapper[4978]: I0225 10:00:46.540800 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:00:51 crc kubenswrapper[4978]: I0225 10:00:51.488826 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:51 crc kubenswrapper[4978]: I0225 10:00:51.489435 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:51 crc kubenswrapper[4978]: I0225 10:00:51.553822 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:51 crc kubenswrapper[4978]: I0225 10:00:51.950904 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:00:52 crc kubenswrapper[4978]: I0225 10:00:52.014308 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d7qwh"] Feb 25 10:00:53 crc kubenswrapper[4978]: I0225 10:00:53.921713 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d7qwh" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" containerName="registry-server" containerID="cri-o://f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90" gracePeriod=2 Feb 25 10:00:54 crc kubenswrapper[4978]: I0225 10:00:54.935865 4978 generic.go:334] "Generic (PLEG): container finished" podID="02b2c7fa-5935-470e-b349-63993285f7c4" containerID="f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90" exitCode=0 Feb 25 10:00:54 crc kubenswrapper[4978]: I0225 10:00:54.935903 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d7qwh" event={"ID":"02b2c7fa-5935-470e-b349-63993285f7c4","Type":"ContainerDied","Data":"f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90"} Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.189654 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29533561-phzdw"] Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.193987 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.213810 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29533561-phzdw"] Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.319279 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-combined-ca-bundle\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.319333 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-config-data\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.319605 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh78l\" (UniqueName: \"kubernetes.io/projected/434f88a0-9d17-474f-a877-35569aba520b-kube-api-access-vh78l\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.319808 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-fernet-keys\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.421583 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh78l\" (UniqueName: \"kubernetes.io/projected/434f88a0-9d17-474f-a877-35569aba520b-kube-api-access-vh78l\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.421917 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-fernet-keys\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.422841 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-combined-ca-bundle\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.422897 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-config-data\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.428586 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-fernet-keys\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.435017 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-combined-ca-bundle\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.436089 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-config-data\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.441736 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh78l\" (UniqueName: \"kubernetes.io/projected/434f88a0-9d17-474f-a877-35569aba520b-kube-api-access-vh78l\") pod \"keystone-cron-29533561-phzdw\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:00 crc kubenswrapper[4978]: I0225 10:01:00.516767 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:01 crc kubenswrapper[4978]: E0225 10:01:01.488706 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90 is running failed: container process not found" containerID="f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90" cmd=["grpc_health_probe","-addr=:50051"] Feb 25 10:01:01 crc kubenswrapper[4978]: E0225 10:01:01.489712 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90 is running failed: container process not found" containerID="f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90" cmd=["grpc_health_probe","-addr=:50051"] Feb 25 10:01:01 crc kubenswrapper[4978]: E0225 10:01:01.490286 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90 is running failed: container process not found" containerID="f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90" cmd=["grpc_health_probe","-addr=:50051"] Feb 25 10:01:01 crc kubenswrapper[4978]: E0225 10:01:01.490329 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-d7qwh" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" containerName="registry-server" Feb 25 10:01:11 crc kubenswrapper[4978]: E0225 10:01:11.488662 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90 is running failed: container process not found" containerID="f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90" cmd=["grpc_health_probe","-addr=:50051"] Feb 25 10:01:11 crc kubenswrapper[4978]: E0225 10:01:11.489506 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90 is running failed: container process not found" containerID="f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90" cmd=["grpc_health_probe","-addr=:50051"] Feb 25 10:01:11 crc kubenswrapper[4978]: E0225 10:01:11.489851 4978 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90 is running failed: container process not found" containerID="f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90" cmd=["grpc_health_probe","-addr=:50051"] Feb 25 10:01:11 crc kubenswrapper[4978]: E0225 10:01:11.489910 4978 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-d7qwh" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" containerName="registry-server" Feb 25 10:01:16 crc kubenswrapper[4978]: E0225 10:01:16.104281 4978 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:4198fe0138d2c46ae51eb6e53fed5006" Feb 25 10:01:16 crc kubenswrapper[4978]: E0225 10:01:16.104949 4978 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:4198fe0138d2c46ae51eb6e53fed5006" Feb 25 10:01:16 crc kubenswrapper[4978]: E0225 10:01:16.106330 4978 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:4198fe0138d2c46ae51eb6e53fed5006,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fxr2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(ec6773fc-9a0a-4d54-a950-85b5be1fe6cd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 25 10:01:16 crc kubenswrapper[4978]: E0225 10:01:16.108077 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" Feb 25 10:01:16 crc kubenswrapper[4978]: E0225 10:01:16.178052 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:4198fe0138d2c46ae51eb6e53fed5006\\\"\"" pod="openstack/tempest-tests-tempest" podUID="ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" Feb 25 10:01:16 crc kubenswrapper[4978]: I0225 10:01:16.540301 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:01:16 crc kubenswrapper[4978]: I0225 10:01:16.540386 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:01:16 crc kubenswrapper[4978]: I0225 10:01:16.540437 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 10:01:16 crc kubenswrapper[4978]: I0225 10:01:16.541217 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ddedb687f1478b20f3544dbc0c4281804392d9541579ad2e88d046540e639049"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 10:01:16 crc kubenswrapper[4978]: I0225 10:01:16.541277 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://ddedb687f1478b20f3544dbc0c4281804392d9541579ad2e88d046540e639049" gracePeriod=600 Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.194995 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="ddedb687f1478b20f3544dbc0c4281804392d9541579ad2e88d046540e639049" exitCode=0 Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.195075 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"ddedb687f1478b20f3544dbc0c4281804392d9541579ad2e88d046540e639049"} Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.195357 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3"} Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.195407 4978 scope.go:117] "RemoveContainer" containerID="9027f3ebfd163728dc2e2df00b55de3ae5e517335200654feb3c8261f3c3706c" Feb 25 10:01:17 crc kubenswrapper[4978]: W0225 10:01:17.253537 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod434f88a0_9d17_474f_a877_35569aba520b.slice/crio-2f95c3df6619d114d82303285861c6bb9c508f71a9b34b4ec561c174ac8f32b7 WatchSource:0}: Error finding container 2f95c3df6619d114d82303285861c6bb9c508f71a9b34b4ec561c174ac8f32b7: Status 404 returned error can't find the container with id 2f95c3df6619d114d82303285861c6bb9c508f71a9b34b4ec561c174ac8f32b7 Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.253980 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29533561-phzdw"] Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.560471 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.682633 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-utilities\") pod \"02b2c7fa-5935-470e-b349-63993285f7c4\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.683160 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-catalog-content\") pod \"02b2c7fa-5935-470e-b349-63993285f7c4\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.683239 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-utilities" (OuterVolumeSpecName: "utilities") pod "02b2c7fa-5935-470e-b349-63993285f7c4" (UID: "02b2c7fa-5935-470e-b349-63993285f7c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.683336 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnrpg\" (UniqueName: \"kubernetes.io/projected/02b2c7fa-5935-470e-b349-63993285f7c4-kube-api-access-lnrpg\") pod \"02b2c7fa-5935-470e-b349-63993285f7c4\" (UID: \"02b2c7fa-5935-470e-b349-63993285f7c4\") " Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.684006 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.689920 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02b2c7fa-5935-470e-b349-63993285f7c4-kube-api-access-lnrpg" (OuterVolumeSpecName: "kube-api-access-lnrpg") pod "02b2c7fa-5935-470e-b349-63993285f7c4" (UID: "02b2c7fa-5935-470e-b349-63993285f7c4"). InnerVolumeSpecName "kube-api-access-lnrpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.699251 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "02b2c7fa-5935-470e-b349-63993285f7c4" (UID: "02b2c7fa-5935-470e-b349-63993285f7c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.786577 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02b2c7fa-5935-470e-b349-63993285f7c4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:01:17 crc kubenswrapper[4978]: I0225 10:01:17.786620 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnrpg\" (UniqueName: \"kubernetes.io/projected/02b2c7fa-5935-470e-b349-63993285f7c4-kube-api-access-lnrpg\") on node \"crc\" DevicePath \"\"" Feb 25 10:01:18 crc kubenswrapper[4978]: I0225 10:01:18.205195 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533561-phzdw" event={"ID":"434f88a0-9d17-474f-a877-35569aba520b","Type":"ContainerStarted","Data":"8133ab76189ef7f9381d402e5c33f655ac4fae02d030d7b77b64253710f8b71c"} Feb 25 10:01:18 crc kubenswrapper[4978]: I0225 10:01:18.205237 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533561-phzdw" event={"ID":"434f88a0-9d17-474f-a877-35569aba520b","Type":"ContainerStarted","Data":"2f95c3df6619d114d82303285861c6bb9c508f71a9b34b4ec561c174ac8f32b7"} Feb 25 10:01:18 crc kubenswrapper[4978]: I0225 10:01:18.211205 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d7qwh" event={"ID":"02b2c7fa-5935-470e-b349-63993285f7c4","Type":"ContainerDied","Data":"9ad9dcd5663e25ed66a25e8cb3f12123f8b2bd179e7a4b31e58e5481f6dbddcd"} Feb 25 10:01:18 crc kubenswrapper[4978]: I0225 10:01:18.211246 4978 scope.go:117] "RemoveContainer" containerID="f714208252da1a78715defd850895f4c463dcbe36fcdad01e50ac02a3b307d90" Feb 25 10:01:18 crc kubenswrapper[4978]: I0225 10:01:18.211351 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d7qwh" Feb 25 10:01:18 crc kubenswrapper[4978]: I0225 10:01:18.221700 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29533561-phzdw" podStartSLOduration=18.221680579 podStartE2EDuration="18.221680579s" podCreationTimestamp="2026-02-25 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 10:01:18.218073977 +0000 UTC m=+11771.657330446" watchObservedRunningTime="2026-02-25 10:01:18.221680579 +0000 UTC m=+11771.660937038" Feb 25 10:01:18 crc kubenswrapper[4978]: I0225 10:01:18.245246 4978 scope.go:117] "RemoveContainer" containerID="129f62157fc102e7bee5da8bfa065510cbc44436f7de641aa33971f8f109b3d9" Feb 25 10:01:18 crc kubenswrapper[4978]: I0225 10:01:18.270569 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d7qwh"] Feb 25 10:01:18 crc kubenswrapper[4978]: I0225 10:01:18.279766 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d7qwh"] Feb 25 10:01:18 crc kubenswrapper[4978]: I0225 10:01:18.328956 4978 scope.go:117] "RemoveContainer" containerID="1308f379631b6d2212c95ba3277d27e2ae3c7619554aa2fe308b702030c0ceaa" Feb 25 10:01:19 crc kubenswrapper[4978]: I0225 10:01:19.338085 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" path="/var/lib/kubelet/pods/02b2c7fa-5935-470e-b349-63993285f7c4/volumes" Feb 25 10:01:21 crc kubenswrapper[4978]: I0225 10:01:21.242900 4978 generic.go:334] "Generic (PLEG): container finished" podID="434f88a0-9d17-474f-a877-35569aba520b" containerID="8133ab76189ef7f9381d402e5c33f655ac4fae02d030d7b77b64253710f8b71c" exitCode=0 Feb 25 10:01:21 crc kubenswrapper[4978]: I0225 10:01:21.242992 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533561-phzdw" event={"ID":"434f88a0-9d17-474f-a877-35569aba520b","Type":"ContainerDied","Data":"8133ab76189ef7f9381d402e5c33f655ac4fae02d030d7b77b64253710f8b71c"} Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.548670 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.601539 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-combined-ca-bundle\") pod \"434f88a0-9d17-474f-a877-35569aba520b\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.601607 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-config-data\") pod \"434f88a0-9d17-474f-a877-35569aba520b\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.601665 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-fernet-keys\") pod \"434f88a0-9d17-474f-a877-35569aba520b\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.601853 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vh78l\" (UniqueName: \"kubernetes.io/projected/434f88a0-9d17-474f-a877-35569aba520b-kube-api-access-vh78l\") pod \"434f88a0-9d17-474f-a877-35569aba520b\" (UID: \"434f88a0-9d17-474f-a877-35569aba520b\") " Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.624559 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "434f88a0-9d17-474f-a877-35569aba520b" (UID: "434f88a0-9d17-474f-a877-35569aba520b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.626300 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434f88a0-9d17-474f-a877-35569aba520b-kube-api-access-vh78l" (OuterVolumeSpecName: "kube-api-access-vh78l") pod "434f88a0-9d17-474f-a877-35569aba520b" (UID: "434f88a0-9d17-474f-a877-35569aba520b"). InnerVolumeSpecName "kube-api-access-vh78l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.639553 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "434f88a0-9d17-474f-a877-35569aba520b" (UID: "434f88a0-9d17-474f-a877-35569aba520b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.674135 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-config-data" (OuterVolumeSpecName: "config-data") pod "434f88a0-9d17-474f-a877-35569aba520b" (UID: "434f88a0-9d17-474f-a877-35569aba520b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.704511 4978 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.704550 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vh78l\" (UniqueName: \"kubernetes.io/projected/434f88a0-9d17-474f-a877-35569aba520b-kube-api-access-vh78l\") on node \"crc\" DevicePath \"\"" Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.704571 4978 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 25 10:01:23 crc kubenswrapper[4978]: I0225 10:01:23.704581 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/434f88a0-9d17-474f-a877-35569aba520b-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 10:01:24 crc kubenswrapper[4978]: I0225 10:01:24.271876 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29533561-phzdw" event={"ID":"434f88a0-9d17-474f-a877-35569aba520b","Type":"ContainerDied","Data":"2f95c3df6619d114d82303285861c6bb9c508f71a9b34b4ec561c174ac8f32b7"} Feb 25 10:01:24 crc kubenswrapper[4978]: I0225 10:01:24.272252 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f95c3df6619d114d82303285861c6bb9c508f71a9b34b4ec561c174ac8f32b7" Feb 25 10:01:24 crc kubenswrapper[4978]: I0225 10:01:24.271938 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29533561-phzdw" Feb 25 10:01:30 crc kubenswrapper[4978]: I0225 10:01:30.520862 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 25 10:01:32 crc kubenswrapper[4978]: I0225 10:01:32.389632 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd","Type":"ContainerStarted","Data":"e91d285db051c1003383dd8288a21f37274d719a030841c5c343f36bbdeb9554"} Feb 25 10:01:32 crc kubenswrapper[4978]: I0225 10:01:32.419081 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=5.014561901 podStartE2EDuration="1m11.419054769s" podCreationTimestamp="2026-02-25 10:00:21 +0000 UTC" firstStartedPulling="2026-02-25 10:00:24.112261656 +0000 UTC m=+11717.551518115" lastFinishedPulling="2026-02-25 10:01:30.516754514 +0000 UTC m=+11783.956010983" observedRunningTime="2026-02-25 10:01:32.406253571 +0000 UTC m=+11785.845510030" watchObservedRunningTime="2026-02-25 10:01:32.419054769 +0000 UTC m=+11785.858311258" Feb 25 10:01:41 crc kubenswrapper[4978]: I0225 10:01:41.871959 4978 scope.go:117] "RemoveContainer" containerID="f3899c480d8bf57e7f37390f900f1a897efe250956bf1ee27c8bce18aa3a1969" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.177748 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533562-5kdj2"] Feb 25 10:02:00 crc kubenswrapper[4978]: E0225 10:02:00.224633 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" containerName="extract-utilities" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.224655 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" containerName="extract-utilities" Feb 25 10:02:00 crc kubenswrapper[4978]: E0225 10:02:00.224687 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" containerName="extract-content" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.224694 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" containerName="extract-content" Feb 25 10:02:00 crc kubenswrapper[4978]: E0225 10:02:00.224711 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" containerName="registry-server" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.224717 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" containerName="registry-server" Feb 25 10:02:00 crc kubenswrapper[4978]: E0225 10:02:00.224727 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434f88a0-9d17-474f-a877-35569aba520b" containerName="keystone-cron" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.224733 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="434f88a0-9d17-474f-a877-35569aba520b" containerName="keystone-cron" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.224932 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="434f88a0-9d17-474f-a877-35569aba520b" containerName="keystone-cron" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.224958 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="02b2c7fa-5935-470e-b349-63993285f7c4" containerName="registry-server" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.225700 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533562-5kdj2" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.229784 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.229946 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.234309 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.249646 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533562-5kdj2"] Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.280428 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkk4s\" (UniqueName: \"kubernetes.io/projected/81243863-3696-4d51-9f80-d52d6e9a2242-kube-api-access-nkk4s\") pod \"auto-csr-approver-29533562-5kdj2\" (UID: \"81243863-3696-4d51-9f80-d52d6e9a2242\") " pod="openshift-infra/auto-csr-approver-29533562-5kdj2" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.382565 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkk4s\" (UniqueName: \"kubernetes.io/projected/81243863-3696-4d51-9f80-d52d6e9a2242-kube-api-access-nkk4s\") pod \"auto-csr-approver-29533562-5kdj2\" (UID: \"81243863-3696-4d51-9f80-d52d6e9a2242\") " pod="openshift-infra/auto-csr-approver-29533562-5kdj2" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.414284 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkk4s\" (UniqueName: \"kubernetes.io/projected/81243863-3696-4d51-9f80-d52d6e9a2242-kube-api-access-nkk4s\") pod \"auto-csr-approver-29533562-5kdj2\" (UID: \"81243863-3696-4d51-9f80-d52d6e9a2242\") " pod="openshift-infra/auto-csr-approver-29533562-5kdj2" Feb 25 10:02:00 crc kubenswrapper[4978]: I0225 10:02:00.553701 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533562-5kdj2" Feb 25 10:02:01 crc kubenswrapper[4978]: I0225 10:02:01.259021 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533562-5kdj2"] Feb 25 10:02:01 crc kubenswrapper[4978]: I0225 10:02:01.676293 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533562-5kdj2" event={"ID":"81243863-3696-4d51-9f80-d52d6e9a2242","Type":"ContainerStarted","Data":"ae770acc3dbf251efb2b977952b6d1aca4d4a4fc8b6c9baa0168f6243e5e6979"} Feb 25 10:02:01 crc kubenswrapper[4978]: I0225 10:02:01.971055 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x4v6k"] Feb 25 10:02:01 crc kubenswrapper[4978]: I0225 10:02:01.974344 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.001031 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x4v6k"] Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.015476 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhhsj\" (UniqueName: \"kubernetes.io/projected/d8444727-c0fd-459c-aeef-bbfe22d12348-kube-api-access-zhhsj\") pod \"community-operators-x4v6k\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.015599 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-utilities\") pod \"community-operators-x4v6k\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.015635 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-catalog-content\") pod \"community-operators-x4v6k\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.117801 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-utilities\") pod \"community-operators-x4v6k\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.117852 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-catalog-content\") pod \"community-operators-x4v6k\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.117968 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhhsj\" (UniqueName: \"kubernetes.io/projected/d8444727-c0fd-459c-aeef-bbfe22d12348-kube-api-access-zhhsj\") pod \"community-operators-x4v6k\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.118400 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-catalog-content\") pod \"community-operators-x4v6k\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.118429 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-utilities\") pod \"community-operators-x4v6k\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.137302 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhhsj\" (UniqueName: \"kubernetes.io/projected/d8444727-c0fd-459c-aeef-bbfe22d12348-kube-api-access-zhhsj\") pod \"community-operators-x4v6k\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:02 crc kubenswrapper[4978]: I0225 10:02:02.307414 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:03 crc kubenswrapper[4978]: I0225 10:02:03.125092 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x4v6k"] Feb 25 10:02:03 crc kubenswrapper[4978]: W0225 10:02:03.155972 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8444727_c0fd_459c_aeef_bbfe22d12348.slice/crio-0eaaf744ec97140031218d1e505fdeee3457527a423ca051d436447d7095f956 WatchSource:0}: Error finding container 0eaaf744ec97140031218d1e505fdeee3457527a423ca051d436447d7095f956: Status 404 returned error can't find the container with id 0eaaf744ec97140031218d1e505fdeee3457527a423ca051d436447d7095f956 Feb 25 10:02:03 crc kubenswrapper[4978]: I0225 10:02:03.708544 4978 generic.go:334] "Generic (PLEG): container finished" podID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerID="1d0caec3c0cb112ae2dab8fd655e65dd3dfd9fa3cf4dbac95d10294b8be4d049" exitCode=0 Feb 25 10:02:03 crc kubenswrapper[4978]: I0225 10:02:03.708795 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4v6k" event={"ID":"d8444727-c0fd-459c-aeef-bbfe22d12348","Type":"ContainerDied","Data":"1d0caec3c0cb112ae2dab8fd655e65dd3dfd9fa3cf4dbac95d10294b8be4d049"} Feb 25 10:02:03 crc kubenswrapper[4978]: I0225 10:02:03.708884 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4v6k" event={"ID":"d8444727-c0fd-459c-aeef-bbfe22d12348","Type":"ContainerStarted","Data":"0eaaf744ec97140031218d1e505fdeee3457527a423ca051d436447d7095f956"} Feb 25 10:02:03 crc kubenswrapper[4978]: I0225 10:02:03.716564 4978 generic.go:334] "Generic (PLEG): container finished" podID="81243863-3696-4d51-9f80-d52d6e9a2242" containerID="6a8491a21b13e90cedfd33cec899773db68a66df6a271f8194be346d7be43d9e" exitCode=0 Feb 25 10:02:03 crc kubenswrapper[4978]: I0225 10:02:03.716609 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533562-5kdj2" event={"ID":"81243863-3696-4d51-9f80-d52d6e9a2242","Type":"ContainerDied","Data":"6a8491a21b13e90cedfd33cec899773db68a66df6a271f8194be346d7be43d9e"} Feb 25 10:02:04 crc kubenswrapper[4978]: I0225 10:02:04.755297 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4v6k" event={"ID":"d8444727-c0fd-459c-aeef-bbfe22d12348","Type":"ContainerStarted","Data":"35fb100a3c152ec4745347ea5b2e362692b2db36fde0f149131e5d3cb4120e5e"} Feb 25 10:02:06 crc kubenswrapper[4978]: I0225 10:02:06.210107 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533562-5kdj2" Feb 25 10:02:06 crc kubenswrapper[4978]: I0225 10:02:06.403305 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkk4s\" (UniqueName: \"kubernetes.io/projected/81243863-3696-4d51-9f80-d52d6e9a2242-kube-api-access-nkk4s\") pod \"81243863-3696-4d51-9f80-d52d6e9a2242\" (UID: \"81243863-3696-4d51-9f80-d52d6e9a2242\") " Feb 25 10:02:06 crc kubenswrapper[4978]: I0225 10:02:06.409146 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81243863-3696-4d51-9f80-d52d6e9a2242-kube-api-access-nkk4s" (OuterVolumeSpecName: "kube-api-access-nkk4s") pod "81243863-3696-4d51-9f80-d52d6e9a2242" (UID: "81243863-3696-4d51-9f80-d52d6e9a2242"). InnerVolumeSpecName "kube-api-access-nkk4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:02:06 crc kubenswrapper[4978]: I0225 10:02:06.506385 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkk4s\" (UniqueName: \"kubernetes.io/projected/81243863-3696-4d51-9f80-d52d6e9a2242-kube-api-access-nkk4s\") on node \"crc\" DevicePath \"\"" Feb 25 10:02:06 crc kubenswrapper[4978]: I0225 10:02:06.775427 4978 generic.go:334] "Generic (PLEG): container finished" podID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerID="35fb100a3c152ec4745347ea5b2e362692b2db36fde0f149131e5d3cb4120e5e" exitCode=0 Feb 25 10:02:06 crc kubenswrapper[4978]: I0225 10:02:06.775502 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4v6k" event={"ID":"d8444727-c0fd-459c-aeef-bbfe22d12348","Type":"ContainerDied","Data":"35fb100a3c152ec4745347ea5b2e362692b2db36fde0f149131e5d3cb4120e5e"} Feb 25 10:02:06 crc kubenswrapper[4978]: I0225 10:02:06.779918 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533562-5kdj2" event={"ID":"81243863-3696-4d51-9f80-d52d6e9a2242","Type":"ContainerDied","Data":"ae770acc3dbf251efb2b977952b6d1aca4d4a4fc8b6c9baa0168f6243e5e6979"} Feb 25 10:02:06 crc kubenswrapper[4978]: I0225 10:02:06.779952 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae770acc3dbf251efb2b977952b6d1aca4d4a4fc8b6c9baa0168f6243e5e6979" Feb 25 10:02:06 crc kubenswrapper[4978]: I0225 10:02:06.780007 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533562-5kdj2" Feb 25 10:02:07 crc kubenswrapper[4978]: I0225 10:02:07.287015 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533556-hn6bn"] Feb 25 10:02:07 crc kubenswrapper[4978]: I0225 10:02:07.296968 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533556-hn6bn"] Feb 25 10:02:07 crc kubenswrapper[4978]: I0225 10:02:07.343632 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54e7776d-9897-4420-ae00-7422083846ee" path="/var/lib/kubelet/pods/54e7776d-9897-4420-ae00-7422083846ee/volumes" Feb 25 10:02:08 crc kubenswrapper[4978]: I0225 10:02:08.802636 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4v6k" event={"ID":"d8444727-c0fd-459c-aeef-bbfe22d12348","Type":"ContainerStarted","Data":"0d6256f80fe38459326b8a70b1eef3681df4e35482d42950f1971b3f755572d1"} Feb 25 10:02:12 crc kubenswrapper[4978]: I0225 10:02:12.308573 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:12 crc kubenswrapper[4978]: I0225 10:02:12.310411 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:13 crc kubenswrapper[4978]: I0225 10:02:13.357651 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-x4v6k" podUID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerName="registry-server" probeResult="failure" output=< Feb 25 10:02:13 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 10:02:13 crc kubenswrapper[4978]: > Feb 25 10:02:22 crc kubenswrapper[4978]: I0225 10:02:22.357033 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:22 crc kubenswrapper[4978]: I0225 10:02:22.391191 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x4v6k" podStartSLOduration=17.463981878 podStartE2EDuration="21.391173868s" podCreationTimestamp="2026-02-25 10:02:01 +0000 UTC" firstStartedPulling="2026-02-25 10:02:03.710910025 +0000 UTC m=+11817.150166484" lastFinishedPulling="2026-02-25 10:02:07.638102005 +0000 UTC m=+11821.077358474" observedRunningTime="2026-02-25 10:02:08.827512405 +0000 UTC m=+11822.266768864" watchObservedRunningTime="2026-02-25 10:02:22.391173868 +0000 UTC m=+11835.830430327" Feb 25 10:02:22 crc kubenswrapper[4978]: I0225 10:02:22.410415 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:22 crc kubenswrapper[4978]: I0225 10:02:22.596758 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x4v6k"] Feb 25 10:02:23 crc kubenswrapper[4978]: I0225 10:02:23.967848 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x4v6k" podUID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerName="registry-server" containerID="cri-o://0d6256f80fe38459326b8a70b1eef3681df4e35482d42950f1971b3f755572d1" gracePeriod=2 Feb 25 10:02:24 crc kubenswrapper[4978]: I0225 10:02:24.980386 4978 generic.go:334] "Generic (PLEG): container finished" podID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerID="0d6256f80fe38459326b8a70b1eef3681df4e35482d42950f1971b3f755572d1" exitCode=0 Feb 25 10:02:24 crc kubenswrapper[4978]: I0225 10:02:24.980478 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4v6k" event={"ID":"d8444727-c0fd-459c-aeef-bbfe22d12348","Type":"ContainerDied","Data":"0d6256f80fe38459326b8a70b1eef3681df4e35482d42950f1971b3f755572d1"} Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.575518 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.723571 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-utilities\") pod \"d8444727-c0fd-459c-aeef-bbfe22d12348\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.723712 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhhsj\" (UniqueName: \"kubernetes.io/projected/d8444727-c0fd-459c-aeef-bbfe22d12348-kube-api-access-zhhsj\") pod \"d8444727-c0fd-459c-aeef-bbfe22d12348\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.723738 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-catalog-content\") pod \"d8444727-c0fd-459c-aeef-bbfe22d12348\" (UID: \"d8444727-c0fd-459c-aeef-bbfe22d12348\") " Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.728501 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-utilities" (OuterVolumeSpecName: "utilities") pod "d8444727-c0fd-459c-aeef-bbfe22d12348" (UID: "d8444727-c0fd-459c-aeef-bbfe22d12348"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.730737 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8444727-c0fd-459c-aeef-bbfe22d12348-kube-api-access-zhhsj" (OuterVolumeSpecName: "kube-api-access-zhhsj") pod "d8444727-c0fd-459c-aeef-bbfe22d12348" (UID: "d8444727-c0fd-459c-aeef-bbfe22d12348"). InnerVolumeSpecName "kube-api-access-zhhsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.824805 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8444727-c0fd-459c-aeef-bbfe22d12348" (UID: "d8444727-c0fd-459c-aeef-bbfe22d12348"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.826728 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.826787 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhhsj\" (UniqueName: \"kubernetes.io/projected/d8444727-c0fd-459c-aeef-bbfe22d12348-kube-api-access-zhhsj\") on node \"crc\" DevicePath \"\"" Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.826805 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8444727-c0fd-459c-aeef-bbfe22d12348-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.992460 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x4v6k" event={"ID":"d8444727-c0fd-459c-aeef-bbfe22d12348","Type":"ContainerDied","Data":"0eaaf744ec97140031218d1e505fdeee3457527a423ca051d436447d7095f956"} Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.992502 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x4v6k" Feb 25 10:02:25 crc kubenswrapper[4978]: I0225 10:02:25.992516 4978 scope.go:117] "RemoveContainer" containerID="0d6256f80fe38459326b8a70b1eef3681df4e35482d42950f1971b3f755572d1" Feb 25 10:02:26 crc kubenswrapper[4978]: I0225 10:02:26.020213 4978 scope.go:117] "RemoveContainer" containerID="35fb100a3c152ec4745347ea5b2e362692b2db36fde0f149131e5d3cb4120e5e" Feb 25 10:02:26 crc kubenswrapper[4978]: I0225 10:02:26.034042 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x4v6k"] Feb 25 10:02:26 crc kubenswrapper[4978]: I0225 10:02:26.045161 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x4v6k"] Feb 25 10:02:26 crc kubenswrapper[4978]: I0225 10:02:26.053614 4978 scope.go:117] "RemoveContainer" containerID="1d0caec3c0cb112ae2dab8fd655e65dd3dfd9fa3cf4dbac95d10294b8be4d049" Feb 25 10:02:27 crc kubenswrapper[4978]: I0225 10:02:27.337166 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8444727-c0fd-459c-aeef-bbfe22d12348" path="/var/lib/kubelet/pods/d8444727-c0fd-459c-aeef-bbfe22d12348/volumes" Feb 25 10:02:41 crc kubenswrapper[4978]: I0225 10:02:41.987517 4978 scope.go:117] "RemoveContainer" containerID="ac56bf969c651b9dbd17ea7204ff9a270cb5765f5b0535e968e418a93605e394" Feb 25 10:02:42 crc kubenswrapper[4978]: I0225 10:02:42.038841 4978 scope.go:117] "RemoveContainer" containerID="faea6c3c7912ced8f428bed325fb94c31feffcc20e42d035f69ab5d01c702edf" Feb 25 10:02:42 crc kubenswrapper[4978]: I0225 10:02:42.092348 4978 scope.go:117] "RemoveContainer" containerID="c9ebc3b74847af6d32f8155c200063a475c00abc84efd1082e4c85a6e76b83ae" Feb 25 10:03:46 crc kubenswrapper[4978]: I0225 10:03:46.540217 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:03:46 crc kubenswrapper[4978]: I0225 10:03:46.540834 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.143854 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533564-9fbj4"] Feb 25 10:04:00 crc kubenswrapper[4978]: E0225 10:04:00.144761 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerName="extract-utilities" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.144774 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerName="extract-utilities" Feb 25 10:04:00 crc kubenswrapper[4978]: E0225 10:04:00.144799 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerName="extract-content" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.144805 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerName="extract-content" Feb 25 10:04:00 crc kubenswrapper[4978]: E0225 10:04:00.144817 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerName="registry-server" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.144823 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerName="registry-server" Feb 25 10:04:00 crc kubenswrapper[4978]: E0225 10:04:00.144846 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81243863-3696-4d51-9f80-d52d6e9a2242" containerName="oc" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.144863 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="81243863-3696-4d51-9f80-d52d6e9a2242" containerName="oc" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.145067 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="81243863-3696-4d51-9f80-d52d6e9a2242" containerName="oc" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.145093 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8444727-c0fd-459c-aeef-bbfe22d12348" containerName="registry-server" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.145811 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533564-9fbj4" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.148155 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.148233 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.148272 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.162840 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533564-9fbj4"] Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.240292 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqnww\" (UniqueName: \"kubernetes.io/projected/d52e09c1-2451-42bc-beb9-2589474a0757-kube-api-access-wqnww\") pod \"auto-csr-approver-29533564-9fbj4\" (UID: \"d52e09c1-2451-42bc-beb9-2589474a0757\") " pod="openshift-infra/auto-csr-approver-29533564-9fbj4" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.342805 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqnww\" (UniqueName: \"kubernetes.io/projected/d52e09c1-2451-42bc-beb9-2589474a0757-kube-api-access-wqnww\") pod \"auto-csr-approver-29533564-9fbj4\" (UID: \"d52e09c1-2451-42bc-beb9-2589474a0757\") " pod="openshift-infra/auto-csr-approver-29533564-9fbj4" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.370029 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqnww\" (UniqueName: \"kubernetes.io/projected/d52e09c1-2451-42bc-beb9-2589474a0757-kube-api-access-wqnww\") pod \"auto-csr-approver-29533564-9fbj4\" (UID: \"d52e09c1-2451-42bc-beb9-2589474a0757\") " pod="openshift-infra/auto-csr-approver-29533564-9fbj4" Feb 25 10:04:00 crc kubenswrapper[4978]: I0225 10:04:00.462275 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533564-9fbj4" Feb 25 10:04:01 crc kubenswrapper[4978]: I0225 10:04:01.397844 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533564-9fbj4"] Feb 25 10:04:01 crc kubenswrapper[4978]: I0225 10:04:01.410991 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 10:04:01 crc kubenswrapper[4978]: I0225 10:04:01.901407 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533564-9fbj4" event={"ID":"d52e09c1-2451-42bc-beb9-2589474a0757","Type":"ContainerStarted","Data":"2668ea51bbcf7c553f5539d08534ced434770c9973d8db9b5508608bf6be68e4"} Feb 25 10:04:02 crc kubenswrapper[4978]: I0225 10:04:02.910899 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533564-9fbj4" event={"ID":"d52e09c1-2451-42bc-beb9-2589474a0757","Type":"ContainerStarted","Data":"e0f944945099de8791e042754aae1d5fce8160929304dbe465058d8a6439adad"} Feb 25 10:04:02 crc kubenswrapper[4978]: I0225 10:04:02.928028 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533564-9fbj4" podStartSLOduration=2.103091285 podStartE2EDuration="2.928012781s" podCreationTimestamp="2026-02-25 10:04:00 +0000 UTC" firstStartedPulling="2026-02-25 10:04:01.410768759 +0000 UTC m=+11934.850025218" lastFinishedPulling="2026-02-25 10:04:02.235690255 +0000 UTC m=+11935.674946714" observedRunningTime="2026-02-25 10:04:02.92156843 +0000 UTC m=+11936.360824889" watchObservedRunningTime="2026-02-25 10:04:02.928012781 +0000 UTC m=+11936.367269240" Feb 25 10:04:03 crc kubenswrapper[4978]: I0225 10:04:03.920585 4978 generic.go:334] "Generic (PLEG): container finished" podID="d52e09c1-2451-42bc-beb9-2589474a0757" containerID="e0f944945099de8791e042754aae1d5fce8160929304dbe465058d8a6439adad" exitCode=0 Feb 25 10:04:03 crc kubenswrapper[4978]: I0225 10:04:03.920640 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533564-9fbj4" event={"ID":"d52e09c1-2451-42bc-beb9-2589474a0757","Type":"ContainerDied","Data":"e0f944945099de8791e042754aae1d5fce8160929304dbe465058d8a6439adad"} Feb 25 10:04:06 crc kubenswrapper[4978]: I0225 10:04:06.588019 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533564-9fbj4" Feb 25 10:04:06 crc kubenswrapper[4978]: I0225 10:04:06.685257 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqnww\" (UniqueName: \"kubernetes.io/projected/d52e09c1-2451-42bc-beb9-2589474a0757-kube-api-access-wqnww\") pod \"d52e09c1-2451-42bc-beb9-2589474a0757\" (UID: \"d52e09c1-2451-42bc-beb9-2589474a0757\") " Feb 25 10:04:06 crc kubenswrapper[4978]: I0225 10:04:06.708659 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d52e09c1-2451-42bc-beb9-2589474a0757-kube-api-access-wqnww" (OuterVolumeSpecName: "kube-api-access-wqnww") pod "d52e09c1-2451-42bc-beb9-2589474a0757" (UID: "d52e09c1-2451-42bc-beb9-2589474a0757"). InnerVolumeSpecName "kube-api-access-wqnww". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:04:06 crc kubenswrapper[4978]: I0225 10:04:06.788342 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqnww\" (UniqueName: \"kubernetes.io/projected/d52e09c1-2451-42bc-beb9-2589474a0757-kube-api-access-wqnww\") on node \"crc\" DevicePath \"\"" Feb 25 10:04:06 crc kubenswrapper[4978]: I0225 10:04:06.947991 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533564-9fbj4" event={"ID":"d52e09c1-2451-42bc-beb9-2589474a0757","Type":"ContainerDied","Data":"2668ea51bbcf7c553f5539d08534ced434770c9973d8db9b5508608bf6be68e4"} Feb 25 10:04:06 crc kubenswrapper[4978]: I0225 10:04:06.948029 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2668ea51bbcf7c553f5539d08534ced434770c9973d8db9b5508608bf6be68e4" Feb 25 10:04:06 crc kubenswrapper[4978]: I0225 10:04:06.948086 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533564-9fbj4" Feb 25 10:04:07 crc kubenswrapper[4978]: I0225 10:04:07.662699 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533558-vwg6p"] Feb 25 10:04:07 crc kubenswrapper[4978]: I0225 10:04:07.676596 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533558-vwg6p"] Feb 25 10:04:09 crc kubenswrapper[4978]: I0225 10:04:09.340146 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e1095ae-2b2d-480d-8a71-e056849fa6b5" path="/var/lib/kubelet/pods/2e1095ae-2b2d-480d-8a71-e056849fa6b5/volumes" Feb 25 10:04:16 crc kubenswrapper[4978]: I0225 10:04:16.539821 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:04:16 crc kubenswrapper[4978]: I0225 10:04:16.540233 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:04:42 crc kubenswrapper[4978]: I0225 10:04:42.274552 4978 scope.go:117] "RemoveContainer" containerID="6533d65023411a5f7bd38c24d6762bf5885bd9e3b9f8a7b9fa39cd976c279378" Feb 25 10:04:46 crc kubenswrapper[4978]: I0225 10:04:46.540744 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:04:46 crc kubenswrapper[4978]: I0225 10:04:46.541346 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:04:46 crc kubenswrapper[4978]: I0225 10:04:46.541442 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 10:04:46 crc kubenswrapper[4978]: I0225 10:04:46.542558 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 10:04:46 crc kubenswrapper[4978]: I0225 10:04:46.542660 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" gracePeriod=600 Feb 25 10:04:46 crc kubenswrapper[4978]: E0225 10:04:46.685230 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:04:47 crc kubenswrapper[4978]: I0225 10:04:47.379431 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" exitCode=0 Feb 25 10:04:47 crc kubenswrapper[4978]: I0225 10:04:47.379500 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3"} Feb 25 10:04:47 crc kubenswrapper[4978]: I0225 10:04:47.379555 4978 scope.go:117] "RemoveContainer" containerID="ddedb687f1478b20f3544dbc0c4281804392d9541579ad2e88d046540e639049" Feb 25 10:04:47 crc kubenswrapper[4978]: I0225 10:04:47.380524 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:04:47 crc kubenswrapper[4978]: E0225 10:04:47.381037 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.608951 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6xwhn"] Feb 25 10:04:53 crc kubenswrapper[4978]: E0225 10:04:53.609750 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d52e09c1-2451-42bc-beb9-2589474a0757" containerName="oc" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.609763 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d52e09c1-2451-42bc-beb9-2589474a0757" containerName="oc" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.609974 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d52e09c1-2451-42bc-beb9-2589474a0757" containerName="oc" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.611431 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.628036 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6xwhn"] Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.714234 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-catalog-content\") pod \"redhat-operators-6xwhn\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.714405 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzq6g\" (UniqueName: \"kubernetes.io/projected/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-kube-api-access-wzq6g\") pod \"redhat-operators-6xwhn\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.714466 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-utilities\") pod \"redhat-operators-6xwhn\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.816937 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzq6g\" (UniqueName: \"kubernetes.io/projected/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-kube-api-access-wzq6g\") pod \"redhat-operators-6xwhn\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.817043 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-utilities\") pod \"redhat-operators-6xwhn\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.817180 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-catalog-content\") pod \"redhat-operators-6xwhn\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.817835 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-catalog-content\") pod \"redhat-operators-6xwhn\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.817835 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-utilities\") pod \"redhat-operators-6xwhn\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.842659 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzq6g\" (UniqueName: \"kubernetes.io/projected/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-kube-api-access-wzq6g\") pod \"redhat-operators-6xwhn\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:53 crc kubenswrapper[4978]: I0225 10:04:53.974572 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:04:55 crc kubenswrapper[4978]: I0225 10:04:55.028046 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6xwhn"] Feb 25 10:04:55 crc kubenswrapper[4978]: I0225 10:04:55.464639 4978 generic.go:334] "Generic (PLEG): container finished" podID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerID="efd13f2cff6e59006552cc5741ec38dfb2f115912d9ad18ffa3d6375dfc75315" exitCode=0 Feb 25 10:04:55 crc kubenswrapper[4978]: I0225 10:04:55.464697 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xwhn" event={"ID":"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1","Type":"ContainerDied","Data":"efd13f2cff6e59006552cc5741ec38dfb2f115912d9ad18ffa3d6375dfc75315"} Feb 25 10:04:55 crc kubenswrapper[4978]: I0225 10:04:55.464921 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xwhn" event={"ID":"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1","Type":"ContainerStarted","Data":"0461b6c1f60b94dd5e6e6bfe27e44f9e196554c00e9b0ffdee95c58d8b3fb8f8"} Feb 25 10:04:56 crc kubenswrapper[4978]: I0225 10:04:56.477440 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xwhn" event={"ID":"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1","Type":"ContainerStarted","Data":"f4d65412867d91db7b338ee34b6e826e24d94dfb89638a20487fc485258491d1"} Feb 25 10:05:01 crc kubenswrapper[4978]: I0225 10:05:01.539936 4978 generic.go:334] "Generic (PLEG): container finished" podID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerID="f4d65412867d91db7b338ee34b6e826e24d94dfb89638a20487fc485258491d1" exitCode=0 Feb 25 10:05:01 crc kubenswrapper[4978]: I0225 10:05:01.540036 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xwhn" event={"ID":"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1","Type":"ContainerDied","Data":"f4d65412867d91db7b338ee34b6e826e24d94dfb89638a20487fc485258491d1"} Feb 25 10:05:02 crc kubenswrapper[4978]: I0225 10:05:02.328541 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:05:02 crc kubenswrapper[4978]: E0225 10:05:02.328905 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:05:03 crc kubenswrapper[4978]: I0225 10:05:03.564524 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xwhn" event={"ID":"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1","Type":"ContainerStarted","Data":"b6504166d679e3a0c15b974ed4a74afefdb95d349286ec23b9bd2df3ba0d8455"} Feb 25 10:05:03 crc kubenswrapper[4978]: I0225 10:05:03.597382 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6xwhn" podStartSLOduration=3.364514852 podStartE2EDuration="10.59735017s" podCreationTimestamp="2026-02-25 10:04:53 +0000 UTC" firstStartedPulling="2026-02-25 10:04:55.466686394 +0000 UTC m=+11988.905942853" lastFinishedPulling="2026-02-25 10:05:02.699521702 +0000 UTC m=+11996.138778171" observedRunningTime="2026-02-25 10:05:03.583795358 +0000 UTC m=+11997.023051817" watchObservedRunningTime="2026-02-25 10:05:03.59735017 +0000 UTC m=+11997.036606629" Feb 25 10:05:03 crc kubenswrapper[4978]: I0225 10:05:03.975918 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:05:03 crc kubenswrapper[4978]: I0225 10:05:03.976174 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:05:05 crc kubenswrapper[4978]: I0225 10:05:05.032878 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6xwhn" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerName="registry-server" probeResult="failure" output=< Feb 25 10:05:05 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 10:05:05 crc kubenswrapper[4978]: > Feb 25 10:05:15 crc kubenswrapper[4978]: I0225 10:05:15.034904 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6xwhn" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerName="registry-server" probeResult="failure" output=< Feb 25 10:05:15 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 10:05:15 crc kubenswrapper[4978]: > Feb 25 10:05:17 crc kubenswrapper[4978]: I0225 10:05:17.335037 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:05:17 crc kubenswrapper[4978]: E0225 10:05:17.335749 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:05:24 crc kubenswrapper[4978]: I0225 10:05:24.040890 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:05:24 crc kubenswrapper[4978]: I0225 10:05:24.119014 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:05:24 crc kubenswrapper[4978]: I0225 10:05:24.820380 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6xwhn"] Feb 25 10:05:25 crc kubenswrapper[4978]: I0225 10:05:25.882878 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6xwhn" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerName="registry-server" containerID="cri-o://b6504166d679e3a0c15b974ed4a74afefdb95d349286ec23b9bd2df3ba0d8455" gracePeriod=2 Feb 25 10:05:26 crc kubenswrapper[4978]: I0225 10:05:26.905131 4978 generic.go:334] "Generic (PLEG): container finished" podID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerID="b6504166d679e3a0c15b974ed4a74afefdb95d349286ec23b9bd2df3ba0d8455" exitCode=0 Feb 25 10:05:26 crc kubenswrapper[4978]: I0225 10:05:26.905210 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xwhn" event={"ID":"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1","Type":"ContainerDied","Data":"b6504166d679e3a0c15b974ed4a74afefdb95d349286ec23b9bd2df3ba0d8455"} Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.666300 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.802565 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzq6g\" (UniqueName: \"kubernetes.io/projected/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-kube-api-access-wzq6g\") pod \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.802715 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-catalog-content\") pod \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.802754 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-utilities\") pod \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\" (UID: \"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1\") " Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.803530 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-utilities" (OuterVolumeSpecName: "utilities") pod "3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" (UID: "3d7c7b78-a1c6-4548-a10a-283f2e94f7d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.822597 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-kube-api-access-wzq6g" (OuterVolumeSpecName: "kube-api-access-wzq6g") pod "3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" (UID: "3d7c7b78-a1c6-4548-a10a-283f2e94f7d1"). InnerVolumeSpecName "kube-api-access-wzq6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.906191 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzq6g\" (UniqueName: \"kubernetes.io/projected/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-kube-api-access-wzq6g\") on node \"crc\" DevicePath \"\"" Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.906225 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.921830 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6xwhn" event={"ID":"3d7c7b78-a1c6-4548-a10a-283f2e94f7d1","Type":"ContainerDied","Data":"0461b6c1f60b94dd5e6e6bfe27e44f9e196554c00e9b0ffdee95c58d8b3fb8f8"} Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.921877 4978 scope.go:117] "RemoveContainer" containerID="b6504166d679e3a0c15b974ed4a74afefdb95d349286ec23b9bd2df3ba0d8455" Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.922035 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6xwhn" Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.947516 4978 scope.go:117] "RemoveContainer" containerID="f4d65412867d91db7b338ee34b6e826e24d94dfb89638a20487fc485258491d1" Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.954839 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" (UID: "3d7c7b78-a1c6-4548-a10a-283f2e94f7d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:05:27 crc kubenswrapper[4978]: I0225 10:05:27.976011 4978 scope.go:117] "RemoveContainer" containerID="efd13f2cff6e59006552cc5741ec38dfb2f115912d9ad18ffa3d6375dfc75315" Feb 25 10:05:28 crc kubenswrapper[4978]: I0225 10:05:28.007715 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:05:28 crc kubenswrapper[4978]: I0225 10:05:28.258836 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6xwhn"] Feb 25 10:05:28 crc kubenswrapper[4978]: I0225 10:05:28.270720 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6xwhn"] Feb 25 10:05:29 crc kubenswrapper[4978]: I0225 10:05:29.327888 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:05:29 crc kubenswrapper[4978]: E0225 10:05:29.328264 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:05:29 crc kubenswrapper[4978]: I0225 10:05:29.338972 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" path="/var/lib/kubelet/pods/3d7c7b78-a1c6-4548-a10a-283f2e94f7d1/volumes" Feb 25 10:05:42 crc kubenswrapper[4978]: I0225 10:05:42.328781 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:05:42 crc kubenswrapper[4978]: E0225 10:05:42.329628 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:05:57 crc kubenswrapper[4978]: I0225 10:05:57.335620 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:05:57 crc kubenswrapper[4978]: E0225 10:05:57.336293 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.143342 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533566-bgzph"] Feb 25 10:06:00 crc kubenswrapper[4978]: E0225 10:06:00.144044 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerName="registry-server" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.144057 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerName="registry-server" Feb 25 10:06:00 crc kubenswrapper[4978]: E0225 10:06:00.144082 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerName="extract-utilities" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.144089 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerName="extract-utilities" Feb 25 10:06:00 crc kubenswrapper[4978]: E0225 10:06:00.144134 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerName="extract-content" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.144315 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerName="extract-content" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.144537 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d7c7b78-a1c6-4548-a10a-283f2e94f7d1" containerName="registry-server" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.145258 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533566-bgzph" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.147722 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.147959 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.148654 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.206285 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533566-bgzph"] Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.314215 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwwjq\" (UniqueName: \"kubernetes.io/projected/1b8f8008-3a3c-42ce-ab79-74c1a56b662d-kube-api-access-zwwjq\") pod \"auto-csr-approver-29533566-bgzph\" (UID: \"1b8f8008-3a3c-42ce-ab79-74c1a56b662d\") " pod="openshift-infra/auto-csr-approver-29533566-bgzph" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.416576 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwwjq\" (UniqueName: \"kubernetes.io/projected/1b8f8008-3a3c-42ce-ab79-74c1a56b662d-kube-api-access-zwwjq\") pod \"auto-csr-approver-29533566-bgzph\" (UID: \"1b8f8008-3a3c-42ce-ab79-74c1a56b662d\") " pod="openshift-infra/auto-csr-approver-29533566-bgzph" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.436040 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwwjq\" (UniqueName: \"kubernetes.io/projected/1b8f8008-3a3c-42ce-ab79-74c1a56b662d-kube-api-access-zwwjq\") pod \"auto-csr-approver-29533566-bgzph\" (UID: \"1b8f8008-3a3c-42ce-ab79-74c1a56b662d\") " pod="openshift-infra/auto-csr-approver-29533566-bgzph" Feb 25 10:06:00 crc kubenswrapper[4978]: I0225 10:06:00.524504 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533566-bgzph" Feb 25 10:06:01 crc kubenswrapper[4978]: I0225 10:06:01.397657 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533566-bgzph"] Feb 25 10:06:02 crc kubenswrapper[4978]: I0225 10:06:02.287441 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533566-bgzph" event={"ID":"1b8f8008-3a3c-42ce-ab79-74c1a56b662d","Type":"ContainerStarted","Data":"1792d3bea31ac06c458941d3154dd433a146f94758dfaa998adee7b394d742b0"} Feb 25 10:06:03 crc kubenswrapper[4978]: I0225 10:06:03.299393 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533566-bgzph" event={"ID":"1b8f8008-3a3c-42ce-ab79-74c1a56b662d","Type":"ContainerStarted","Data":"5edf96b8861d25d4ca216ba8c955f40ccfba53333e96a8c8a186ed5883d51465"} Feb 25 10:06:03 crc kubenswrapper[4978]: I0225 10:06:03.313328 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533566-bgzph" podStartSLOduration=2.426023994 podStartE2EDuration="3.313308573s" podCreationTimestamp="2026-02-25 10:06:00 +0000 UTC" firstStartedPulling="2026-02-25 10:06:01.361140604 +0000 UTC m=+12054.800397063" lastFinishedPulling="2026-02-25 10:06:02.248425183 +0000 UTC m=+12055.687681642" observedRunningTime="2026-02-25 10:06:03.312417026 +0000 UTC m=+12056.751673515" watchObservedRunningTime="2026-02-25 10:06:03.313308573 +0000 UTC m=+12056.752565032" Feb 25 10:06:04 crc kubenswrapper[4978]: I0225 10:06:04.311157 4978 generic.go:334] "Generic (PLEG): container finished" podID="1b8f8008-3a3c-42ce-ab79-74c1a56b662d" containerID="5edf96b8861d25d4ca216ba8c955f40ccfba53333e96a8c8a186ed5883d51465" exitCode=0 Feb 25 10:06:04 crc kubenswrapper[4978]: I0225 10:06:04.311206 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533566-bgzph" event={"ID":"1b8f8008-3a3c-42ce-ab79-74c1a56b662d","Type":"ContainerDied","Data":"5edf96b8861d25d4ca216ba8c955f40ccfba53333e96a8c8a186ed5883d51465"} Feb 25 10:06:06 crc kubenswrapper[4978]: I0225 10:06:06.800171 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533566-bgzph" Feb 25 10:06:06 crc kubenswrapper[4978]: I0225 10:06:06.967258 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwwjq\" (UniqueName: \"kubernetes.io/projected/1b8f8008-3a3c-42ce-ab79-74c1a56b662d-kube-api-access-zwwjq\") pod \"1b8f8008-3a3c-42ce-ab79-74c1a56b662d\" (UID: \"1b8f8008-3a3c-42ce-ab79-74c1a56b662d\") " Feb 25 10:06:06 crc kubenswrapper[4978]: I0225 10:06:06.976180 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b8f8008-3a3c-42ce-ab79-74c1a56b662d-kube-api-access-zwwjq" (OuterVolumeSpecName: "kube-api-access-zwwjq") pod "1b8f8008-3a3c-42ce-ab79-74c1a56b662d" (UID: "1b8f8008-3a3c-42ce-ab79-74c1a56b662d"). InnerVolumeSpecName "kube-api-access-zwwjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:06:07 crc kubenswrapper[4978]: I0225 10:06:07.069847 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwwjq\" (UniqueName: \"kubernetes.io/projected/1b8f8008-3a3c-42ce-ab79-74c1a56b662d-kube-api-access-zwwjq\") on node \"crc\" DevicePath \"\"" Feb 25 10:06:07 crc kubenswrapper[4978]: I0225 10:06:07.357029 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533566-bgzph" Feb 25 10:06:07 crc kubenswrapper[4978]: I0225 10:06:07.358676 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533566-bgzph" event={"ID":"1b8f8008-3a3c-42ce-ab79-74c1a56b662d","Type":"ContainerDied","Data":"1792d3bea31ac06c458941d3154dd433a146f94758dfaa998adee7b394d742b0"} Feb 25 10:06:07 crc kubenswrapper[4978]: I0225 10:06:07.358704 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1792d3bea31ac06c458941d3154dd433a146f94758dfaa998adee7b394d742b0" Feb 25 10:06:07 crc kubenswrapper[4978]: I0225 10:06:07.872992 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533560-464p8"] Feb 25 10:06:07 crc kubenswrapper[4978]: I0225 10:06:07.882138 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533560-464p8"] Feb 25 10:06:09 crc kubenswrapper[4978]: I0225 10:06:09.341524 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bd2ef26-a6e6-4787-b8e9-96841e04ae52" path="/var/lib/kubelet/pods/4bd2ef26-a6e6-4787-b8e9-96841e04ae52/volumes" Feb 25 10:06:12 crc kubenswrapper[4978]: I0225 10:06:12.328208 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:06:12 crc kubenswrapper[4978]: E0225 10:06:12.328752 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:06:17 crc kubenswrapper[4978]: E0225 10:06:17.403147 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b8f8008_3a3c_42ce_ab79_74c1a56b662d.slice\": RecentStats: unable to find data in memory cache]" Feb 25 10:06:24 crc kubenswrapper[4978]: I0225 10:06:24.328433 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:06:24 crc kubenswrapper[4978]: E0225 10:06:24.329293 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:06:27 crc kubenswrapper[4978]: E0225 10:06:27.759118 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b8f8008_3a3c_42ce_ab79_74c1a56b662d.slice\": RecentStats: unable to find data in memory cache]" Feb 25 10:06:36 crc kubenswrapper[4978]: I0225 10:06:36.327654 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:06:36 crc kubenswrapper[4978]: E0225 10:06:36.328341 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:06:38 crc kubenswrapper[4978]: E0225 10:06:38.116670 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b8f8008_3a3c_42ce_ab79_74c1a56b662d.slice\": RecentStats: unable to find data in memory cache]" Feb 25 10:06:42 crc kubenswrapper[4978]: I0225 10:06:42.524909 4978 scope.go:117] "RemoveContainer" containerID="f5a38eca26fd819b52414a41986f336e8b7b8ba03e039bdd58f1d4a015bfd00e" Feb 25 10:06:42 crc kubenswrapper[4978]: I0225 10:06:42.578022 4978 scope.go:117] "RemoveContainer" containerID="14de3b94951640888fd04512eeb7a4fd8d2049646328284d8507fe9c04ae56a6" Feb 25 10:06:48 crc kubenswrapper[4978]: I0225 10:06:48.327984 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:06:48 crc kubenswrapper[4978]: E0225 10:06:48.328804 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:06:48 crc kubenswrapper[4978]: E0225 10:06:48.413709 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b8f8008_3a3c_42ce_ab79_74c1a56b662d.slice\": RecentStats: unable to find data in memory cache]" Feb 25 10:06:58 crc kubenswrapper[4978]: E0225 10:06:58.697117 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b8f8008_3a3c_42ce_ab79_74c1a56b662d.slice\": RecentStats: unable to find data in memory cache]" Feb 25 10:07:01 crc kubenswrapper[4978]: I0225 10:07:01.327453 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:07:01 crc kubenswrapper[4978]: E0225 10:07:01.328042 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:07:12 crc kubenswrapper[4978]: I0225 10:07:12.328353 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:07:12 crc kubenswrapper[4978]: E0225 10:07:12.329026 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:07:27 crc kubenswrapper[4978]: I0225 10:07:27.335818 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:07:27 crc kubenswrapper[4978]: E0225 10:07:27.336762 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:07:42 crc kubenswrapper[4978]: I0225 10:07:42.327825 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:07:42 crc kubenswrapper[4978]: E0225 10:07:42.328490 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:07:57 crc kubenswrapper[4978]: I0225 10:07:57.335186 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:07:57 crc kubenswrapper[4978]: E0225 10:07:57.336724 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.223297 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533568-vbxnx"] Feb 25 10:08:00 crc kubenswrapper[4978]: E0225 10:08:00.223951 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b8f8008-3a3c-42ce-ab79-74c1a56b662d" containerName="oc" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.223964 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b8f8008-3a3c-42ce-ab79-74c1a56b662d" containerName="oc" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.224174 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b8f8008-3a3c-42ce-ab79-74c1a56b662d" containerName="oc" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.227409 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533568-vbxnx" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.231084 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.231164 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.231266 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.305818 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533568-vbxnx"] Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.339775 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbh8l\" (UniqueName: \"kubernetes.io/projected/5f151643-7415-47f9-88fe-4f7db486d57c-kube-api-access-lbh8l\") pod \"auto-csr-approver-29533568-vbxnx\" (UID: \"5f151643-7415-47f9-88fe-4f7db486d57c\") " pod="openshift-infra/auto-csr-approver-29533568-vbxnx" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.441323 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbh8l\" (UniqueName: \"kubernetes.io/projected/5f151643-7415-47f9-88fe-4f7db486d57c-kube-api-access-lbh8l\") pod \"auto-csr-approver-29533568-vbxnx\" (UID: \"5f151643-7415-47f9-88fe-4f7db486d57c\") " pod="openshift-infra/auto-csr-approver-29533568-vbxnx" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.473436 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbh8l\" (UniqueName: \"kubernetes.io/projected/5f151643-7415-47f9-88fe-4f7db486d57c-kube-api-access-lbh8l\") pod \"auto-csr-approver-29533568-vbxnx\" (UID: \"5f151643-7415-47f9-88fe-4f7db486d57c\") " pod="openshift-infra/auto-csr-approver-29533568-vbxnx" Feb 25 10:08:00 crc kubenswrapper[4978]: I0225 10:08:00.559383 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533568-vbxnx" Feb 25 10:08:02 crc kubenswrapper[4978]: I0225 10:08:02.097394 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533568-vbxnx"] Feb 25 10:08:02 crc kubenswrapper[4978]: I0225 10:08:02.536388 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533568-vbxnx" event={"ID":"5f151643-7415-47f9-88fe-4f7db486d57c","Type":"ContainerStarted","Data":"ee5ad9cd94d6a55ec13355c84ce686d90a4b865b9f5b737bd0b4351807370192"} Feb 25 10:08:03 crc kubenswrapper[4978]: I0225 10:08:03.549853 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533568-vbxnx" event={"ID":"5f151643-7415-47f9-88fe-4f7db486d57c","Type":"ContainerStarted","Data":"a47b9d26afe9e18fa8e10fecd098a201e4a4f4c45e38f1dc5b8946e493245a34"} Feb 25 10:08:03 crc kubenswrapper[4978]: I0225 10:08:03.578056 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533568-vbxnx" podStartSLOduration=2.726980778 podStartE2EDuration="3.578042228s" podCreationTimestamp="2026-02-25 10:08:00 +0000 UTC" firstStartedPulling="2026-02-25 10:08:02.101644069 +0000 UTC m=+12175.540900528" lastFinishedPulling="2026-02-25 10:08:02.952705509 +0000 UTC m=+12176.391961978" observedRunningTime="2026-02-25 10:08:03.575344274 +0000 UTC m=+12177.014600733" watchObservedRunningTime="2026-02-25 10:08:03.578042228 +0000 UTC m=+12177.017298687" Feb 25 10:08:05 crc kubenswrapper[4978]: I0225 10:08:05.583280 4978 generic.go:334] "Generic (PLEG): container finished" podID="5f151643-7415-47f9-88fe-4f7db486d57c" containerID="a47b9d26afe9e18fa8e10fecd098a201e4a4f4c45e38f1dc5b8946e493245a34" exitCode=0 Feb 25 10:08:05 crc kubenswrapper[4978]: I0225 10:08:05.583460 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533568-vbxnx" event={"ID":"5f151643-7415-47f9-88fe-4f7db486d57c","Type":"ContainerDied","Data":"a47b9d26afe9e18fa8e10fecd098a201e4a4f4c45e38f1dc5b8946e493245a34"} Feb 25 10:08:08 crc kubenswrapper[4978]: I0225 10:08:08.134415 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533568-vbxnx" Feb 25 10:08:08 crc kubenswrapper[4978]: I0225 10:08:08.241237 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbh8l\" (UniqueName: \"kubernetes.io/projected/5f151643-7415-47f9-88fe-4f7db486d57c-kube-api-access-lbh8l\") pod \"5f151643-7415-47f9-88fe-4f7db486d57c\" (UID: \"5f151643-7415-47f9-88fe-4f7db486d57c\") " Feb 25 10:08:08 crc kubenswrapper[4978]: I0225 10:08:08.253634 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f151643-7415-47f9-88fe-4f7db486d57c-kube-api-access-lbh8l" (OuterVolumeSpecName: "kube-api-access-lbh8l") pod "5f151643-7415-47f9-88fe-4f7db486d57c" (UID: "5f151643-7415-47f9-88fe-4f7db486d57c"). InnerVolumeSpecName "kube-api-access-lbh8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:08:08 crc kubenswrapper[4978]: I0225 10:08:08.344358 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbh8l\" (UniqueName: \"kubernetes.io/projected/5f151643-7415-47f9-88fe-4f7db486d57c-kube-api-access-lbh8l\") on node \"crc\" DevicePath \"\"" Feb 25 10:08:08 crc kubenswrapper[4978]: I0225 10:08:08.610866 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533568-vbxnx" event={"ID":"5f151643-7415-47f9-88fe-4f7db486d57c","Type":"ContainerDied","Data":"ee5ad9cd94d6a55ec13355c84ce686d90a4b865b9f5b737bd0b4351807370192"} Feb 25 10:08:08 crc kubenswrapper[4978]: I0225 10:08:08.611317 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533568-vbxnx" Feb 25 10:08:08 crc kubenswrapper[4978]: I0225 10:08:08.611889 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee5ad9cd94d6a55ec13355c84ce686d90a4b865b9f5b737bd0b4351807370192" Feb 25 10:08:09 crc kubenswrapper[4978]: I0225 10:08:09.239421 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533562-5kdj2"] Feb 25 10:08:09 crc kubenswrapper[4978]: I0225 10:08:09.254505 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533562-5kdj2"] Feb 25 10:08:09 crc kubenswrapper[4978]: I0225 10:08:09.339597 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81243863-3696-4d51-9f80-d52d6e9a2242" path="/var/lib/kubelet/pods/81243863-3696-4d51-9f80-d52d6e9a2242/volumes" Feb 25 10:08:11 crc kubenswrapper[4978]: I0225 10:08:11.328072 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:08:11 crc kubenswrapper[4978]: E0225 10:08:11.328705 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:08:23 crc kubenswrapper[4978]: I0225 10:08:23.328094 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:08:23 crc kubenswrapper[4978]: E0225 10:08:23.328846 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:08:36 crc kubenswrapper[4978]: I0225 10:08:36.327579 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:08:36 crc kubenswrapper[4978]: E0225 10:08:36.328352 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:08:42 crc kubenswrapper[4978]: I0225 10:08:42.700581 4978 scope.go:117] "RemoveContainer" containerID="6a8491a21b13e90cedfd33cec899773db68a66df6a271f8194be346d7be43d9e" Feb 25 10:08:50 crc kubenswrapper[4978]: I0225 10:08:50.328749 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:08:50 crc kubenswrapper[4978]: E0225 10:08:50.329514 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:09:04 crc kubenswrapper[4978]: I0225 10:09:04.328680 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:09:04 crc kubenswrapper[4978]: E0225 10:09:04.329740 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:09:16 crc kubenswrapper[4978]: I0225 10:09:16.328196 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:09:16 crc kubenswrapper[4978]: E0225 10:09:16.328968 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:09:27 crc kubenswrapper[4978]: I0225 10:09:27.338064 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:09:27 crc kubenswrapper[4978]: E0225 10:09:27.340042 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:09:39 crc kubenswrapper[4978]: I0225 10:09:39.327912 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:09:39 crc kubenswrapper[4978]: E0225 10:09:39.328688 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:09:52 crc kubenswrapper[4978]: I0225 10:09:52.327643 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:09:52 crc kubenswrapper[4978]: I0225 10:09:52.645732 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"976935188a773bf5312bacbd18db355df365f9b002eecd014780a6407dc6feb0"} Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.165420 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533570-qmqzd"] Feb 25 10:10:00 crc kubenswrapper[4978]: E0225 10:10:00.170298 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f151643-7415-47f9-88fe-4f7db486d57c" containerName="oc" Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.170334 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f151643-7415-47f9-88fe-4f7db486d57c" containerName="oc" Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.172157 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f151643-7415-47f9-88fe-4f7db486d57c" containerName="oc" Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.177185 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533570-qmqzd" Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.185017 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.185014 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.185017 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.187674 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533570-qmqzd"] Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.280664 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78sz5\" (UniqueName: \"kubernetes.io/projected/4d197cca-65f4-4550-8ae2-afa723821792-kube-api-access-78sz5\") pod \"auto-csr-approver-29533570-qmqzd\" (UID: \"4d197cca-65f4-4550-8ae2-afa723821792\") " pod="openshift-infra/auto-csr-approver-29533570-qmqzd" Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.383027 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78sz5\" (UniqueName: \"kubernetes.io/projected/4d197cca-65f4-4550-8ae2-afa723821792-kube-api-access-78sz5\") pod \"auto-csr-approver-29533570-qmqzd\" (UID: \"4d197cca-65f4-4550-8ae2-afa723821792\") " pod="openshift-infra/auto-csr-approver-29533570-qmqzd" Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.406442 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78sz5\" (UniqueName: \"kubernetes.io/projected/4d197cca-65f4-4550-8ae2-afa723821792-kube-api-access-78sz5\") pod \"auto-csr-approver-29533570-qmqzd\" (UID: \"4d197cca-65f4-4550-8ae2-afa723821792\") " pod="openshift-infra/auto-csr-approver-29533570-qmqzd" Feb 25 10:10:00 crc kubenswrapper[4978]: I0225 10:10:00.500692 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533570-qmqzd" Feb 25 10:10:01 crc kubenswrapper[4978]: I0225 10:10:01.808620 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533570-qmqzd"] Feb 25 10:10:01 crc kubenswrapper[4978]: I0225 10:10:01.860079 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 10:10:02 crc kubenswrapper[4978]: I0225 10:10:02.747335 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533570-qmqzd" event={"ID":"4d197cca-65f4-4550-8ae2-afa723821792","Type":"ContainerStarted","Data":"56765352e3d857d2961fec8b8b992b1f8bf991aa6f89c8422b9004138d2fe807"} Feb 25 10:10:04 crc kubenswrapper[4978]: I0225 10:10:04.778525 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533570-qmqzd" event={"ID":"4d197cca-65f4-4550-8ae2-afa723821792","Type":"ContainerStarted","Data":"004f6532eb56c46f1a5156b33543482daffc1dfa00a5ae4b221a4be22addfbbe"} Feb 25 10:10:04 crc kubenswrapper[4978]: I0225 10:10:04.794717 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533570-qmqzd" podStartSLOduration=3.772512244 podStartE2EDuration="4.794677834s" podCreationTimestamp="2026-02-25 10:10:00 +0000 UTC" firstStartedPulling="2026-02-25 10:10:01.859852145 +0000 UTC m=+12295.299108604" lastFinishedPulling="2026-02-25 10:10:02.882017745 +0000 UTC m=+12296.321274194" observedRunningTime="2026-02-25 10:10:04.790806873 +0000 UTC m=+12298.230063332" watchObservedRunningTime="2026-02-25 10:10:04.794677834 +0000 UTC m=+12298.233934323" Feb 25 10:10:05 crc kubenswrapper[4978]: I0225 10:10:05.790639 4978 generic.go:334] "Generic (PLEG): container finished" podID="4d197cca-65f4-4550-8ae2-afa723821792" containerID="004f6532eb56c46f1a5156b33543482daffc1dfa00a5ae4b221a4be22addfbbe" exitCode=0 Feb 25 10:10:05 crc kubenswrapper[4978]: I0225 10:10:05.790739 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533570-qmqzd" event={"ID":"4d197cca-65f4-4550-8ae2-afa723821792","Type":"ContainerDied","Data":"004f6532eb56c46f1a5156b33543482daffc1dfa00a5ae4b221a4be22addfbbe"} Feb 25 10:10:07 crc kubenswrapper[4978]: I0225 10:10:07.947466 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533570-qmqzd" Feb 25 10:10:08 crc kubenswrapper[4978]: I0225 10:10:08.040809 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78sz5\" (UniqueName: \"kubernetes.io/projected/4d197cca-65f4-4550-8ae2-afa723821792-kube-api-access-78sz5\") pod \"4d197cca-65f4-4550-8ae2-afa723821792\" (UID: \"4d197cca-65f4-4550-8ae2-afa723821792\") " Feb 25 10:10:08 crc kubenswrapper[4978]: I0225 10:10:08.054713 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d197cca-65f4-4550-8ae2-afa723821792-kube-api-access-78sz5" (OuterVolumeSpecName: "kube-api-access-78sz5") pod "4d197cca-65f4-4550-8ae2-afa723821792" (UID: "4d197cca-65f4-4550-8ae2-afa723821792"). InnerVolumeSpecName "kube-api-access-78sz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:10:08 crc kubenswrapper[4978]: I0225 10:10:08.143082 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78sz5\" (UniqueName: \"kubernetes.io/projected/4d197cca-65f4-4550-8ae2-afa723821792-kube-api-access-78sz5\") on node \"crc\" DevicePath \"\"" Feb 25 10:10:08 crc kubenswrapper[4978]: I0225 10:10:08.820690 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533570-qmqzd" event={"ID":"4d197cca-65f4-4550-8ae2-afa723821792","Type":"ContainerDied","Data":"56765352e3d857d2961fec8b8b992b1f8bf991aa6f89c8422b9004138d2fe807"} Feb 25 10:10:08 crc kubenswrapper[4978]: I0225 10:10:08.820725 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56765352e3d857d2961fec8b8b992b1f8bf991aa6f89c8422b9004138d2fe807" Feb 25 10:10:08 crc kubenswrapper[4978]: I0225 10:10:08.820775 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533570-qmqzd" Feb 25 10:10:09 crc kubenswrapper[4978]: I0225 10:10:09.035264 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533564-9fbj4"] Feb 25 10:10:09 crc kubenswrapper[4978]: I0225 10:10:09.045613 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533564-9fbj4"] Feb 25 10:10:09 crc kubenswrapper[4978]: I0225 10:10:09.338575 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d52e09c1-2451-42bc-beb9-2589474a0757" path="/var/lib/kubelet/pods/d52e09c1-2451-42bc-beb9-2589474a0757/volumes" Feb 25 10:10:43 crc kubenswrapper[4978]: I0225 10:10:43.048435 4978 scope.go:117] "RemoveContainer" containerID="e0f944945099de8791e042754aae1d5fce8160929304dbe465058d8a6439adad" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.726969 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhg4"] Feb 25 10:11:03 crc kubenswrapper[4978]: E0225 10:11:03.729650 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d197cca-65f4-4550-8ae2-afa723821792" containerName="oc" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.729674 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d197cca-65f4-4550-8ae2-afa723821792" containerName="oc" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.730539 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d197cca-65f4-4550-8ae2-afa723821792" containerName="oc" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.733822 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.745007 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhg4"] Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.785838 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7z9w\" (UniqueName: \"kubernetes.io/projected/42c31f5e-6fec-4985-a75e-11a8216a6ce5-kube-api-access-l7z9w\") pod \"redhat-marketplace-5hhg4\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.785890 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-catalog-content\") pod \"redhat-marketplace-5hhg4\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.785980 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-utilities\") pod \"redhat-marketplace-5hhg4\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.887398 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-utilities\") pod \"redhat-marketplace-5hhg4\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.887551 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7z9w\" (UniqueName: \"kubernetes.io/projected/42c31f5e-6fec-4985-a75e-11a8216a6ce5-kube-api-access-l7z9w\") pod \"redhat-marketplace-5hhg4\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.887597 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-catalog-content\") pod \"redhat-marketplace-5hhg4\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.887821 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-utilities\") pod \"redhat-marketplace-5hhg4\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.888020 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-catalog-content\") pod \"redhat-marketplace-5hhg4\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:03 crc kubenswrapper[4978]: I0225 10:11:03.915457 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7z9w\" (UniqueName: \"kubernetes.io/projected/42c31f5e-6fec-4985-a75e-11a8216a6ce5-kube-api-access-l7z9w\") pod \"redhat-marketplace-5hhg4\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:04 crc kubenswrapper[4978]: I0225 10:11:04.071612 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:05 crc kubenswrapper[4978]: I0225 10:11:05.598192 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhg4"] Feb 25 10:11:06 crc kubenswrapper[4978]: I0225 10:11:06.475321 4978 generic.go:334] "Generic (PLEG): container finished" podID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerID="6779e1e660a6023f68e41dcd5a5e976abfc066fba8a5ea76f1289e425c3a0ecd" exitCode=0 Feb 25 10:11:06 crc kubenswrapper[4978]: I0225 10:11:06.475411 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhg4" event={"ID":"42c31f5e-6fec-4985-a75e-11a8216a6ce5","Type":"ContainerDied","Data":"6779e1e660a6023f68e41dcd5a5e976abfc066fba8a5ea76f1289e425c3a0ecd"} Feb 25 10:11:06 crc kubenswrapper[4978]: I0225 10:11:06.475971 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhg4" event={"ID":"42c31f5e-6fec-4985-a75e-11a8216a6ce5","Type":"ContainerStarted","Data":"20bf33370165ef7ae0ec67a2c643dfa584e91afcd1e4be9e33f174d6d2e3abbe"} Feb 25 10:11:07 crc kubenswrapper[4978]: I0225 10:11:07.522580 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhg4" event={"ID":"42c31f5e-6fec-4985-a75e-11a8216a6ce5","Type":"ContainerStarted","Data":"c490de2d12cca6d20ecd36d00e8a054af68a9fd72f7457d36e11b36a2085282b"} Feb 25 10:11:08 crc kubenswrapper[4978]: I0225 10:11:08.534828 4978 generic.go:334] "Generic (PLEG): container finished" podID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerID="c490de2d12cca6d20ecd36d00e8a054af68a9fd72f7457d36e11b36a2085282b" exitCode=0 Feb 25 10:11:08 crc kubenswrapper[4978]: I0225 10:11:08.534892 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhg4" event={"ID":"42c31f5e-6fec-4985-a75e-11a8216a6ce5","Type":"ContainerDied","Data":"c490de2d12cca6d20ecd36d00e8a054af68a9fd72f7457d36e11b36a2085282b"} Feb 25 10:11:09 crc kubenswrapper[4978]: I0225 10:11:09.547353 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhg4" event={"ID":"42c31f5e-6fec-4985-a75e-11a8216a6ce5","Type":"ContainerStarted","Data":"914b8d040757c1b7a959b8ac9658697c8aa7a7cebf46f5033d8aa5d5f41406b1"} Feb 25 10:11:09 crc kubenswrapper[4978]: I0225 10:11:09.572791 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5hhg4" podStartSLOduration=4.133675696 podStartE2EDuration="6.572764614s" podCreationTimestamp="2026-02-25 10:11:03 +0000 UTC" firstStartedPulling="2026-02-25 10:11:06.477658039 +0000 UTC m=+12359.916914498" lastFinishedPulling="2026-02-25 10:11:08.916746957 +0000 UTC m=+12362.356003416" observedRunningTime="2026-02-25 10:11:09.561879026 +0000 UTC m=+12363.001135505" watchObservedRunningTime="2026-02-25 10:11:09.572764614 +0000 UTC m=+12363.012021093" Feb 25 10:11:14 crc kubenswrapper[4978]: I0225 10:11:14.072481 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:14 crc kubenswrapper[4978]: I0225 10:11:14.072547 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:15 crc kubenswrapper[4978]: I0225 10:11:15.143121 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-5hhg4" podUID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerName="registry-server" probeResult="failure" output=< Feb 25 10:11:15 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 10:11:15 crc kubenswrapper[4978]: > Feb 25 10:11:24 crc kubenswrapper[4978]: I0225 10:11:24.122416 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:24 crc kubenswrapper[4978]: I0225 10:11:24.177844 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:24 crc kubenswrapper[4978]: I0225 10:11:24.385309 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhg4"] Feb 25 10:11:25 crc kubenswrapper[4978]: I0225 10:11:25.696945 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5hhg4" podUID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerName="registry-server" containerID="cri-o://914b8d040757c1b7a959b8ac9658697c8aa7a7cebf46f5033d8aa5d5f41406b1" gracePeriod=2 Feb 25 10:11:26 crc kubenswrapper[4978]: I0225 10:11:26.771734 4978 generic.go:334] "Generic (PLEG): container finished" podID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerID="914b8d040757c1b7a959b8ac9658697c8aa7a7cebf46f5033d8aa5d5f41406b1" exitCode=0 Feb 25 10:11:26 crc kubenswrapper[4978]: I0225 10:11:26.772041 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhg4" event={"ID":"42c31f5e-6fec-4985-a75e-11a8216a6ce5","Type":"ContainerDied","Data":"914b8d040757c1b7a959b8ac9658697c8aa7a7cebf46f5033d8aa5d5f41406b1"} Feb 25 10:11:27 crc kubenswrapper[4978]: I0225 10:11:27.789971 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhg4" event={"ID":"42c31f5e-6fec-4985-a75e-11a8216a6ce5","Type":"ContainerDied","Data":"20bf33370165ef7ae0ec67a2c643dfa584e91afcd1e4be9e33f174d6d2e3abbe"} Feb 25 10:11:27 crc kubenswrapper[4978]: I0225 10:11:27.798397 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20bf33370165ef7ae0ec67a2c643dfa584e91afcd1e4be9e33f174d6d2e3abbe" Feb 25 10:11:27 crc kubenswrapper[4978]: I0225 10:11:27.821365 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.021463 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-catalog-content\") pod \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.021766 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7z9w\" (UniqueName: \"kubernetes.io/projected/42c31f5e-6fec-4985-a75e-11a8216a6ce5-kube-api-access-l7z9w\") pod \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.021807 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-utilities\") pod \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\" (UID: \"42c31f5e-6fec-4985-a75e-11a8216a6ce5\") " Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.032276 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-utilities" (OuterVolumeSpecName: "utilities") pod "42c31f5e-6fec-4985-a75e-11a8216a6ce5" (UID: "42c31f5e-6fec-4985-a75e-11a8216a6ce5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.058176 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42c31f5e-6fec-4985-a75e-11a8216a6ce5" (UID: "42c31f5e-6fec-4985-a75e-11a8216a6ce5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.061157 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42c31f5e-6fec-4985-a75e-11a8216a6ce5-kube-api-access-l7z9w" (OuterVolumeSpecName: "kube-api-access-l7z9w") pod "42c31f5e-6fec-4985-a75e-11a8216a6ce5" (UID: "42c31f5e-6fec-4985-a75e-11a8216a6ce5"). InnerVolumeSpecName "kube-api-access-l7z9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.125314 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7z9w\" (UniqueName: \"kubernetes.io/projected/42c31f5e-6fec-4985-a75e-11a8216a6ce5-kube-api-access-l7z9w\") on node \"crc\" DevicePath \"\"" Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.125350 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.125362 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42c31f5e-6fec-4985-a75e-11a8216a6ce5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.797579 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hhg4" Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.835256 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhg4"] Feb 25 10:11:28 crc kubenswrapper[4978]: I0225 10:11:28.850631 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhg4"] Feb 25 10:11:29 crc kubenswrapper[4978]: I0225 10:11:29.338440 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" path="/var/lib/kubelet/pods/42c31f5e-6fec-4985-a75e-11a8216a6ce5/volumes" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.581569 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533572-rlrwl"] Feb 25 10:12:00 crc kubenswrapper[4978]: E0225 10:12:00.599610 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerName="registry-server" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.599898 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerName="registry-server" Feb 25 10:12:00 crc kubenswrapper[4978]: E0225 10:12:00.601758 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerName="extract-utilities" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.601773 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerName="extract-utilities" Feb 25 10:12:00 crc kubenswrapper[4978]: E0225 10:12:00.601826 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerName="extract-content" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.601832 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerName="extract-content" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.603790 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="42c31f5e-6fec-4985-a75e-11a8216a6ce5" containerName="registry-server" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.617614 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533572-rlrwl" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.631889 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.633672 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.634661 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.724819 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533572-rlrwl"] Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.725753 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvpdr\" (UniqueName: \"kubernetes.io/projected/792fa502-b983-430e-972d-8f44d765f92a-kube-api-access-hvpdr\") pod \"auto-csr-approver-29533572-rlrwl\" (UID: \"792fa502-b983-430e-972d-8f44d765f92a\") " pod="openshift-infra/auto-csr-approver-29533572-rlrwl" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.827614 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvpdr\" (UniqueName: \"kubernetes.io/projected/792fa502-b983-430e-972d-8f44d765f92a-kube-api-access-hvpdr\") pod \"auto-csr-approver-29533572-rlrwl\" (UID: \"792fa502-b983-430e-972d-8f44d765f92a\") " pod="openshift-infra/auto-csr-approver-29533572-rlrwl" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.902240 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvpdr\" (UniqueName: \"kubernetes.io/projected/792fa502-b983-430e-972d-8f44d765f92a-kube-api-access-hvpdr\") pod \"auto-csr-approver-29533572-rlrwl\" (UID: \"792fa502-b983-430e-972d-8f44d765f92a\") " pod="openshift-infra/auto-csr-approver-29533572-rlrwl" Feb 25 10:12:00 crc kubenswrapper[4978]: I0225 10:12:00.996206 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533572-rlrwl" Feb 25 10:12:03 crc kubenswrapper[4978]: I0225 10:12:03.279835 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533572-rlrwl"] Feb 25 10:12:04 crc kubenswrapper[4978]: I0225 10:12:04.182090 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533572-rlrwl" event={"ID":"792fa502-b983-430e-972d-8f44d765f92a","Type":"ContainerStarted","Data":"da4a59e92466c197a835e8728d0cd03a4f45fa182ac9a10c76362ecd3a4e5759"} Feb 25 10:12:06 crc kubenswrapper[4978]: I0225 10:12:06.204392 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533572-rlrwl" event={"ID":"792fa502-b983-430e-972d-8f44d765f92a","Type":"ContainerStarted","Data":"ff50c432bc39c99cffc609f82e85ec6305669b1902f30792897da74b6a2e2b54"} Feb 25 10:12:06 crc kubenswrapper[4978]: I0225 10:12:06.242490 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533572-rlrwl" podStartSLOduration=5.2819179290000005 podStartE2EDuration="6.237997195s" podCreationTimestamp="2026-02-25 10:12:00 +0000 UTC" firstStartedPulling="2026-02-25 10:12:03.328251347 +0000 UTC m=+12416.767507806" lastFinishedPulling="2026-02-25 10:12:04.284330613 +0000 UTC m=+12417.723587072" observedRunningTime="2026-02-25 10:12:06.23298829 +0000 UTC m=+12419.672244759" watchObservedRunningTime="2026-02-25 10:12:06.237997195 +0000 UTC m=+12419.677253654" Feb 25 10:12:08 crc kubenswrapper[4978]: I0225 10:12:08.223416 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533572-rlrwl" event={"ID":"792fa502-b983-430e-972d-8f44d765f92a","Type":"ContainerDied","Data":"ff50c432bc39c99cffc609f82e85ec6305669b1902f30792897da74b6a2e2b54"} Feb 25 10:12:08 crc kubenswrapper[4978]: I0225 10:12:08.224133 4978 generic.go:334] "Generic (PLEG): container finished" podID="792fa502-b983-430e-972d-8f44d765f92a" containerID="ff50c432bc39c99cffc609f82e85ec6305669b1902f30792897da74b6a2e2b54" exitCode=0 Feb 25 10:12:11 crc kubenswrapper[4978]: I0225 10:12:11.217555 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533572-rlrwl" Feb 25 10:12:11 crc kubenswrapper[4978]: I0225 10:12:11.269476 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533572-rlrwl" event={"ID":"792fa502-b983-430e-972d-8f44d765f92a","Type":"ContainerDied","Data":"da4a59e92466c197a835e8728d0cd03a4f45fa182ac9a10c76362ecd3a4e5759"} Feb 25 10:12:11 crc kubenswrapper[4978]: I0225 10:12:11.269515 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533572-rlrwl" Feb 25 10:12:11 crc kubenswrapper[4978]: I0225 10:12:11.272012 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da4a59e92466c197a835e8728d0cd03a4f45fa182ac9a10c76362ecd3a4e5759" Feb 25 10:12:11 crc kubenswrapper[4978]: I0225 10:12:11.352320 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvpdr\" (UniqueName: \"kubernetes.io/projected/792fa502-b983-430e-972d-8f44d765f92a-kube-api-access-hvpdr\") pod \"792fa502-b983-430e-972d-8f44d765f92a\" (UID: \"792fa502-b983-430e-972d-8f44d765f92a\") " Feb 25 10:12:11 crc kubenswrapper[4978]: I0225 10:12:11.377218 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/792fa502-b983-430e-972d-8f44d765f92a-kube-api-access-hvpdr" (OuterVolumeSpecName: "kube-api-access-hvpdr") pod "792fa502-b983-430e-972d-8f44d765f92a" (UID: "792fa502-b983-430e-972d-8f44d765f92a"). InnerVolumeSpecName "kube-api-access-hvpdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:12:11 crc kubenswrapper[4978]: I0225 10:12:11.454837 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvpdr\" (UniqueName: \"kubernetes.io/projected/792fa502-b983-430e-972d-8f44d765f92a-kube-api-access-hvpdr\") on node \"crc\" DevicePath \"\"" Feb 25 10:12:12 crc kubenswrapper[4978]: I0225 10:12:12.385783 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533566-bgzph"] Feb 25 10:12:12 crc kubenswrapper[4978]: I0225 10:12:12.399349 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533566-bgzph"] Feb 25 10:12:13 crc kubenswrapper[4978]: I0225 10:12:13.346527 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b8f8008-3a3c-42ce-ab79-74c1a56b662d" path="/var/lib/kubelet/pods/1b8f8008-3a3c-42ce-ab79-74c1a56b662d/volumes" Feb 25 10:12:16 crc kubenswrapper[4978]: I0225 10:12:16.540740 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:12:16 crc kubenswrapper[4978]: I0225 10:12:16.541686 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:12:43 crc kubenswrapper[4978]: I0225 10:12:43.372118 4978 scope.go:117] "RemoveContainer" containerID="5edf96b8861d25d4ca216ba8c955f40ccfba53333e96a8c8a186ed5883d51465" Feb 25 10:12:46 crc kubenswrapper[4978]: I0225 10:12:46.540973 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:12:46 crc kubenswrapper[4978]: I0225 10:12:46.541468 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:13:16 crc kubenswrapper[4978]: I0225 10:13:16.540464 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:13:16 crc kubenswrapper[4978]: I0225 10:13:16.540886 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:13:16 crc kubenswrapper[4978]: I0225 10:13:16.540932 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 10:13:16 crc kubenswrapper[4978]: I0225 10:13:16.542520 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"976935188a773bf5312bacbd18db355df365f9b002eecd014780a6407dc6feb0"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 10:13:16 crc kubenswrapper[4978]: I0225 10:13:16.542583 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://976935188a773bf5312bacbd18db355df365f9b002eecd014780a6407dc6feb0" gracePeriod=600 Feb 25 10:13:16 crc kubenswrapper[4978]: I0225 10:13:16.907345 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="976935188a773bf5312bacbd18db355df365f9b002eecd014780a6407dc6feb0" exitCode=0 Feb 25 10:13:16 crc kubenswrapper[4978]: I0225 10:13:16.907574 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"976935188a773bf5312bacbd18db355df365f9b002eecd014780a6407dc6feb0"} Feb 25 10:13:16 crc kubenswrapper[4978]: I0225 10:13:16.907978 4978 scope.go:117] "RemoveContainer" containerID="725b8e559976f77f673e15f17c0f67a0028b65b580142b70cdbad663e9d2c9b3" Feb 25 10:13:17 crc kubenswrapper[4978]: I0225 10:13:17.917335 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638"} Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.158433 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533574-7tkgm"] Feb 25 10:14:00 crc kubenswrapper[4978]: E0225 10:14:00.159603 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792fa502-b983-430e-972d-8f44d765f92a" containerName="oc" Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.159621 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="792fa502-b983-430e-972d-8f44d765f92a" containerName="oc" Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.161808 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="792fa502-b983-430e-972d-8f44d765f92a" containerName="oc" Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.167248 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533574-7tkgm" Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.174473 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.174482 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.174478 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.190906 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533574-7tkgm"] Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.258866 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqbtn\" (UniqueName: \"kubernetes.io/projected/cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0-kube-api-access-hqbtn\") pod \"auto-csr-approver-29533574-7tkgm\" (UID: \"cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0\") " pod="openshift-infra/auto-csr-approver-29533574-7tkgm" Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.361382 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqbtn\" (UniqueName: \"kubernetes.io/projected/cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0-kube-api-access-hqbtn\") pod \"auto-csr-approver-29533574-7tkgm\" (UID: \"cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0\") " pod="openshift-infra/auto-csr-approver-29533574-7tkgm" Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.384862 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqbtn\" (UniqueName: \"kubernetes.io/projected/cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0-kube-api-access-hqbtn\") pod \"auto-csr-approver-29533574-7tkgm\" (UID: \"cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0\") " pod="openshift-infra/auto-csr-approver-29533574-7tkgm" Feb 25 10:14:00 crc kubenswrapper[4978]: I0225 10:14:00.486998 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533574-7tkgm" Feb 25 10:14:01 crc kubenswrapper[4978]: I0225 10:14:01.355273 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533574-7tkgm"] Feb 25 10:14:01 crc kubenswrapper[4978]: I0225 10:14:01.371710 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533574-7tkgm" event={"ID":"cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0","Type":"ContainerStarted","Data":"1c7be67032631cc3082fb73602b564b1d094d1ce4929b4ca1e104a33bf3e50f0"} Feb 25 10:14:03 crc kubenswrapper[4978]: I0225 10:14:03.389063 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533574-7tkgm" event={"ID":"cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0","Type":"ContainerStarted","Data":"46c2f61669395902edc871a4076d335d3c788e4a027e1862e37f620c3c4627d5"} Feb 25 10:14:03 crc kubenswrapper[4978]: I0225 10:14:03.414769 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533574-7tkgm" podStartSLOduration=2.599859858 podStartE2EDuration="3.414745589s" podCreationTimestamp="2026-02-25 10:14:00 +0000 UTC" firstStartedPulling="2026-02-25 10:14:01.359143467 +0000 UTC m=+12534.798399926" lastFinishedPulling="2026-02-25 10:14:02.174029198 +0000 UTC m=+12535.613285657" observedRunningTime="2026-02-25 10:14:03.405142922 +0000 UTC m=+12536.844399381" watchObservedRunningTime="2026-02-25 10:14:03.414745589 +0000 UTC m=+12536.854002058" Feb 25 10:14:04 crc kubenswrapper[4978]: I0225 10:14:04.404589 4978 generic.go:334] "Generic (PLEG): container finished" podID="cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0" containerID="46c2f61669395902edc871a4076d335d3c788e4a027e1862e37f620c3c4627d5" exitCode=0 Feb 25 10:14:04 crc kubenswrapper[4978]: I0225 10:14:04.404638 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533574-7tkgm" event={"ID":"cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0","Type":"ContainerDied","Data":"46c2f61669395902edc871a4076d335d3c788e4a027e1862e37f620c3c4627d5"} Feb 25 10:14:06 crc kubenswrapper[4978]: I0225 10:14:06.874613 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533574-7tkgm" Feb 25 10:14:07 crc kubenswrapper[4978]: I0225 10:14:07.034327 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqbtn\" (UniqueName: \"kubernetes.io/projected/cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0-kube-api-access-hqbtn\") pod \"cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0\" (UID: \"cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0\") " Feb 25 10:14:07 crc kubenswrapper[4978]: I0225 10:14:07.051594 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0-kube-api-access-hqbtn" (OuterVolumeSpecName: "kube-api-access-hqbtn") pod "cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0" (UID: "cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0"). InnerVolumeSpecName "kube-api-access-hqbtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:14:07 crc kubenswrapper[4978]: I0225 10:14:07.136849 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqbtn\" (UniqueName: \"kubernetes.io/projected/cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0-kube-api-access-hqbtn\") on node \"crc\" DevicePath \"\"" Feb 25 10:14:07 crc kubenswrapper[4978]: I0225 10:14:07.432560 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533574-7tkgm" event={"ID":"cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0","Type":"ContainerDied","Data":"1c7be67032631cc3082fb73602b564b1d094d1ce4929b4ca1e104a33bf3e50f0"} Feb 25 10:14:07 crc kubenswrapper[4978]: I0225 10:14:07.432609 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c7be67032631cc3082fb73602b564b1d094d1ce4929b4ca1e104a33bf3e50f0" Feb 25 10:14:07 crc kubenswrapper[4978]: I0225 10:14:07.432878 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533574-7tkgm" Feb 25 10:14:07 crc kubenswrapper[4978]: I0225 10:14:07.938452 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533568-vbxnx"] Feb 25 10:14:07 crc kubenswrapper[4978]: I0225 10:14:07.947060 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533568-vbxnx"] Feb 25 10:14:09 crc kubenswrapper[4978]: I0225 10:14:09.339880 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f151643-7415-47f9-88fe-4f7db486d57c" path="/var/lib/kubelet/pods/5f151643-7415-47f9-88fe-4f7db486d57c/volumes" Feb 25 10:14:43 crc kubenswrapper[4978]: I0225 10:14:43.561165 4978 scope.go:117] "RemoveContainer" containerID="a47b9d26afe9e18fa8e10fecd098a201e4a4f4c45e38f1dc5b8946e493245a34" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.174499 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm"] Feb 25 10:15:00 crc kubenswrapper[4978]: E0225 10:15:00.175848 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0" containerName="oc" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.175875 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0" containerName="oc" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.176196 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0" containerName="oc" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.177513 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.186675 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm"] Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.189839 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.190121 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.336077 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q74l7\" (UniqueName: \"kubernetes.io/projected/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-kube-api-access-q74l7\") pod \"collect-profiles-29533575-q7rcm\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.336162 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-config-volume\") pod \"collect-profiles-29533575-q7rcm\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.336354 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-secret-volume\") pod \"collect-profiles-29533575-q7rcm\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.438073 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q74l7\" (UniqueName: \"kubernetes.io/projected/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-kube-api-access-q74l7\") pod \"collect-profiles-29533575-q7rcm\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.438411 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-config-volume\") pod \"collect-profiles-29533575-q7rcm\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.438625 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-secret-volume\") pod \"collect-profiles-29533575-q7rcm\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.439547 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-config-volume\") pod \"collect-profiles-29533575-q7rcm\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.444491 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-secret-volume\") pod \"collect-profiles-29533575-q7rcm\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.460114 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q74l7\" (UniqueName: \"kubernetes.io/projected/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-kube-api-access-q74l7\") pod \"collect-profiles-29533575-q7rcm\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:00 crc kubenswrapper[4978]: I0225 10:15:00.533205 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.313920 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm"] Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.436466 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-628sz"] Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.439348 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.446654 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-628sz"] Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.561062 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-catalog-content\") pod \"redhat-operators-628sz\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.561344 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-utilities\") pod \"redhat-operators-628sz\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.561473 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn5kz\" (UniqueName: \"kubernetes.io/projected/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-kube-api-access-qn5kz\") pod \"redhat-operators-628sz\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.663615 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-utilities\") pod \"redhat-operators-628sz\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.663681 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn5kz\" (UniqueName: \"kubernetes.io/projected/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-kube-api-access-qn5kz\") pod \"redhat-operators-628sz\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.663762 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-catalog-content\") pod \"redhat-operators-628sz\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.664181 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-utilities\") pod \"redhat-operators-628sz\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.664269 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-catalog-content\") pod \"redhat-operators-628sz\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.703817 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn5kz\" (UniqueName: \"kubernetes.io/projected/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-kube-api-access-qn5kz\") pod \"redhat-operators-628sz\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:01 crc kubenswrapper[4978]: I0225 10:15:01.803233 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:02 crc kubenswrapper[4978]: I0225 10:15:02.072666 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" event={"ID":"3a7888d7-9e77-4747-9ad0-b0b3a3da086a","Type":"ContainerStarted","Data":"c4fc1fa4f4fbcc2c3004cf1655e83eb4aa76e5c1c5bd6ef7bd58f30b8d40be44"} Feb 25 10:15:02 crc kubenswrapper[4978]: I0225 10:15:02.073034 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" event={"ID":"3a7888d7-9e77-4747-9ad0-b0b3a3da086a","Type":"ContainerStarted","Data":"978f4aa2fbe47f6f856615ca0cae9d37697af462a54b65f489e5cd521881ec60"} Feb 25 10:15:02 crc kubenswrapper[4978]: I0225 10:15:02.094115 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" podStartSLOduration=2.094098206 podStartE2EDuration="2.094098206s" podCreationTimestamp="2026-02-25 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 10:15:02.086225422 +0000 UTC m=+12595.525481891" watchObservedRunningTime="2026-02-25 10:15:02.094098206 +0000 UTC m=+12595.533354665" Feb 25 10:15:02 crc kubenswrapper[4978]: I0225 10:15:02.555744 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-628sz"] Feb 25 10:15:03 crc kubenswrapper[4978]: I0225 10:15:03.102629 4978 generic.go:334] "Generic (PLEG): container finished" podID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerID="2930c355a1bae92638dd3b40e04fbf9c8dfe67284eb3b3552f01b030958d20f4" exitCode=0 Feb 25 10:15:03 crc kubenswrapper[4978]: I0225 10:15:03.102691 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-628sz" event={"ID":"21bc97fc-00a0-4f2c-8e6d-60d61b989c44","Type":"ContainerDied","Data":"2930c355a1bae92638dd3b40e04fbf9c8dfe67284eb3b3552f01b030958d20f4"} Feb 25 10:15:03 crc kubenswrapper[4978]: I0225 10:15:03.102715 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-628sz" event={"ID":"21bc97fc-00a0-4f2c-8e6d-60d61b989c44","Type":"ContainerStarted","Data":"291f2b978fcdeb15b025f52909c083c9da2d0f1979ae07c73ae230919b71d3e6"} Feb 25 10:15:03 crc kubenswrapper[4978]: I0225 10:15:03.105284 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 10:15:03 crc kubenswrapper[4978]: I0225 10:15:03.107258 4978 generic.go:334] "Generic (PLEG): container finished" podID="3a7888d7-9e77-4747-9ad0-b0b3a3da086a" containerID="c4fc1fa4f4fbcc2c3004cf1655e83eb4aa76e5c1c5bd6ef7bd58f30b8d40be44" exitCode=0 Feb 25 10:15:03 crc kubenswrapper[4978]: I0225 10:15:03.107307 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" event={"ID":"3a7888d7-9e77-4747-9ad0-b0b3a3da086a","Type":"ContainerDied","Data":"c4fc1fa4f4fbcc2c3004cf1655e83eb4aa76e5c1c5bd6ef7bd58f30b8d40be44"} Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.133206 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-628sz" event={"ID":"21bc97fc-00a0-4f2c-8e6d-60d61b989c44","Type":"ContainerStarted","Data":"46f9c8ec1d1326fe5d134e73cf2eac0c42f5b29087949d79e6bc855191f4a554"} Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.663880 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.747359 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-config-volume\") pod \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.747418 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-secret-volume\") pod \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.747480 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q74l7\" (UniqueName: \"kubernetes.io/projected/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-kube-api-access-q74l7\") pod \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\" (UID: \"3a7888d7-9e77-4747-9ad0-b0b3a3da086a\") " Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.747975 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-config-volume" (OuterVolumeSpecName: "config-volume") pod "3a7888d7-9e77-4747-9ad0-b0b3a3da086a" (UID: "3a7888d7-9e77-4747-9ad0-b0b3a3da086a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.753227 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-kube-api-access-q74l7" (OuterVolumeSpecName: "kube-api-access-q74l7") pod "3a7888d7-9e77-4747-9ad0-b0b3a3da086a" (UID: "3a7888d7-9e77-4747-9ad0-b0b3a3da086a"). InnerVolumeSpecName "kube-api-access-q74l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.761528 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3a7888d7-9e77-4747-9ad0-b0b3a3da086a" (UID: "3a7888d7-9e77-4747-9ad0-b0b3a3da086a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.849673 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.849712 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:05 crc kubenswrapper[4978]: I0225 10:15:05.849727 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q74l7\" (UniqueName: \"kubernetes.io/projected/3a7888d7-9e77-4747-9ad0-b0b3a3da086a-kube-api-access-q74l7\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:06 crc kubenswrapper[4978]: I0225 10:15:06.147245 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" Feb 25 10:15:06 crc kubenswrapper[4978]: I0225 10:15:06.151506 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533575-q7rcm" event={"ID":"3a7888d7-9e77-4747-9ad0-b0b3a3da086a","Type":"ContainerDied","Data":"978f4aa2fbe47f6f856615ca0cae9d37697af462a54b65f489e5cd521881ec60"} Feb 25 10:15:06 crc kubenswrapper[4978]: I0225 10:15:06.151559 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="978f4aa2fbe47f6f856615ca0cae9d37697af462a54b65f489e5cd521881ec60" Feb 25 10:15:06 crc kubenswrapper[4978]: I0225 10:15:06.764857 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr"] Feb 25 10:15:06 crc kubenswrapper[4978]: I0225 10:15:06.779703 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533530-nqgpr"] Feb 25 10:15:07 crc kubenswrapper[4978]: I0225 10:15:07.354033 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbd51857-44ff-4aa8-b388-5848273ba5eb" path="/var/lib/kubelet/pods/bbd51857-44ff-4aa8-b388-5848273ba5eb/volumes" Feb 25 10:15:08 crc kubenswrapper[4978]: I0225 10:15:08.168173 4978 generic.go:334] "Generic (PLEG): container finished" podID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerID="46f9c8ec1d1326fe5d134e73cf2eac0c42f5b29087949d79e6bc855191f4a554" exitCode=0 Feb 25 10:15:08 crc kubenswrapper[4978]: I0225 10:15:08.168211 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-628sz" event={"ID":"21bc97fc-00a0-4f2c-8e6d-60d61b989c44","Type":"ContainerDied","Data":"46f9c8ec1d1326fe5d134e73cf2eac0c42f5b29087949d79e6bc855191f4a554"} Feb 25 10:15:09 crc kubenswrapper[4978]: I0225 10:15:09.179499 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-628sz" event={"ID":"21bc97fc-00a0-4f2c-8e6d-60d61b989c44","Type":"ContainerStarted","Data":"cf7437579ea3004fc079c42352cd4c77e3fb216c423a24046493c66c8e679a33"} Feb 25 10:15:09 crc kubenswrapper[4978]: I0225 10:15:09.199983 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-628sz" podStartSLOduration=2.728947305 podStartE2EDuration="8.199968121s" podCreationTimestamp="2026-02-25 10:15:01 +0000 UTC" firstStartedPulling="2026-02-25 10:15:03.105057285 +0000 UTC m=+12596.544313744" lastFinishedPulling="2026-02-25 10:15:08.576078101 +0000 UTC m=+12602.015334560" observedRunningTime="2026-02-25 10:15:09.197200735 +0000 UTC m=+12602.636457204" watchObservedRunningTime="2026-02-25 10:15:09.199968121 +0000 UTC m=+12602.639224570" Feb 25 10:15:11 crc kubenswrapper[4978]: I0225 10:15:11.805003 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:11 crc kubenswrapper[4978]: I0225 10:15:11.805067 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:12 crc kubenswrapper[4978]: I0225 10:15:12.857403 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-628sz" podUID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerName="registry-server" probeResult="failure" output=< Feb 25 10:15:12 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 10:15:12 crc kubenswrapper[4978]: > Feb 25 10:15:16 crc kubenswrapper[4978]: I0225 10:15:16.540591 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:15:16 crc kubenswrapper[4978]: I0225 10:15:16.541096 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:15:18 crc kubenswrapper[4978]: I0225 10:15:18.269783 4978 generic.go:334] "Generic (PLEG): container finished" podID="ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" containerID="e91d285db051c1003383dd8288a21f37274d719a030841c5c343f36bbdeb9554" exitCode=0 Feb 25 10:15:18 crc kubenswrapper[4978]: I0225 10:15:18.269886 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd","Type":"ContainerDied","Data":"e91d285db051c1003383dd8288a21f37274d719a030841c5c343f36bbdeb9554"} Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.740125 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.873032 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config-secret\") pod \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.873220 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-temporary\") pod \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.873347 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ca-certs\") pod \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.873392 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxr2m\" (UniqueName: \"kubernetes.io/projected/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-kube-api-access-fxr2m\") pod \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.873416 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-config-data\") pod \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.873443 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ssh-key\") pod \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.873488 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config\") pod \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.873518 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-workdir\") pod \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.873543 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\" (UID: \"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd\") " Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.874343 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" (UID: "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.878750 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-config-data" (OuterVolumeSpecName: "config-data") pod "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" (UID: "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.880043 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" (UID: "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.883686 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-kube-api-access-fxr2m" (OuterVolumeSpecName: "kube-api-access-fxr2m") pod "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" (UID: "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd"). InnerVolumeSpecName "kube-api-access-fxr2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.885045 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" (UID: "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.911800 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" (UID: "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.914928 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" (UID: "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.930326 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" (UID: "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.947723 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" (UID: "ec6773fc-9a0a-4d54-a950-85b5be1fe6cd"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.975547 4978 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.975580 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.975593 4978 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.975604 4978 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.975613 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxr2m\" (UniqueName: \"kubernetes.io/projected/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-kube-api-access-fxr2m\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.975622 4978 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-config-data\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.975630 4978 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.975640 4978 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:20 crc kubenswrapper[4978]: I0225 10:15:20.975649 4978 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/ec6773fc-9a0a-4d54-a950-85b5be1fe6cd-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:21 crc kubenswrapper[4978]: I0225 10:15:21.001549 4978 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 25 10:15:21 crc kubenswrapper[4978]: I0225 10:15:21.077769 4978 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:21 crc kubenswrapper[4978]: I0225 10:15:21.304653 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"ec6773fc-9a0a-4d54-a950-85b5be1fe6cd","Type":"ContainerDied","Data":"09461336f7e6ab390807996a0637070f0954dbc79bf0245695b4a6000fce7e24"} Feb 25 10:15:21 crc kubenswrapper[4978]: I0225 10:15:21.304698 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09461336f7e6ab390807996a0637070f0954dbc79bf0245695b4a6000fce7e24" Feb 25 10:15:21 crc kubenswrapper[4978]: I0225 10:15:21.304753 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 25 10:15:21 crc kubenswrapper[4978]: I0225 10:15:21.862832 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:21 crc kubenswrapper[4978]: I0225 10:15:21.923610 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:22 crc kubenswrapper[4978]: I0225 10:15:22.103187 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-628sz"] Feb 25 10:15:23 crc kubenswrapper[4978]: I0225 10:15:23.328492 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-628sz" podUID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerName="registry-server" containerID="cri-o://cf7437579ea3004fc079c42352cd4c77e3fb216c423a24046493c66c8e679a33" gracePeriod=2 Feb 25 10:15:24 crc kubenswrapper[4978]: I0225 10:15:24.339380 4978 generic.go:334] "Generic (PLEG): container finished" podID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerID="cf7437579ea3004fc079c42352cd4c77e3fb216c423a24046493c66c8e679a33" exitCode=0 Feb 25 10:15:24 crc kubenswrapper[4978]: I0225 10:15:24.339416 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-628sz" event={"ID":"21bc97fc-00a0-4f2c-8e6d-60d61b989c44","Type":"ContainerDied","Data":"cf7437579ea3004fc079c42352cd4c77e3fb216c423a24046493c66c8e679a33"} Feb 25 10:15:24 crc kubenswrapper[4978]: I0225 10:15:24.833409 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:24 crc kubenswrapper[4978]: I0225 10:15:24.952064 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-utilities\") pod \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " Feb 25 10:15:24 crc kubenswrapper[4978]: I0225 10:15:24.952175 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-catalog-content\") pod \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " Feb 25 10:15:24 crc kubenswrapper[4978]: I0225 10:15:24.952307 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn5kz\" (UniqueName: \"kubernetes.io/projected/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-kube-api-access-qn5kz\") pod \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\" (UID: \"21bc97fc-00a0-4f2c-8e6d-60d61b989c44\") " Feb 25 10:15:24 crc kubenswrapper[4978]: I0225 10:15:24.955234 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-utilities" (OuterVolumeSpecName: "utilities") pod "21bc97fc-00a0-4f2c-8e6d-60d61b989c44" (UID: "21bc97fc-00a0-4f2c-8e6d-60d61b989c44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:15:24 crc kubenswrapper[4978]: I0225 10:15:24.968430 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-kube-api-access-qn5kz" (OuterVolumeSpecName: "kube-api-access-qn5kz") pod "21bc97fc-00a0-4f2c-8e6d-60d61b989c44" (UID: "21bc97fc-00a0-4f2c-8e6d-60d61b989c44"). InnerVolumeSpecName "kube-api-access-qn5kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.054932 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn5kz\" (UniqueName: \"kubernetes.io/projected/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-kube-api-access-qn5kz\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.054966 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.084079 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21bc97fc-00a0-4f2c-8e6d-60d61b989c44" (UID: "21bc97fc-00a0-4f2c-8e6d-60d61b989c44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.156790 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21bc97fc-00a0-4f2c-8e6d-60d61b989c44-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.351600 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-628sz" event={"ID":"21bc97fc-00a0-4f2c-8e6d-60d61b989c44","Type":"ContainerDied","Data":"291f2b978fcdeb15b025f52909c083c9da2d0f1979ae07c73ae230919b71d3e6"} Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.351671 4978 scope.go:117] "RemoveContainer" containerID="cf7437579ea3004fc079c42352cd4c77e3fb216c423a24046493c66c8e679a33" Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.351672 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-628sz" Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.396930 4978 scope.go:117] "RemoveContainer" containerID="46f9c8ec1d1326fe5d134e73cf2eac0c42f5b29087949d79e6bc855191f4a554" Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.398807 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-628sz"] Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.421095 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-628sz"] Feb 25 10:15:25 crc kubenswrapper[4978]: I0225 10:15:25.431755 4978 scope.go:117] "RemoveContainer" containerID="2930c355a1bae92638dd3b40e04fbf9c8dfe67284eb3b3552f01b030958d20f4" Feb 25 10:15:27 crc kubenswrapper[4978]: I0225 10:15:27.338478 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" path="/var/lib/kubelet/pods/21bc97fc-00a0-4f2c-8e6d-60d61b989c44/volumes" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.550989 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 25 10:15:32 crc kubenswrapper[4978]: E0225 10:15:32.552357 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerName="extract-content" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.552406 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerName="extract-content" Feb 25 10:15:32 crc kubenswrapper[4978]: E0225 10:15:32.552456 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerName="registry-server" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.552468 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerName="registry-server" Feb 25 10:15:32 crc kubenswrapper[4978]: E0225 10:15:32.552494 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" containerName="tempest-tests-tempest-tests-runner" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.552506 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" containerName="tempest-tests-tempest-tests-runner" Feb 25 10:15:32 crc kubenswrapper[4978]: E0225 10:15:32.552555 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a7888d7-9e77-4747-9ad0-b0b3a3da086a" containerName="collect-profiles" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.552566 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a7888d7-9e77-4747-9ad0-b0b3a3da086a" containerName="collect-profiles" Feb 25 10:15:32 crc kubenswrapper[4978]: E0225 10:15:32.552585 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerName="extract-utilities" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.552595 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerName="extract-utilities" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.552952 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec6773fc-9a0a-4d54-a950-85b5be1fe6cd" containerName="tempest-tests-tempest-tests-runner" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.553050 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a7888d7-9e77-4747-9ad0-b0b3a3da086a" containerName="collect-profiles" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.553068 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="21bc97fc-00a0-4f2c-8e6d-60d61b989c44" containerName="registry-server" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.554168 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.556579 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fm969" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.567837 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.615866 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbzvz\" (UniqueName: \"kubernetes.io/projected/8f150fab-adf6-49b7-8c94-447452f23850-kube-api-access-nbzvz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8f150fab-adf6-49b7-8c94-447452f23850\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.616031 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8f150fab-adf6-49b7-8c94-447452f23850\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.717555 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbzvz\" (UniqueName: \"kubernetes.io/projected/8f150fab-adf6-49b7-8c94-447452f23850-kube-api-access-nbzvz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8f150fab-adf6-49b7-8c94-447452f23850\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.717892 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8f150fab-adf6-49b7-8c94-447452f23850\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.719758 4978 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8f150fab-adf6-49b7-8c94-447452f23850\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.749293 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbzvz\" (UniqueName: \"kubernetes.io/projected/8f150fab-adf6-49b7-8c94-447452f23850-kube-api-access-nbzvz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8f150fab-adf6-49b7-8c94-447452f23850\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.768029 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8f150fab-adf6-49b7-8c94-447452f23850\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 10:15:32 crc kubenswrapper[4978]: I0225 10:15:32.886425 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 25 10:15:33 crc kubenswrapper[4978]: I0225 10:15:33.623407 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 25 10:15:34 crc kubenswrapper[4978]: I0225 10:15:34.462722 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"8f150fab-adf6-49b7-8c94-447452f23850","Type":"ContainerStarted","Data":"37bcb324167d354905efb1e0653f5fb35e1e3ecbb27bce9e6ea41e36935bf552"} Feb 25 10:15:36 crc kubenswrapper[4978]: I0225 10:15:36.488482 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"8f150fab-adf6-49b7-8c94-447452f23850","Type":"ContainerStarted","Data":"bf018689935780c75c2cf31ec32e48cd74e27acd56e70b8608f675dcb4082a6b"} Feb 25 10:15:36 crc kubenswrapper[4978]: I0225 10:15:36.522556 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.562148421 podStartE2EDuration="4.522533671s" podCreationTimestamp="2026-02-25 10:15:32 +0000 UTC" firstStartedPulling="2026-02-25 10:15:33.627747795 +0000 UTC m=+12627.067004274" lastFinishedPulling="2026-02-25 10:15:35.588133075 +0000 UTC m=+12629.027389524" observedRunningTime="2026-02-25 10:15:36.503391137 +0000 UTC m=+12629.942647596" watchObservedRunningTime="2026-02-25 10:15:36.522533671 +0000 UTC m=+12629.961790140" Feb 25 10:15:43 crc kubenswrapper[4978]: I0225 10:15:43.694114 4978 scope.go:117] "RemoveContainer" containerID="4496ac9d54ffb77e8d4eeea4f3beb20acbc36ea662c186f151130fdefae29eb8" Feb 25 10:15:46 crc kubenswrapper[4978]: I0225 10:15:46.542652 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:15:46 crc kubenswrapper[4978]: I0225 10:15:46.544565 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:16:00 crc kubenswrapper[4978]: I0225 10:16:00.147887 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533576-hr7r8"] Feb 25 10:16:00 crc kubenswrapper[4978]: I0225 10:16:00.149907 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533576-hr7r8" Feb 25 10:16:00 crc kubenswrapper[4978]: I0225 10:16:00.151971 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:16:00 crc kubenswrapper[4978]: I0225 10:16:00.152076 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:16:00 crc kubenswrapper[4978]: I0225 10:16:00.153200 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:16:00 crc kubenswrapper[4978]: I0225 10:16:00.160268 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533576-hr7r8"] Feb 25 10:16:00 crc kubenswrapper[4978]: I0225 10:16:00.240177 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb4rn\" (UniqueName: \"kubernetes.io/projected/cd9f079d-35bf-45c0-850f-d8270674c947-kube-api-access-wb4rn\") pod \"auto-csr-approver-29533576-hr7r8\" (UID: \"cd9f079d-35bf-45c0-850f-d8270674c947\") " pod="openshift-infra/auto-csr-approver-29533576-hr7r8" Feb 25 10:16:00 crc kubenswrapper[4978]: I0225 10:16:00.341706 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb4rn\" (UniqueName: \"kubernetes.io/projected/cd9f079d-35bf-45c0-850f-d8270674c947-kube-api-access-wb4rn\") pod \"auto-csr-approver-29533576-hr7r8\" (UID: \"cd9f079d-35bf-45c0-850f-d8270674c947\") " pod="openshift-infra/auto-csr-approver-29533576-hr7r8" Feb 25 10:16:00 crc kubenswrapper[4978]: I0225 10:16:00.366104 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb4rn\" (UniqueName: \"kubernetes.io/projected/cd9f079d-35bf-45c0-850f-d8270674c947-kube-api-access-wb4rn\") pod \"auto-csr-approver-29533576-hr7r8\" (UID: \"cd9f079d-35bf-45c0-850f-d8270674c947\") " pod="openshift-infra/auto-csr-approver-29533576-hr7r8" Feb 25 10:16:00 crc kubenswrapper[4978]: I0225 10:16:00.471651 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533576-hr7r8" Feb 25 10:16:01 crc kubenswrapper[4978]: I0225 10:16:01.412154 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533576-hr7r8"] Feb 25 10:16:01 crc kubenswrapper[4978]: I0225 10:16:01.808220 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533576-hr7r8" event={"ID":"cd9f079d-35bf-45c0-850f-d8270674c947","Type":"ContainerStarted","Data":"979e167eeeaaeaa80ce6419ecd1a173225faab2079228e172c637635b296748a"} Feb 25 10:16:02 crc kubenswrapper[4978]: I0225 10:16:02.819295 4978 generic.go:334] "Generic (PLEG): container finished" podID="cd9f079d-35bf-45c0-850f-d8270674c947" containerID="20885ce537e96d0ac8f4dd743fa139eaf4a5d77e14ada984740fbdece9e5a952" exitCode=0 Feb 25 10:16:02 crc kubenswrapper[4978]: I0225 10:16:02.819339 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533576-hr7r8" event={"ID":"cd9f079d-35bf-45c0-850f-d8270674c947","Type":"ContainerDied","Data":"20885ce537e96d0ac8f4dd743fa139eaf4a5d77e14ada984740fbdece9e5a952"} Feb 25 10:16:05 crc kubenswrapper[4978]: I0225 10:16:05.207084 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533576-hr7r8" Feb 25 10:16:05 crc kubenswrapper[4978]: I0225 10:16:05.359920 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb4rn\" (UniqueName: \"kubernetes.io/projected/cd9f079d-35bf-45c0-850f-d8270674c947-kube-api-access-wb4rn\") pod \"cd9f079d-35bf-45c0-850f-d8270674c947\" (UID: \"cd9f079d-35bf-45c0-850f-d8270674c947\") " Feb 25 10:16:05 crc kubenswrapper[4978]: I0225 10:16:05.365580 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd9f079d-35bf-45c0-850f-d8270674c947-kube-api-access-wb4rn" (OuterVolumeSpecName: "kube-api-access-wb4rn") pod "cd9f079d-35bf-45c0-850f-d8270674c947" (UID: "cd9f079d-35bf-45c0-850f-d8270674c947"). InnerVolumeSpecName "kube-api-access-wb4rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:16:05 crc kubenswrapper[4978]: I0225 10:16:05.462457 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb4rn\" (UniqueName: \"kubernetes.io/projected/cd9f079d-35bf-45c0-850f-d8270674c947-kube-api-access-wb4rn\") on node \"crc\" DevicePath \"\"" Feb 25 10:16:05 crc kubenswrapper[4978]: I0225 10:16:05.866302 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533576-hr7r8" event={"ID":"cd9f079d-35bf-45c0-850f-d8270674c947","Type":"ContainerDied","Data":"979e167eeeaaeaa80ce6419ecd1a173225faab2079228e172c637635b296748a"} Feb 25 10:16:05 crc kubenswrapper[4978]: I0225 10:16:05.866742 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="979e167eeeaaeaa80ce6419ecd1a173225faab2079228e172c637635b296748a" Feb 25 10:16:05 crc kubenswrapper[4978]: I0225 10:16:05.866619 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533576-hr7r8" Feb 25 10:16:06 crc kubenswrapper[4978]: I0225 10:16:06.284904 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533570-qmqzd"] Feb 25 10:16:06 crc kubenswrapper[4978]: I0225 10:16:06.295211 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533570-qmqzd"] Feb 25 10:16:07 crc kubenswrapper[4978]: I0225 10:16:07.340687 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d197cca-65f4-4550-8ae2-afa723821792" path="/var/lib/kubelet/pods/4d197cca-65f4-4550-8ae2-afa723821792/volumes" Feb 25 10:16:16 crc kubenswrapper[4978]: I0225 10:16:16.539919 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:16:16 crc kubenswrapper[4978]: I0225 10:16:16.540515 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:16:16 crc kubenswrapper[4978]: I0225 10:16:16.540566 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 10:16:16 crc kubenswrapper[4978]: I0225 10:16:16.541430 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 10:16:16 crc kubenswrapper[4978]: I0225 10:16:16.541484 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" gracePeriod=600 Feb 25 10:16:16 crc kubenswrapper[4978]: I0225 10:16:16.998039 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" exitCode=0 Feb 25 10:16:16 crc kubenswrapper[4978]: I0225 10:16:16.998085 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638"} Feb 25 10:16:16 crc kubenswrapper[4978]: I0225 10:16:16.998398 4978 scope.go:117] "RemoveContainer" containerID="976935188a773bf5312bacbd18db355df365f9b002eecd014780a6407dc6feb0" Feb 25 10:16:17 crc kubenswrapper[4978]: E0225 10:16:17.184411 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:16:18 crc kubenswrapper[4978]: I0225 10:16:18.011628 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:16:18 crc kubenswrapper[4978]: E0225 10:16:18.012461 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:16:32 crc kubenswrapper[4978]: I0225 10:16:32.328050 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:16:32 crc kubenswrapper[4978]: E0225 10:16:32.328813 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:16:43 crc kubenswrapper[4978]: I0225 10:16:43.329330 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:16:43 crc kubenswrapper[4978]: E0225 10:16:43.332331 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:16:43 crc kubenswrapper[4978]: I0225 10:16:43.783760 4978 scope.go:117] "RemoveContainer" containerID="004f6532eb56c46f1a5156b33543482daffc1dfa00a5ae4b221a4be22addfbbe" Feb 25 10:16:55 crc kubenswrapper[4978]: I0225 10:16:55.327777 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:16:55 crc kubenswrapper[4978]: E0225 10:16:55.328566 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:16:57 crc kubenswrapper[4978]: I0225 10:16:57.991813 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pv922/must-gather-7jwqp"] Feb 25 10:16:57 crc kubenswrapper[4978]: E0225 10:16:57.992790 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd9f079d-35bf-45c0-850f-d8270674c947" containerName="oc" Feb 25 10:16:57 crc kubenswrapper[4978]: I0225 10:16:57.992812 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd9f079d-35bf-45c0-850f-d8270674c947" containerName="oc" Feb 25 10:16:57 crc kubenswrapper[4978]: I0225 10:16:57.993113 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd9f079d-35bf-45c0-850f-d8270674c947" containerName="oc" Feb 25 10:16:57 crc kubenswrapper[4978]: I0225 10:16:57.994686 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/must-gather-7jwqp" Feb 25 10:16:57 crc kubenswrapper[4978]: I0225 10:16:57.996691 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pv922"/"default-dockercfg-5wlk4" Feb 25 10:16:58 crc kubenswrapper[4978]: I0225 10:16:58.000680 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pv922"/"kube-root-ca.crt" Feb 25 10:16:58 crc kubenswrapper[4978]: I0225 10:16:58.001482 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pv922"/"openshift-service-ca.crt" Feb 25 10:16:58 crc kubenswrapper[4978]: I0225 10:16:58.016224 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pv922/must-gather-7jwqp"] Feb 25 10:16:58 crc kubenswrapper[4978]: I0225 10:16:58.109504 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq4kl\" (UniqueName: \"kubernetes.io/projected/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-kube-api-access-cq4kl\") pod \"must-gather-7jwqp\" (UID: \"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3\") " pod="openshift-must-gather-pv922/must-gather-7jwqp" Feb 25 10:16:58 crc kubenswrapper[4978]: I0225 10:16:58.109973 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-must-gather-output\") pod \"must-gather-7jwqp\" (UID: \"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3\") " pod="openshift-must-gather-pv922/must-gather-7jwqp" Feb 25 10:16:58 crc kubenswrapper[4978]: I0225 10:16:58.212650 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq4kl\" (UniqueName: \"kubernetes.io/projected/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-kube-api-access-cq4kl\") pod \"must-gather-7jwqp\" (UID: \"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3\") " pod="openshift-must-gather-pv922/must-gather-7jwqp" Feb 25 10:16:58 crc kubenswrapper[4978]: I0225 10:16:58.212817 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-must-gather-output\") pod \"must-gather-7jwqp\" (UID: \"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3\") " pod="openshift-must-gather-pv922/must-gather-7jwqp" Feb 25 10:16:58 crc kubenswrapper[4978]: I0225 10:16:58.213352 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-must-gather-output\") pod \"must-gather-7jwqp\" (UID: \"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3\") " pod="openshift-must-gather-pv922/must-gather-7jwqp" Feb 25 10:16:58 crc kubenswrapper[4978]: I0225 10:16:58.234221 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq4kl\" (UniqueName: \"kubernetes.io/projected/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-kube-api-access-cq4kl\") pod \"must-gather-7jwqp\" (UID: \"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3\") " pod="openshift-must-gather-pv922/must-gather-7jwqp" Feb 25 10:16:58 crc kubenswrapper[4978]: I0225 10:16:58.326304 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/must-gather-7jwqp" Feb 25 10:16:59 crc kubenswrapper[4978]: I0225 10:16:59.107166 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pv922/must-gather-7jwqp"] Feb 25 10:16:59 crc kubenswrapper[4978]: I0225 10:16:59.476063 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/must-gather-7jwqp" event={"ID":"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3","Type":"ContainerStarted","Data":"27ad2cacd0dd847a6af3224c4d755c7589730b1b91756080cbb4bdd983aee6a2"} Feb 25 10:17:07 crc kubenswrapper[4978]: I0225 10:17:07.571013 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/must-gather-7jwqp" event={"ID":"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3","Type":"ContainerStarted","Data":"17fe87146b908d0c834c610b28e9febcb8fda4c87e32a8ce509dc56de3a340ee"} Feb 25 10:17:07 crc kubenswrapper[4978]: I0225 10:17:07.571550 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/must-gather-7jwqp" event={"ID":"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3","Type":"ContainerStarted","Data":"a1f38cec30df06dca5c749dd908e691af42f62bf795536d8c4c0f1870be1cc2f"} Feb 25 10:17:07 crc kubenswrapper[4978]: I0225 10:17:07.591775 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pv922/must-gather-7jwqp" podStartSLOduration=2.881927973 podStartE2EDuration="10.591756949s" podCreationTimestamp="2026-02-25 10:16:57 +0000 UTC" firstStartedPulling="2026-02-25 10:16:59.113948926 +0000 UTC m=+12712.553205385" lastFinishedPulling="2026-02-25 10:17:06.823777902 +0000 UTC m=+12720.263034361" observedRunningTime="2026-02-25 10:17:07.585815145 +0000 UTC m=+12721.025071604" watchObservedRunningTime="2026-02-25 10:17:07.591756949 +0000 UTC m=+12721.031013408" Feb 25 10:17:09 crc kubenswrapper[4978]: I0225 10:17:09.328730 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:17:09 crc kubenswrapper[4978]: E0225 10:17:09.329272 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:17:13 crc kubenswrapper[4978]: I0225 10:17:13.182263 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pv922/crc-debug-nzcjw"] Feb 25 10:17:13 crc kubenswrapper[4978]: I0225 10:17:13.184498 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-nzcjw" Feb 25 10:17:13 crc kubenswrapper[4978]: I0225 10:17:13.242957 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92bkl\" (UniqueName: \"kubernetes.io/projected/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-kube-api-access-92bkl\") pod \"crc-debug-nzcjw\" (UID: \"4c8b2b7d-b132-4dce-9b57-9edca2201cc2\") " pod="openshift-must-gather-pv922/crc-debug-nzcjw" Feb 25 10:17:13 crc kubenswrapper[4978]: I0225 10:17:13.243174 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-host\") pod \"crc-debug-nzcjw\" (UID: \"4c8b2b7d-b132-4dce-9b57-9edca2201cc2\") " pod="openshift-must-gather-pv922/crc-debug-nzcjw" Feb 25 10:17:13 crc kubenswrapper[4978]: I0225 10:17:13.345161 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-host\") pod \"crc-debug-nzcjw\" (UID: \"4c8b2b7d-b132-4dce-9b57-9edca2201cc2\") " pod="openshift-must-gather-pv922/crc-debug-nzcjw" Feb 25 10:17:13 crc kubenswrapper[4978]: I0225 10:17:13.345247 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92bkl\" (UniqueName: \"kubernetes.io/projected/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-kube-api-access-92bkl\") pod \"crc-debug-nzcjw\" (UID: \"4c8b2b7d-b132-4dce-9b57-9edca2201cc2\") " pod="openshift-must-gather-pv922/crc-debug-nzcjw" Feb 25 10:17:13 crc kubenswrapper[4978]: I0225 10:17:13.346196 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-host\") pod \"crc-debug-nzcjw\" (UID: \"4c8b2b7d-b132-4dce-9b57-9edca2201cc2\") " pod="openshift-must-gather-pv922/crc-debug-nzcjw" Feb 25 10:17:13 crc kubenswrapper[4978]: I0225 10:17:13.369091 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92bkl\" (UniqueName: \"kubernetes.io/projected/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-kube-api-access-92bkl\") pod \"crc-debug-nzcjw\" (UID: \"4c8b2b7d-b132-4dce-9b57-9edca2201cc2\") " pod="openshift-must-gather-pv922/crc-debug-nzcjw" Feb 25 10:17:13 crc kubenswrapper[4978]: I0225 10:17:13.509721 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-nzcjw" Feb 25 10:17:13 crc kubenswrapper[4978]: W0225 10:17:13.547110 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c8b2b7d_b132_4dce_9b57_9edca2201cc2.slice/crio-1e0f0ef0762818616c902ab5cebd1789c27281c557f2060459ec7b4866d9b279 WatchSource:0}: Error finding container 1e0f0ef0762818616c902ab5cebd1789c27281c557f2060459ec7b4866d9b279: Status 404 returned error can't find the container with id 1e0f0ef0762818616c902ab5cebd1789c27281c557f2060459ec7b4866d9b279 Feb 25 10:17:13 crc kubenswrapper[4978]: I0225 10:17:13.633595 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/crc-debug-nzcjw" event={"ID":"4c8b2b7d-b132-4dce-9b57-9edca2201cc2","Type":"ContainerStarted","Data":"1e0f0ef0762818616c902ab5cebd1789c27281c557f2060459ec7b4866d9b279"} Feb 25 10:17:20 crc kubenswrapper[4978]: I0225 10:17:20.328085 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:17:20 crc kubenswrapper[4978]: E0225 10:17:20.328769 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:17:24 crc kubenswrapper[4978]: I0225 10:17:24.755648 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/crc-debug-nzcjw" event={"ID":"4c8b2b7d-b132-4dce-9b57-9edca2201cc2","Type":"ContainerStarted","Data":"958ef308add9ecabcedb70be71262a09d658cbe64605e78301502d761a1bcf8c"} Feb 25 10:17:24 crc kubenswrapper[4978]: I0225 10:17:24.775987 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pv922/crc-debug-nzcjw" podStartSLOduration=1.731038939 podStartE2EDuration="11.775969901s" podCreationTimestamp="2026-02-25 10:17:13 +0000 UTC" firstStartedPulling="2026-02-25 10:17:13.549652387 +0000 UTC m=+12726.988908846" lastFinishedPulling="2026-02-25 10:17:23.594583349 +0000 UTC m=+12737.033839808" observedRunningTime="2026-02-25 10:17:24.773858285 +0000 UTC m=+12738.213114774" watchObservedRunningTime="2026-02-25 10:17:24.775969901 +0000 UTC m=+12738.215226350" Feb 25 10:17:34 crc kubenswrapper[4978]: I0225 10:17:34.327794 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:17:34 crc kubenswrapper[4978]: E0225 10:17:34.328608 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:17:43 crc kubenswrapper[4978]: I0225 10:17:43.870356 4978 scope.go:117] "RemoveContainer" containerID="c490de2d12cca6d20ecd36d00e8a054af68a9fd72f7457d36e11b36a2085282b" Feb 25 10:17:43 crc kubenswrapper[4978]: I0225 10:17:43.926813 4978 scope.go:117] "RemoveContainer" containerID="6779e1e660a6023f68e41dcd5a5e976abfc066fba8a5ea76f1289e425c3a0ecd" Feb 25 10:17:43 crc kubenswrapper[4978]: I0225 10:17:43.997694 4978 scope.go:117] "RemoveContainer" containerID="914b8d040757c1b7a959b8ac9658697c8aa7a7cebf46f5033d8aa5d5f41406b1" Feb 25 10:17:46 crc kubenswrapper[4978]: I0225 10:17:46.328363 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:17:46 crc kubenswrapper[4978]: E0225 10:17:46.329062 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.158738 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533578-mpvsg"] Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.162411 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533578-mpvsg" Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.167761 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.168117 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.168600 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533578-mpvsg"] Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.168630 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.284390 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-474d7\" (UniqueName: \"kubernetes.io/projected/a3aa68c1-2c13-4bed-bc6e-87342302509f-kube-api-access-474d7\") pod \"auto-csr-approver-29533578-mpvsg\" (UID: \"a3aa68c1-2c13-4bed-bc6e-87342302509f\") " pod="openshift-infra/auto-csr-approver-29533578-mpvsg" Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.329641 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:18:00 crc kubenswrapper[4978]: E0225 10:18:00.331165 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.387137 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-474d7\" (UniqueName: \"kubernetes.io/projected/a3aa68c1-2c13-4bed-bc6e-87342302509f-kube-api-access-474d7\") pod \"auto-csr-approver-29533578-mpvsg\" (UID: \"a3aa68c1-2c13-4bed-bc6e-87342302509f\") " pod="openshift-infra/auto-csr-approver-29533578-mpvsg" Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.416514 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-474d7\" (UniqueName: \"kubernetes.io/projected/a3aa68c1-2c13-4bed-bc6e-87342302509f-kube-api-access-474d7\") pod \"auto-csr-approver-29533578-mpvsg\" (UID: \"a3aa68c1-2c13-4bed-bc6e-87342302509f\") " pod="openshift-infra/auto-csr-approver-29533578-mpvsg" Feb 25 10:18:00 crc kubenswrapper[4978]: I0225 10:18:00.501561 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533578-mpvsg" Feb 25 10:18:01 crc kubenswrapper[4978]: I0225 10:18:01.261061 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533578-mpvsg"] Feb 25 10:18:02 crc kubenswrapper[4978]: I0225 10:18:02.232790 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533578-mpvsg" event={"ID":"a3aa68c1-2c13-4bed-bc6e-87342302509f","Type":"ContainerStarted","Data":"c2b82226c0256aa79a32711e0cdf9036ba2c196c4057ed1f96a1e0db122f1ef2"} Feb 25 10:18:03 crc kubenswrapper[4978]: I0225 10:18:03.243673 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533578-mpvsg" event={"ID":"a3aa68c1-2c13-4bed-bc6e-87342302509f","Type":"ContainerStarted","Data":"4e125d1aec87b4b1c2f404989d66d76ca9469688131870420a2372d61fed8027"} Feb 25 10:18:03 crc kubenswrapper[4978]: I0225 10:18:03.267049 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533578-mpvsg" podStartSLOduration=2.399484789 podStartE2EDuration="3.267028022s" podCreationTimestamp="2026-02-25 10:18:00 +0000 UTC" firstStartedPulling="2026-02-25 10:18:01.23976507 +0000 UTC m=+12774.679021529" lastFinishedPulling="2026-02-25 10:18:02.107308303 +0000 UTC m=+12775.546564762" observedRunningTime="2026-02-25 10:18:03.255533147 +0000 UTC m=+12776.694789606" watchObservedRunningTime="2026-02-25 10:18:03.267028022 +0000 UTC m=+12776.706284481" Feb 25 10:18:04 crc kubenswrapper[4978]: I0225 10:18:04.255588 4978 generic.go:334] "Generic (PLEG): container finished" podID="a3aa68c1-2c13-4bed-bc6e-87342302509f" containerID="4e125d1aec87b4b1c2f404989d66d76ca9469688131870420a2372d61fed8027" exitCode=0 Feb 25 10:18:04 crc kubenswrapper[4978]: I0225 10:18:04.255633 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533578-mpvsg" event={"ID":"a3aa68c1-2c13-4bed-bc6e-87342302509f","Type":"ContainerDied","Data":"4e125d1aec87b4b1c2f404989d66d76ca9469688131870420a2372d61fed8027"} Feb 25 10:18:07 crc kubenswrapper[4978]: I0225 10:18:07.000619 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533578-mpvsg" Feb 25 10:18:07 crc kubenswrapper[4978]: I0225 10:18:07.127693 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-474d7\" (UniqueName: \"kubernetes.io/projected/a3aa68c1-2c13-4bed-bc6e-87342302509f-kube-api-access-474d7\") pod \"a3aa68c1-2c13-4bed-bc6e-87342302509f\" (UID: \"a3aa68c1-2c13-4bed-bc6e-87342302509f\") " Feb 25 10:18:07 crc kubenswrapper[4978]: I0225 10:18:07.132649 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3aa68c1-2c13-4bed-bc6e-87342302509f-kube-api-access-474d7" (OuterVolumeSpecName: "kube-api-access-474d7") pod "a3aa68c1-2c13-4bed-bc6e-87342302509f" (UID: "a3aa68c1-2c13-4bed-bc6e-87342302509f"). InnerVolumeSpecName "kube-api-access-474d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:18:07 crc kubenswrapper[4978]: I0225 10:18:07.229907 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-474d7\" (UniqueName: \"kubernetes.io/projected/a3aa68c1-2c13-4bed-bc6e-87342302509f-kube-api-access-474d7\") on node \"crc\" DevicePath \"\"" Feb 25 10:18:07 crc kubenswrapper[4978]: I0225 10:18:07.316647 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533578-mpvsg" event={"ID":"a3aa68c1-2c13-4bed-bc6e-87342302509f","Type":"ContainerDied","Data":"c2b82226c0256aa79a32711e0cdf9036ba2c196c4057ed1f96a1e0db122f1ef2"} Feb 25 10:18:07 crc kubenswrapper[4978]: I0225 10:18:07.316694 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2b82226c0256aa79a32711e0cdf9036ba2c196c4057ed1f96a1e0db122f1ef2" Feb 25 10:18:07 crc kubenswrapper[4978]: I0225 10:18:07.317000 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533578-mpvsg" Feb 25 10:18:08 crc kubenswrapper[4978]: I0225 10:18:08.085345 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533572-rlrwl"] Feb 25 10:18:08 crc kubenswrapper[4978]: I0225 10:18:08.097934 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533572-rlrwl"] Feb 25 10:18:09 crc kubenswrapper[4978]: I0225 10:18:09.339526 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="792fa502-b983-430e-972d-8f44d765f92a" path="/var/lib/kubelet/pods/792fa502-b983-430e-972d-8f44d765f92a/volumes" Feb 25 10:18:14 crc kubenswrapper[4978]: I0225 10:18:14.327691 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:18:14 crc kubenswrapper[4978]: E0225 10:18:14.328288 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:18:16 crc kubenswrapper[4978]: I0225 10:18:16.404493 4978 generic.go:334] "Generic (PLEG): container finished" podID="4c8b2b7d-b132-4dce-9b57-9edca2201cc2" containerID="958ef308add9ecabcedb70be71262a09d658cbe64605e78301502d761a1bcf8c" exitCode=0 Feb 25 10:18:16 crc kubenswrapper[4978]: I0225 10:18:16.404579 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/crc-debug-nzcjw" event={"ID":"4c8b2b7d-b132-4dce-9b57-9edca2201cc2","Type":"ContainerDied","Data":"958ef308add9ecabcedb70be71262a09d658cbe64605e78301502d761a1bcf8c"} Feb 25 10:18:17 crc kubenswrapper[4978]: I0225 10:18:17.544817 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-nzcjw" Feb 25 10:18:17 crc kubenswrapper[4978]: I0225 10:18:17.586240 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pv922/crc-debug-nzcjw"] Feb 25 10:18:17 crc kubenswrapper[4978]: I0225 10:18:17.591443 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92bkl\" (UniqueName: \"kubernetes.io/projected/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-kube-api-access-92bkl\") pod \"4c8b2b7d-b132-4dce-9b57-9edca2201cc2\" (UID: \"4c8b2b7d-b132-4dce-9b57-9edca2201cc2\") " Feb 25 10:18:17 crc kubenswrapper[4978]: I0225 10:18:17.591617 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-host\") pod \"4c8b2b7d-b132-4dce-9b57-9edca2201cc2\" (UID: \"4c8b2b7d-b132-4dce-9b57-9edca2201cc2\") " Feb 25 10:18:17 crc kubenswrapper[4978]: I0225 10:18:17.591731 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-host" (OuterVolumeSpecName: "host") pod "4c8b2b7d-b132-4dce-9b57-9edca2201cc2" (UID: "4c8b2b7d-b132-4dce-9b57-9edca2201cc2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 10:18:17 crc kubenswrapper[4978]: I0225 10:18:17.592018 4978 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-host\") on node \"crc\" DevicePath \"\"" Feb 25 10:18:17 crc kubenswrapper[4978]: I0225 10:18:17.595120 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pv922/crc-debug-nzcjw"] Feb 25 10:18:17 crc kubenswrapper[4978]: I0225 10:18:17.598311 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-kube-api-access-92bkl" (OuterVolumeSpecName: "kube-api-access-92bkl") pod "4c8b2b7d-b132-4dce-9b57-9edca2201cc2" (UID: "4c8b2b7d-b132-4dce-9b57-9edca2201cc2"). InnerVolumeSpecName "kube-api-access-92bkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:18:17 crc kubenswrapper[4978]: I0225 10:18:17.694159 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92bkl\" (UniqueName: \"kubernetes.io/projected/4c8b2b7d-b132-4dce-9b57-9edca2201cc2-kube-api-access-92bkl\") on node \"crc\" DevicePath \"\"" Feb 25 10:18:18 crc kubenswrapper[4978]: I0225 10:18:18.422816 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e0f0ef0762818616c902ab5cebd1789c27281c557f2060459ec7b4866d9b279" Feb 25 10:18:18 crc kubenswrapper[4978]: I0225 10:18:18.422868 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-nzcjw" Feb 25 10:18:18 crc kubenswrapper[4978]: I0225 10:18:18.957150 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pv922/crc-debug-dkr9g"] Feb 25 10:18:18 crc kubenswrapper[4978]: E0225 10:18:18.958051 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8b2b7d-b132-4dce-9b57-9edca2201cc2" containerName="container-00" Feb 25 10:18:18 crc kubenswrapper[4978]: I0225 10:18:18.958071 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8b2b7d-b132-4dce-9b57-9edca2201cc2" containerName="container-00" Feb 25 10:18:18 crc kubenswrapper[4978]: E0225 10:18:18.958114 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3aa68c1-2c13-4bed-bc6e-87342302509f" containerName="oc" Feb 25 10:18:18 crc kubenswrapper[4978]: I0225 10:18:18.958123 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3aa68c1-2c13-4bed-bc6e-87342302509f" containerName="oc" Feb 25 10:18:18 crc kubenswrapper[4978]: I0225 10:18:18.958445 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3aa68c1-2c13-4bed-bc6e-87342302509f" containerName="oc" Feb 25 10:18:18 crc kubenswrapper[4978]: I0225 10:18:18.958483 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c8b2b7d-b132-4dce-9b57-9edca2201cc2" containerName="container-00" Feb 25 10:18:18 crc kubenswrapper[4978]: I0225 10:18:18.959333 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-dkr9g" Feb 25 10:18:19 crc kubenswrapper[4978]: I0225 10:18:19.048211 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-host\") pod \"crc-debug-dkr9g\" (UID: \"10fb0d06-4f2b-44c2-aebf-7075d918b3e2\") " pod="openshift-must-gather-pv922/crc-debug-dkr9g" Feb 25 10:18:19 crc kubenswrapper[4978]: I0225 10:18:19.048297 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmxz9\" (UniqueName: \"kubernetes.io/projected/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-kube-api-access-lmxz9\") pod \"crc-debug-dkr9g\" (UID: \"10fb0d06-4f2b-44c2-aebf-7075d918b3e2\") " pod="openshift-must-gather-pv922/crc-debug-dkr9g" Feb 25 10:18:19 crc kubenswrapper[4978]: I0225 10:18:19.150118 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-host\") pod \"crc-debug-dkr9g\" (UID: \"10fb0d06-4f2b-44c2-aebf-7075d918b3e2\") " pod="openshift-must-gather-pv922/crc-debug-dkr9g" Feb 25 10:18:19 crc kubenswrapper[4978]: I0225 10:18:19.150207 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmxz9\" (UniqueName: \"kubernetes.io/projected/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-kube-api-access-lmxz9\") pod \"crc-debug-dkr9g\" (UID: \"10fb0d06-4f2b-44c2-aebf-7075d918b3e2\") " pod="openshift-must-gather-pv922/crc-debug-dkr9g" Feb 25 10:18:19 crc kubenswrapper[4978]: I0225 10:18:19.150234 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-host\") pod \"crc-debug-dkr9g\" (UID: \"10fb0d06-4f2b-44c2-aebf-7075d918b3e2\") " pod="openshift-must-gather-pv922/crc-debug-dkr9g" Feb 25 10:18:19 crc kubenswrapper[4978]: I0225 10:18:19.177798 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmxz9\" (UniqueName: \"kubernetes.io/projected/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-kube-api-access-lmxz9\") pod \"crc-debug-dkr9g\" (UID: \"10fb0d06-4f2b-44c2-aebf-7075d918b3e2\") " pod="openshift-must-gather-pv922/crc-debug-dkr9g" Feb 25 10:18:19 crc kubenswrapper[4978]: I0225 10:18:19.275736 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-dkr9g" Feb 25 10:18:19 crc kubenswrapper[4978]: W0225 10:18:19.313982 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10fb0d06_4f2b_44c2_aebf_7075d918b3e2.slice/crio-3e2b8d3ea4d35884511ebd7aa83c20cec922ae2ce1eef04369c5f9446cc418e2 WatchSource:0}: Error finding container 3e2b8d3ea4d35884511ebd7aa83c20cec922ae2ce1eef04369c5f9446cc418e2: Status 404 returned error can't find the container with id 3e2b8d3ea4d35884511ebd7aa83c20cec922ae2ce1eef04369c5f9446cc418e2 Feb 25 10:18:19 crc kubenswrapper[4978]: I0225 10:18:19.346992 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c8b2b7d-b132-4dce-9b57-9edca2201cc2" path="/var/lib/kubelet/pods/4c8b2b7d-b132-4dce-9b57-9edca2201cc2/volumes" Feb 25 10:18:19 crc kubenswrapper[4978]: I0225 10:18:19.435163 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/crc-debug-dkr9g" event={"ID":"10fb0d06-4f2b-44c2-aebf-7075d918b3e2","Type":"ContainerStarted","Data":"3e2b8d3ea4d35884511ebd7aa83c20cec922ae2ce1eef04369c5f9446cc418e2"} Feb 25 10:18:20 crc kubenswrapper[4978]: I0225 10:18:20.445044 4978 generic.go:334] "Generic (PLEG): container finished" podID="10fb0d06-4f2b-44c2-aebf-7075d918b3e2" containerID="a195044fb5c61a961d79848819aa5025ad19ca13589da9a8f91eadc29a2167a4" exitCode=0 Feb 25 10:18:20 crc kubenswrapper[4978]: I0225 10:18:20.445103 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/crc-debug-dkr9g" event={"ID":"10fb0d06-4f2b-44c2-aebf-7075d918b3e2","Type":"ContainerDied","Data":"a195044fb5c61a961d79848819aa5025ad19ca13589da9a8f91eadc29a2167a4"} Feb 25 10:18:21 crc kubenswrapper[4978]: I0225 10:18:21.598249 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-dkr9g" Feb 25 10:18:21 crc kubenswrapper[4978]: I0225 10:18:21.615870 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmxz9\" (UniqueName: \"kubernetes.io/projected/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-kube-api-access-lmxz9\") pod \"10fb0d06-4f2b-44c2-aebf-7075d918b3e2\" (UID: \"10fb0d06-4f2b-44c2-aebf-7075d918b3e2\") " Feb 25 10:18:21 crc kubenswrapper[4978]: I0225 10:18:21.616113 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-host\") pod \"10fb0d06-4f2b-44c2-aebf-7075d918b3e2\" (UID: \"10fb0d06-4f2b-44c2-aebf-7075d918b3e2\") " Feb 25 10:18:21 crc kubenswrapper[4978]: I0225 10:18:21.616548 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-host" (OuterVolumeSpecName: "host") pod "10fb0d06-4f2b-44c2-aebf-7075d918b3e2" (UID: "10fb0d06-4f2b-44c2-aebf-7075d918b3e2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 10:18:21 crc kubenswrapper[4978]: I0225 10:18:21.622125 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-kube-api-access-lmxz9" (OuterVolumeSpecName: "kube-api-access-lmxz9") pod "10fb0d06-4f2b-44c2-aebf-7075d918b3e2" (UID: "10fb0d06-4f2b-44c2-aebf-7075d918b3e2"). InnerVolumeSpecName "kube-api-access-lmxz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:18:21 crc kubenswrapper[4978]: I0225 10:18:21.718093 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmxz9\" (UniqueName: \"kubernetes.io/projected/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-kube-api-access-lmxz9\") on node \"crc\" DevicePath \"\"" Feb 25 10:18:21 crc kubenswrapper[4978]: I0225 10:18:21.718127 4978 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/10fb0d06-4f2b-44c2-aebf-7075d918b3e2-host\") on node \"crc\" DevicePath \"\"" Feb 25 10:18:22 crc kubenswrapper[4978]: I0225 10:18:22.470701 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/crc-debug-dkr9g" event={"ID":"10fb0d06-4f2b-44c2-aebf-7075d918b3e2","Type":"ContainerDied","Data":"3e2b8d3ea4d35884511ebd7aa83c20cec922ae2ce1eef04369c5f9446cc418e2"} Feb 25 10:18:22 crc kubenswrapper[4978]: I0225 10:18:22.470744 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e2b8d3ea4d35884511ebd7aa83c20cec922ae2ce1eef04369c5f9446cc418e2" Feb 25 10:18:22 crc kubenswrapper[4978]: I0225 10:18:22.470808 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-dkr9g" Feb 25 10:18:23 crc kubenswrapper[4978]: I0225 10:18:23.240700 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pv922/crc-debug-dkr9g"] Feb 25 10:18:23 crc kubenswrapper[4978]: I0225 10:18:23.253406 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pv922/crc-debug-dkr9g"] Feb 25 10:18:23 crc kubenswrapper[4978]: I0225 10:18:23.338841 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10fb0d06-4f2b-44c2-aebf-7075d918b3e2" path="/var/lib/kubelet/pods/10fb0d06-4f2b-44c2-aebf-7075d918b3e2/volumes" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.538215 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pv922/crc-debug-66l97"] Feb 25 10:18:24 crc kubenswrapper[4978]: E0225 10:18:24.538737 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10fb0d06-4f2b-44c2-aebf-7075d918b3e2" containerName="container-00" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.538750 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="10fb0d06-4f2b-44c2-aebf-7075d918b3e2" containerName="container-00" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.538999 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="10fb0d06-4f2b-44c2-aebf-7075d918b3e2" containerName="container-00" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.542396 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-66l97" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.575570 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rznnh\" (UniqueName: \"kubernetes.io/projected/0fac5539-ba26-4ff9-b435-3c91b18c5e19-kube-api-access-rznnh\") pod \"crc-debug-66l97\" (UID: \"0fac5539-ba26-4ff9-b435-3c91b18c5e19\") " pod="openshift-must-gather-pv922/crc-debug-66l97" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.575620 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0fac5539-ba26-4ff9-b435-3c91b18c5e19-host\") pod \"crc-debug-66l97\" (UID: \"0fac5539-ba26-4ff9-b435-3c91b18c5e19\") " pod="openshift-must-gather-pv922/crc-debug-66l97" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.677430 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rznnh\" (UniqueName: \"kubernetes.io/projected/0fac5539-ba26-4ff9-b435-3c91b18c5e19-kube-api-access-rznnh\") pod \"crc-debug-66l97\" (UID: \"0fac5539-ba26-4ff9-b435-3c91b18c5e19\") " pod="openshift-must-gather-pv922/crc-debug-66l97" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.677507 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0fac5539-ba26-4ff9-b435-3c91b18c5e19-host\") pod \"crc-debug-66l97\" (UID: \"0fac5539-ba26-4ff9-b435-3c91b18c5e19\") " pod="openshift-must-gather-pv922/crc-debug-66l97" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.677794 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0fac5539-ba26-4ff9-b435-3c91b18c5e19-host\") pod \"crc-debug-66l97\" (UID: \"0fac5539-ba26-4ff9-b435-3c91b18c5e19\") " pod="openshift-must-gather-pv922/crc-debug-66l97" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.715055 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rznnh\" (UniqueName: \"kubernetes.io/projected/0fac5539-ba26-4ff9-b435-3c91b18c5e19-kube-api-access-rznnh\") pod \"crc-debug-66l97\" (UID: \"0fac5539-ba26-4ff9-b435-3c91b18c5e19\") " pod="openshift-must-gather-pv922/crc-debug-66l97" Feb 25 10:18:24 crc kubenswrapper[4978]: I0225 10:18:24.862012 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-66l97" Feb 25 10:18:25 crc kubenswrapper[4978]: I0225 10:18:25.500660 4978 generic.go:334] "Generic (PLEG): container finished" podID="0fac5539-ba26-4ff9-b435-3c91b18c5e19" containerID="a01edfc4d9d6d0b3e5f03e3ecbf5a390dca2759bd9aab8eca0abcda1a7afb972" exitCode=0 Feb 25 10:18:25 crc kubenswrapper[4978]: I0225 10:18:25.500743 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/crc-debug-66l97" event={"ID":"0fac5539-ba26-4ff9-b435-3c91b18c5e19","Type":"ContainerDied","Data":"a01edfc4d9d6d0b3e5f03e3ecbf5a390dca2759bd9aab8eca0abcda1a7afb972"} Feb 25 10:18:25 crc kubenswrapper[4978]: I0225 10:18:25.501008 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/crc-debug-66l97" event={"ID":"0fac5539-ba26-4ff9-b435-3c91b18c5e19","Type":"ContainerStarted","Data":"a9aa590c7479854eca840b8839995143150737be599c0573a66f282ebdaecdcd"} Feb 25 10:18:25 crc kubenswrapper[4978]: I0225 10:18:25.572114 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pv922/crc-debug-66l97"] Feb 25 10:18:25 crc kubenswrapper[4978]: I0225 10:18:25.598975 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pv922/crc-debug-66l97"] Feb 25 10:18:26 crc kubenswrapper[4978]: I0225 10:18:26.644420 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-66l97" Feb 25 10:18:26 crc kubenswrapper[4978]: I0225 10:18:26.718180 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0fac5539-ba26-4ff9-b435-3c91b18c5e19-host\") pod \"0fac5539-ba26-4ff9-b435-3c91b18c5e19\" (UID: \"0fac5539-ba26-4ff9-b435-3c91b18c5e19\") " Feb 25 10:18:26 crc kubenswrapper[4978]: I0225 10:18:26.718237 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rznnh\" (UniqueName: \"kubernetes.io/projected/0fac5539-ba26-4ff9-b435-3c91b18c5e19-kube-api-access-rznnh\") pod \"0fac5539-ba26-4ff9-b435-3c91b18c5e19\" (UID: \"0fac5539-ba26-4ff9-b435-3c91b18c5e19\") " Feb 25 10:18:26 crc kubenswrapper[4978]: I0225 10:18:26.720105 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0fac5539-ba26-4ff9-b435-3c91b18c5e19-host" (OuterVolumeSpecName: "host") pod "0fac5539-ba26-4ff9-b435-3c91b18c5e19" (UID: "0fac5539-ba26-4ff9-b435-3c91b18c5e19"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 25 10:18:26 crc kubenswrapper[4978]: I0225 10:18:26.740569 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fac5539-ba26-4ff9-b435-3c91b18c5e19-kube-api-access-rznnh" (OuterVolumeSpecName: "kube-api-access-rznnh") pod "0fac5539-ba26-4ff9-b435-3c91b18c5e19" (UID: "0fac5539-ba26-4ff9-b435-3c91b18c5e19"). InnerVolumeSpecName "kube-api-access-rznnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:18:26 crc kubenswrapper[4978]: I0225 10:18:26.821267 4978 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0fac5539-ba26-4ff9-b435-3c91b18c5e19-host\") on node \"crc\" DevicePath \"\"" Feb 25 10:18:26 crc kubenswrapper[4978]: I0225 10:18:26.821320 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rznnh\" (UniqueName: \"kubernetes.io/projected/0fac5539-ba26-4ff9-b435-3c91b18c5e19-kube-api-access-rznnh\") on node \"crc\" DevicePath \"\"" Feb 25 10:18:27 crc kubenswrapper[4978]: I0225 10:18:27.340013 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fac5539-ba26-4ff9-b435-3c91b18c5e19" path="/var/lib/kubelet/pods/0fac5539-ba26-4ff9-b435-3c91b18c5e19/volumes" Feb 25 10:18:27 crc kubenswrapper[4978]: I0225 10:18:27.543912 4978 scope.go:117] "RemoveContainer" containerID="a01edfc4d9d6d0b3e5f03e3ecbf5a390dca2759bd9aab8eca0abcda1a7afb972" Feb 25 10:18:27 crc kubenswrapper[4978]: I0225 10:18:27.544022 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/crc-debug-66l97" Feb 25 10:18:28 crc kubenswrapper[4978]: I0225 10:18:28.328160 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:18:28 crc kubenswrapper[4978]: E0225 10:18:28.328707 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:18:40 crc kubenswrapper[4978]: I0225 10:18:40.327769 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:18:40 crc kubenswrapper[4978]: E0225 10:18:40.328531 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:18:44 crc kubenswrapper[4978]: I0225 10:18:44.091739 4978 scope.go:117] "RemoveContainer" containerID="ff50c432bc39c99cffc609f82e85ec6305669b1902f30792897da74b6a2e2b54" Feb 25 10:18:51 crc kubenswrapper[4978]: I0225 10:18:51.327397 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:18:51 crc kubenswrapper[4978]: E0225 10:18:51.328385 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:19:02 crc kubenswrapper[4978]: I0225 10:19:02.327661 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:19:02 crc kubenswrapper[4978]: E0225 10:19:02.328273 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.623361 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dncwv"] Feb 25 10:19:11 crc kubenswrapper[4978]: E0225 10:19:11.624287 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fac5539-ba26-4ff9-b435-3c91b18c5e19" containerName="container-00" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.624300 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fac5539-ba26-4ff9-b435-3c91b18c5e19" containerName="container-00" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.624569 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fac5539-ba26-4ff9-b435-3c91b18c5e19" containerName="container-00" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.626117 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.635572 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dncwv"] Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.754856 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01c8bc46-7771-4b6a-b27c-49a507df6627-utilities\") pod \"certified-operators-dncwv\" (UID: \"01c8bc46-7771-4b6a-b27c-49a507df6627\") " pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.755049 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw69l\" (UniqueName: \"kubernetes.io/projected/01c8bc46-7771-4b6a-b27c-49a507df6627-kube-api-access-xw69l\") pod \"certified-operators-dncwv\" (UID: \"01c8bc46-7771-4b6a-b27c-49a507df6627\") " pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.755285 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01c8bc46-7771-4b6a-b27c-49a507df6627-catalog-content\") pod \"certified-operators-dncwv\" (UID: \"01c8bc46-7771-4b6a-b27c-49a507df6627\") " pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.857063 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01c8bc46-7771-4b6a-b27c-49a507df6627-catalog-content\") pod \"certified-operators-dncwv\" (UID: \"01c8bc46-7771-4b6a-b27c-49a507df6627\") " pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.857857 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01c8bc46-7771-4b6a-b27c-49a507df6627-utilities\") pod \"certified-operators-dncwv\" (UID: \"01c8bc46-7771-4b6a-b27c-49a507df6627\") " pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.857908 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw69l\" (UniqueName: \"kubernetes.io/projected/01c8bc46-7771-4b6a-b27c-49a507df6627-kube-api-access-xw69l\") pod \"certified-operators-dncwv\" (UID: \"01c8bc46-7771-4b6a-b27c-49a507df6627\") " pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.857727 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01c8bc46-7771-4b6a-b27c-49a507df6627-catalog-content\") pod \"certified-operators-dncwv\" (UID: \"01c8bc46-7771-4b6a-b27c-49a507df6627\") " pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.858563 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01c8bc46-7771-4b6a-b27c-49a507df6627-utilities\") pod \"certified-operators-dncwv\" (UID: \"01c8bc46-7771-4b6a-b27c-49a507df6627\") " pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.876518 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw69l\" (UniqueName: \"kubernetes.io/projected/01c8bc46-7771-4b6a-b27c-49a507df6627-kube-api-access-xw69l\") pod \"certified-operators-dncwv\" (UID: \"01c8bc46-7771-4b6a-b27c-49a507df6627\") " pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:11 crc kubenswrapper[4978]: I0225 10:19:11.960032 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:12 crc kubenswrapper[4978]: I0225 10:19:12.665955 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dncwv"] Feb 25 10:19:13 crc kubenswrapper[4978]: I0225 10:19:13.012003 4978 generic.go:334] "Generic (PLEG): container finished" podID="01c8bc46-7771-4b6a-b27c-49a507df6627" containerID="9f48bd76cf19df100b28c9c9e79348700f2c8bb917fa874f6727d61b7bf3c4c0" exitCode=0 Feb 25 10:19:13 crc kubenswrapper[4978]: I0225 10:19:13.012265 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dncwv" event={"ID":"01c8bc46-7771-4b6a-b27c-49a507df6627","Type":"ContainerDied","Data":"9f48bd76cf19df100b28c9c9e79348700f2c8bb917fa874f6727d61b7bf3c4c0"} Feb 25 10:19:13 crc kubenswrapper[4978]: I0225 10:19:13.012652 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dncwv" event={"ID":"01c8bc46-7771-4b6a-b27c-49a507df6627","Type":"ContainerStarted","Data":"1787ed3bf7f0004066b9210511a2d36b7965f2442e31f6ae89811ebbc5868bef"} Feb 25 10:19:15 crc kubenswrapper[4978]: I0225 10:19:15.328802 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:19:15 crc kubenswrapper[4978]: E0225 10:19:15.329317 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:19:21 crc kubenswrapper[4978]: I0225 10:19:21.118872 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dncwv" event={"ID":"01c8bc46-7771-4b6a-b27c-49a507df6627","Type":"ContainerStarted","Data":"094bff693783946639f46fbc4eb40f33eaf91e66afd4490f60b26a780a7953a4"} Feb 25 10:19:22 crc kubenswrapper[4978]: I0225 10:19:22.132211 4978 generic.go:334] "Generic (PLEG): container finished" podID="01c8bc46-7771-4b6a-b27c-49a507df6627" containerID="094bff693783946639f46fbc4eb40f33eaf91e66afd4490f60b26a780a7953a4" exitCode=0 Feb 25 10:19:22 crc kubenswrapper[4978]: I0225 10:19:22.132439 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dncwv" event={"ID":"01c8bc46-7771-4b6a-b27c-49a507df6627","Type":"ContainerDied","Data":"094bff693783946639f46fbc4eb40f33eaf91e66afd4490f60b26a780a7953a4"} Feb 25 10:19:23 crc kubenswrapper[4978]: I0225 10:19:23.146056 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dncwv" event={"ID":"01c8bc46-7771-4b6a-b27c-49a507df6627","Type":"ContainerStarted","Data":"72cafb6e6c39e2ece5290e3f3d41242ba8eb0c865e1887a9b908d21e4986dce1"} Feb 25 10:19:23 crc kubenswrapper[4978]: I0225 10:19:23.162940 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dncwv" podStartSLOduration=2.296230513 podStartE2EDuration="12.16292526s" podCreationTimestamp="2026-02-25 10:19:11 +0000 UTC" firstStartedPulling="2026-02-25 10:19:13.013607402 +0000 UTC m=+12846.452863861" lastFinishedPulling="2026-02-25 10:19:22.880302139 +0000 UTC m=+12856.319558608" observedRunningTime="2026-02-25 10:19:23.162537338 +0000 UTC m=+12856.601793817" watchObservedRunningTime="2026-02-25 10:19:23.16292526 +0000 UTC m=+12856.602181719" Feb 25 10:19:28 crc kubenswrapper[4978]: I0225 10:19:28.328160 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:19:28 crc kubenswrapper[4978]: E0225 10:19:28.329066 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:19:31 crc kubenswrapper[4978]: I0225 10:19:31.960566 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:31 crc kubenswrapper[4978]: I0225 10:19:31.961704 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:32 crc kubenswrapper[4978]: I0225 10:19:32.016831 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:32 crc kubenswrapper[4978]: I0225 10:19:32.297311 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dncwv" Feb 25 10:19:32 crc kubenswrapper[4978]: I0225 10:19:32.391313 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dncwv"] Feb 25 10:19:32 crc kubenswrapper[4978]: I0225 10:19:32.446243 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6qjw5"] Feb 25 10:19:32 crc kubenswrapper[4978]: I0225 10:19:32.446974 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6qjw5" podUID="969040a8-e02f-4b30-a579-db2185d5734d" containerName="registry-server" containerID="cri-o://3dea3d3c0d9e14a8c98cf34b101bdeb6cc5f7535af8d47913aa5aac799d38aed" gracePeriod=2 Feb 25 10:19:33 crc kubenswrapper[4978]: I0225 10:19:33.247583 4978 generic.go:334] "Generic (PLEG): container finished" podID="969040a8-e02f-4b30-a579-db2185d5734d" containerID="3dea3d3c0d9e14a8c98cf34b101bdeb6cc5f7535af8d47913aa5aac799d38aed" exitCode=0 Feb 25 10:19:33 crc kubenswrapper[4978]: I0225 10:19:33.247742 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qjw5" event={"ID":"969040a8-e02f-4b30-a579-db2185d5734d","Type":"ContainerDied","Data":"3dea3d3c0d9e14a8c98cf34b101bdeb6cc5f7535af8d47913aa5aac799d38aed"} Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.149633 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.213973 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdxcn\" (UniqueName: \"kubernetes.io/projected/969040a8-e02f-4b30-a579-db2185d5734d-kube-api-access-wdxcn\") pod \"969040a8-e02f-4b30-a579-db2185d5734d\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.214153 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-catalog-content\") pod \"969040a8-e02f-4b30-a579-db2185d5734d\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.214257 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-utilities\") pod \"969040a8-e02f-4b30-a579-db2185d5734d\" (UID: \"969040a8-e02f-4b30-a579-db2185d5734d\") " Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.215299 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-utilities" (OuterVolumeSpecName: "utilities") pod "969040a8-e02f-4b30-a579-db2185d5734d" (UID: "969040a8-e02f-4b30-a579-db2185d5734d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.221953 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/969040a8-e02f-4b30-a579-db2185d5734d-kube-api-access-wdxcn" (OuterVolumeSpecName: "kube-api-access-wdxcn") pod "969040a8-e02f-4b30-a579-db2185d5734d" (UID: "969040a8-e02f-4b30-a579-db2185d5734d"). InnerVolumeSpecName "kube-api-access-wdxcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.261284 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qjw5" event={"ID":"969040a8-e02f-4b30-a579-db2185d5734d","Type":"ContainerDied","Data":"5d26afefe13212ac7189c4b0238a37c910b7eb4f2e807ec78f6b502dd2135712"} Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.261390 4978 scope.go:117] "RemoveContainer" containerID="3dea3d3c0d9e14a8c98cf34b101bdeb6cc5f7535af8d47913aa5aac799d38aed" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.261309 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qjw5" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.279033 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "969040a8-e02f-4b30-a579-db2185d5734d" (UID: "969040a8-e02f-4b30-a579-db2185d5734d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.301533 4978 scope.go:117] "RemoveContainer" containerID="91cfbb0c04743a40f7bf750a5c1bcec6b8dd0f811d9f390bee40b1fcb05235bf" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.317765 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdxcn\" (UniqueName: \"kubernetes.io/projected/969040a8-e02f-4b30-a579-db2185d5734d-kube-api-access-wdxcn\") on node \"crc\" DevicePath \"\"" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.317807 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.317817 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/969040a8-e02f-4b30-a579-db2185d5734d-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.326796 4978 scope.go:117] "RemoveContainer" containerID="28d0c5c859ce593fdb9f9d1d0633804dc1d61a049ecb727639adae54c298711e" Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.594985 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6qjw5"] Feb 25 10:19:34 crc kubenswrapper[4978]: I0225 10:19:34.607543 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6qjw5"] Feb 25 10:19:35 crc kubenswrapper[4978]: I0225 10:19:35.338168 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="969040a8-e02f-4b30-a579-db2185d5734d" path="/var/lib/kubelet/pods/969040a8-e02f-4b30-a579-db2185d5734d/volumes" Feb 25 10:19:43 crc kubenswrapper[4978]: I0225 10:19:43.334196 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:19:43 crc kubenswrapper[4978]: E0225 10:19:43.339554 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:19:54 crc kubenswrapper[4978]: I0225 10:19:54.328325 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:19:54 crc kubenswrapper[4978]: E0225 10:19:54.329167 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.167280 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533580-4f8k8"] Feb 25 10:20:00 crc kubenswrapper[4978]: E0225 10:20:00.169533 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="969040a8-e02f-4b30-a579-db2185d5734d" containerName="extract-content" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.169639 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="969040a8-e02f-4b30-a579-db2185d5734d" containerName="extract-content" Feb 25 10:20:00 crc kubenswrapper[4978]: E0225 10:20:00.169777 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="969040a8-e02f-4b30-a579-db2185d5734d" containerName="extract-utilities" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.169862 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="969040a8-e02f-4b30-a579-db2185d5734d" containerName="extract-utilities" Feb 25 10:20:00 crc kubenswrapper[4978]: E0225 10:20:00.169947 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="969040a8-e02f-4b30-a579-db2185d5734d" containerName="registry-server" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.170029 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="969040a8-e02f-4b30-a579-db2185d5734d" containerName="registry-server" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.170363 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="969040a8-e02f-4b30-a579-db2185d5734d" containerName="registry-server" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.171240 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533580-4f8k8" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.173360 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.175056 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.178197 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.182340 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533580-4f8k8"] Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.276636 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kh47\" (UniqueName: \"kubernetes.io/projected/6d4aee5d-a45a-4672-ae14-e1af0287f17d-kube-api-access-9kh47\") pod \"auto-csr-approver-29533580-4f8k8\" (UID: \"6d4aee5d-a45a-4672-ae14-e1af0287f17d\") " pod="openshift-infra/auto-csr-approver-29533580-4f8k8" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.379012 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kh47\" (UniqueName: \"kubernetes.io/projected/6d4aee5d-a45a-4672-ae14-e1af0287f17d-kube-api-access-9kh47\") pod \"auto-csr-approver-29533580-4f8k8\" (UID: \"6d4aee5d-a45a-4672-ae14-e1af0287f17d\") " pod="openshift-infra/auto-csr-approver-29533580-4f8k8" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.401666 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kh47\" (UniqueName: \"kubernetes.io/projected/6d4aee5d-a45a-4672-ae14-e1af0287f17d-kube-api-access-9kh47\") pod \"auto-csr-approver-29533580-4f8k8\" (UID: \"6d4aee5d-a45a-4672-ae14-e1af0287f17d\") " pod="openshift-infra/auto-csr-approver-29533580-4f8k8" Feb 25 10:20:00 crc kubenswrapper[4978]: I0225 10:20:00.499660 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533580-4f8k8" Feb 25 10:20:01 crc kubenswrapper[4978]: I0225 10:20:01.261214 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533580-4f8k8"] Feb 25 10:20:01 crc kubenswrapper[4978]: I0225 10:20:01.577967 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533580-4f8k8" event={"ID":"6d4aee5d-a45a-4672-ae14-e1af0287f17d","Type":"ContainerStarted","Data":"5688d2a8344d477f132c484f910e1e3bc7014029904b57a44e294a7254e43063"} Feb 25 10:20:02 crc kubenswrapper[4978]: I0225 10:20:02.608833 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533580-4f8k8" podStartSLOduration=1.6237309629999999 podStartE2EDuration="2.60881203s" podCreationTimestamp="2026-02-25 10:20:00 +0000 UTC" firstStartedPulling="2026-02-25 10:20:01.272139984 +0000 UTC m=+12894.711396453" lastFinishedPulling="2026-02-25 10:20:02.257221051 +0000 UTC m=+12895.696477520" observedRunningTime="2026-02-25 10:20:02.607501219 +0000 UTC m=+12896.046757688" watchObservedRunningTime="2026-02-25 10:20:02.60881203 +0000 UTC m=+12896.048068489" Feb 25 10:20:03 crc kubenswrapper[4978]: I0225 10:20:03.608810 4978 generic.go:334] "Generic (PLEG): container finished" podID="6d4aee5d-a45a-4672-ae14-e1af0287f17d" containerID="91092ca3188fe8ae4937f7f257581369ac360c51c531b685ae629e814456c6ba" exitCode=0 Feb 25 10:20:03 crc kubenswrapper[4978]: I0225 10:20:03.608927 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533580-4f8k8" event={"ID":"6d4aee5d-a45a-4672-ae14-e1af0287f17d","Type":"ContainerDied","Data":"91092ca3188fe8ae4937f7f257581369ac360c51c531b685ae629e814456c6ba"} Feb 25 10:20:06 crc kubenswrapper[4978]: I0225 10:20:06.062623 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533580-4f8k8" Feb 25 10:20:06 crc kubenswrapper[4978]: I0225 10:20:06.209311 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kh47\" (UniqueName: \"kubernetes.io/projected/6d4aee5d-a45a-4672-ae14-e1af0287f17d-kube-api-access-9kh47\") pod \"6d4aee5d-a45a-4672-ae14-e1af0287f17d\" (UID: \"6d4aee5d-a45a-4672-ae14-e1af0287f17d\") " Feb 25 10:20:06 crc kubenswrapper[4978]: I0225 10:20:06.214684 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d4aee5d-a45a-4672-ae14-e1af0287f17d-kube-api-access-9kh47" (OuterVolumeSpecName: "kube-api-access-9kh47") pod "6d4aee5d-a45a-4672-ae14-e1af0287f17d" (UID: "6d4aee5d-a45a-4672-ae14-e1af0287f17d"). InnerVolumeSpecName "kube-api-access-9kh47". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:20:06 crc kubenswrapper[4978]: I0225 10:20:06.312960 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kh47\" (UniqueName: \"kubernetes.io/projected/6d4aee5d-a45a-4672-ae14-e1af0287f17d-kube-api-access-9kh47\") on node \"crc\" DevicePath \"\"" Feb 25 10:20:06 crc kubenswrapper[4978]: I0225 10:20:06.645803 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533580-4f8k8" event={"ID":"6d4aee5d-a45a-4672-ae14-e1af0287f17d","Type":"ContainerDied","Data":"5688d2a8344d477f132c484f910e1e3bc7014029904b57a44e294a7254e43063"} Feb 25 10:20:06 crc kubenswrapper[4978]: I0225 10:20:06.645851 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5688d2a8344d477f132c484f910e1e3bc7014029904b57a44e294a7254e43063" Feb 25 10:20:06 crc kubenswrapper[4978]: I0225 10:20:06.645858 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533580-4f8k8" Feb 25 10:20:07 crc kubenswrapper[4978]: I0225 10:20:07.148498 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533574-7tkgm"] Feb 25 10:20:07 crc kubenswrapper[4978]: I0225 10:20:07.161494 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533574-7tkgm"] Feb 25 10:20:07 crc kubenswrapper[4978]: I0225 10:20:07.350633 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0" path="/var/lib/kubelet/pods/cfa31967-bff8-4c3f-a01f-3fbc5a27e9d0/volumes" Feb 25 10:20:08 crc kubenswrapper[4978]: I0225 10:20:08.327617 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:20:08 crc kubenswrapper[4978]: E0225 10:20:08.328008 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:20:19 crc kubenswrapper[4978]: I0225 10:20:19.328243 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:20:19 crc kubenswrapper[4978]: E0225 10:20:19.329203 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:20:32 crc kubenswrapper[4978]: I0225 10:20:32.328412 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:20:32 crc kubenswrapper[4978]: E0225 10:20:32.329211 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:20:44 crc kubenswrapper[4978]: I0225 10:20:44.235339 4978 scope.go:117] "RemoveContainer" containerID="46c2f61669395902edc871a4076d335d3c788e4a027e1862e37f620c3c4627d5" Feb 25 10:20:46 crc kubenswrapper[4978]: I0225 10:20:46.329113 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:20:46 crc kubenswrapper[4978]: E0225 10:20:46.329664 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:20:58 crc kubenswrapper[4978]: I0225 10:20:58.328135 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:20:58 crc kubenswrapper[4978]: E0225 10:20:58.328863 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:21:12 crc kubenswrapper[4978]: I0225 10:21:12.483997 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_55251903-0563-4a00-8652-96350cc28320/init-config-reloader/0.log" Feb 25 10:21:12 crc kubenswrapper[4978]: I0225 10:21:12.709311 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_55251903-0563-4a00-8652-96350cc28320/init-config-reloader/0.log" Feb 25 10:21:12 crc kubenswrapper[4978]: I0225 10:21:12.739275 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_55251903-0563-4a00-8652-96350cc28320/alertmanager/0.log" Feb 25 10:21:12 crc kubenswrapper[4978]: I0225 10:21:12.798977 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_55251903-0563-4a00-8652-96350cc28320/config-reloader/0.log" Feb 25 10:21:12 crc kubenswrapper[4978]: I0225 10:21:12.985594 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_9e08daef-04f0-4303-80a4-a8b21bf0da3d/aodh-api/0.log" Feb 25 10:21:13 crc kubenswrapper[4978]: I0225 10:21:13.023337 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_9e08daef-04f0-4303-80a4-a8b21bf0da3d/aodh-listener/0.log" Feb 25 10:21:13 crc kubenswrapper[4978]: I0225 10:21:13.085318 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_9e08daef-04f0-4303-80a4-a8b21bf0da3d/aodh-evaluator/0.log" Feb 25 10:21:13 crc kubenswrapper[4978]: I0225 10:21:13.140067 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_9e08daef-04f0-4303-80a4-a8b21bf0da3d/aodh-notifier/0.log" Feb 25 10:21:13 crc kubenswrapper[4978]: I0225 10:21:13.327303 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:21:13 crc kubenswrapper[4978]: E0225 10:21:13.327599 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:21:13 crc kubenswrapper[4978]: I0225 10:21:13.329013 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c85b7bd74-wf92q_38283175-934a-4e62-a5af-d873db0c22aa/barbican-api/0.log" Feb 25 10:21:13 crc kubenswrapper[4978]: I0225 10:21:13.352352 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-c85b7bd74-wf92q_38283175-934a-4e62-a5af-d873db0c22aa/barbican-api-log/0.log" Feb 25 10:21:13 crc kubenswrapper[4978]: I0225 10:21:13.541758 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5684bb8b46-xpzqh_0648e0c9-8537-4093-9f70-f91e62f1ff62/barbican-keystone-listener/0.log" Feb 25 10:21:13 crc kubenswrapper[4978]: I0225 10:21:13.713454 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f99df5645-r62ts_7128dcbe-8297-4c05-ae36-ec9d12364871/barbican-worker/0.log" Feb 25 10:21:13 crc kubenswrapper[4978]: I0225 10:21:13.952587 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5684bb8b46-xpzqh_0648e0c9-8537-4093-9f70-f91e62f1ff62/barbican-keystone-listener-log/0.log" Feb 25 10:21:13 crc kubenswrapper[4978]: I0225 10:21:13.981560 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f99df5645-r62ts_7128dcbe-8297-4c05-ae36-ec9d12364871/barbican-worker-log/0.log" Feb 25 10:21:14 crc kubenswrapper[4978]: I0225 10:21:14.084734 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-jjn5l_060597fe-f0f4-45f6-bea3-5faa8b657c41/bootstrap-openstack-openstack-cell1/0.log" Feb 25 10:21:14 crc kubenswrapper[4978]: I0225 10:21:14.259723 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-networker-vhgfz_c811c196-e2b2-4230-b91d-89c22efbd036/bootstrap-openstack-openstack-networker/0.log" Feb 25 10:21:14 crc kubenswrapper[4978]: I0225 10:21:14.390946 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41cbe834-6422-4dc3-b32f-f97042673369/ceilometer-central-agent/0.log" Feb 25 10:21:14 crc kubenswrapper[4978]: I0225 10:21:14.701033 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41cbe834-6422-4dc3-b32f-f97042673369/sg-core/0.log" Feb 25 10:21:14 crc kubenswrapper[4978]: I0225 10:21:14.773747 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41cbe834-6422-4dc3-b32f-f97042673369/proxy-httpd/0.log" Feb 25 10:21:15 crc kubenswrapper[4978]: I0225 10:21:15.106249 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_038563a2-70b7-4050-9183-a4af40007288/cinder-api-log/0.log" Feb 25 10:21:15 crc kubenswrapper[4978]: I0225 10:21:15.160724 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_038563a2-70b7-4050-9183-a4af40007288/cinder-api/0.log" Feb 25 10:21:15 crc kubenswrapper[4978]: I0225 10:21:15.184018 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_41cbe834-6422-4dc3-b32f-f97042673369/ceilometer-notification-agent/0.log" Feb 25 10:21:15 crc kubenswrapper[4978]: I0225 10:21:15.377192 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2a2818b2-3da2-45cb-b9b1-d27341a11934/cinder-scheduler/0.log" Feb 25 10:21:15 crc kubenswrapper[4978]: I0225 10:21:15.504289 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_2a2818b2-3da2-45cb-b9b1-d27341a11934/probe/0.log" Feb 25 10:21:15 crc kubenswrapper[4978]: I0225 10:21:15.623080 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-gc692_03aef306-3186-4b8a-9a1e-523ebca2e8fc/configure-network-openstack-openstack-cell1/0.log" Feb 25 10:21:15 crc kubenswrapper[4978]: I0225 10:21:15.826395 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-networker-4nb7l_04b8c344-5cce-4f9c-9e0b-181950cb4e8b/configure-network-openstack-openstack-networker/0.log" Feb 25 10:21:15 crc kubenswrapper[4978]: I0225 10:21:15.979307 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-xbb8f_07f0da72-81c7-4452-8662-bc63584cad91/configure-os-openstack-openstack-cell1/0.log" Feb 25 10:21:16 crc kubenswrapper[4978]: I0225 10:21:16.248704 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-networker-lgws7_ec14586b-e00a-4d26-9afc-87ef605ddfc4/configure-os-openstack-openstack-networker/0.log" Feb 25 10:21:16 crc kubenswrapper[4978]: I0225 10:21:16.270670 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6997548745-tb5n6_605215fb-7a4c-49fb-a715-5c7251cad82e/init/0.log" Feb 25 10:21:16 crc kubenswrapper[4978]: I0225 10:21:16.619963 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-hzw4k_99ce1a38-0659-4c3d-be3f-5311d2a4cd8e/download-cache-openstack-openstack-cell1/0.log" Feb 25 10:21:16 crc kubenswrapper[4978]: I0225 10:21:16.720463 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6997548745-tb5n6_605215fb-7a4c-49fb-a715-5c7251cad82e/init/0.log" Feb 25 10:21:17 crc kubenswrapper[4978]: I0225 10:21:17.486672 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-networker-77j4w_c8c5a88a-2673-4723-8967-9f209d416174/download-cache-openstack-openstack-networker/0.log" Feb 25 10:21:17 crc kubenswrapper[4978]: I0225 10:21:17.816638 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6bd498f7-e4ee-4062-8db7-cade3fe823af/glance-log/0.log" Feb 25 10:21:17 crc kubenswrapper[4978]: I0225 10:21:17.981393 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_84c35447-f585-4c1d-b04e-a0e6a86b8541/glance-httpd/0.log" Feb 25 10:21:18 crc kubenswrapper[4978]: I0225 10:21:18.056310 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6bd498f7-e4ee-4062-8db7-cade3fe823af/glance-httpd/0.log" Feb 25 10:21:18 crc kubenswrapper[4978]: I0225 10:21:18.137185 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_84c35447-f585-4c1d-b04e-a0e6a86b8541/glance-log/0.log" Feb 25 10:21:18 crc kubenswrapper[4978]: I0225 10:21:18.905607 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-7d87784d56-fghgj_7ffd8086-b8c0-410d-a4cb-e5d0c2ce634b/heat-engine/0.log" Feb 25 10:21:19 crc kubenswrapper[4978]: I0225 10:21:19.373335 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d9787759d-6l97q_f4ea666a-1e63-421a-a9fe-c085feabb6ff/horizon/0.log" Feb 25 10:21:19 crc kubenswrapper[4978]: I0225 10:21:19.474395 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5b459dcfff-rkn7k_e36d5e82-a1a3-431e-ad1b-a585ca8ae0bf/heat-cfnapi/0.log" Feb 25 10:21:19 crc kubenswrapper[4978]: I0225 10:21:19.623504 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-7fc7d9db6-jkz4h_128089fe-663c-42b0-80a2-b32e5793d555/heat-api/0.log" Feb 25 10:21:19 crc kubenswrapper[4978]: I0225 10:21:19.837396 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-7fcwt_b722f89d-c453-4235-8d4c-8d9990f0551b/install-certs-openstack-openstack-cell1/0.log" Feb 25 10:21:20 crc kubenswrapper[4978]: I0225 10:21:20.030536 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-networker-q99kk_4555309b-6018-4eb3-ae3b-0f7ab8d79e10/install-certs-openstack-openstack-networker/0.log" Feb 25 10:21:20 crc kubenswrapper[4978]: I0225 10:21:20.222703 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-97vfl_0aaa93ca-3269-4ade-bb1d-ac862e9738c5/install-os-openstack-openstack-cell1/0.log" Feb 25 10:21:20 crc kubenswrapper[4978]: I0225 10:21:20.547584 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d9787759d-6l97q_f4ea666a-1e63-421a-a9fe-c085feabb6ff/horizon-log/0.log" Feb 25 10:21:20 crc kubenswrapper[4978]: I0225 10:21:20.562717 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-networker-fwglm_50114eec-786f-467f-b235-24829d7c3e63/install-os-openstack-openstack-networker/0.log" Feb 25 10:21:20 crc kubenswrapper[4978]: I0225 10:21:20.891636 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29533501-f2btx_5d5bf4f2-3ff6-405c-85a5-4423fbd1d88e/keystone-cron/0.log" Feb 25 10:21:21 crc kubenswrapper[4978]: I0225 10:21:21.232466 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29533561-phzdw_434f88a0-9d17-474f-a877-35569aba520b/keystone-cron/0.log" Feb 25 10:21:21 crc kubenswrapper[4978]: I0225 10:21:21.285213 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_902de2fe-e738-4665-a008-a9a9ed92c7c1/kube-state-metrics/0.log" Feb 25 10:21:21 crc kubenswrapper[4978]: I0225 10:21:21.598162 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-2cksr_41f59382-722c-42ab-89f1-0e738135edcd/libvirt-openstack-openstack-cell1/0.log" Feb 25 10:21:21 crc kubenswrapper[4978]: I0225 10:21:21.807312 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6997548745-tb5n6_605215fb-7a4c-49fb-a715-5c7251cad82e/dnsmasq-dns/0.log" Feb 25 10:21:21 crc kubenswrapper[4978]: I0225 10:21:21.894785 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-8d576976-9bm5w_25443aff-625b-4e6a-b03c-1697c5e2199e/keystone-api/0.log" Feb 25 10:21:22 crc kubenswrapper[4978]: I0225 10:21:22.341472 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7cc578599c-fmh6q_882b5c88-3ecd-4baa-91c6-682450eb37fe/neutron-httpd/0.log" Feb 25 10:21:22 crc kubenswrapper[4978]: I0225 10:21:22.432989 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-qnnvm_a5f51a06-f15a-4967-a8e6-25c422c09c32/neutron-dhcp-openstack-openstack-cell1/0.log" Feb 25 10:21:22 crc kubenswrapper[4978]: I0225 10:21:22.621736 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7cc578599c-fmh6q_882b5c88-3ecd-4baa-91c6-682450eb37fe/neutron-api/0.log" Feb 25 10:21:22 crc kubenswrapper[4978]: I0225 10:21:22.655451 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-xs7rz_0d691a60-c2ae-48a6-8d7d-b8ea03d5178a/neutron-metadata-openstack-openstack-cell1/0.log" Feb 25 10:21:22 crc kubenswrapper[4978]: I0225 10:21:22.921564 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-networker-wvtwt_29abc7cf-e9b4-47c2-9a9a-5462a7a3f43d/neutron-metadata-openstack-openstack-networker/0.log" Feb 25 10:21:22 crc kubenswrapper[4978]: I0225 10:21:22.964074 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-rd5lr_01398882-0eec-4662-b7fd-5d7c97308a4d/neutron-sriov-openstack-openstack-cell1/0.log" Feb 25 10:21:23 crc kubenswrapper[4978]: I0225 10:21:23.712165 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_621cc593-53a8-4319-91c7-7ec49790758a/nova-cell0-conductor-conductor/0.log" Feb 25 10:21:23 crc kubenswrapper[4978]: I0225 10:21:23.803341 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d2ad5e9e-1c73-4316-940f-5b58e5c39076/nova-api-log/0.log" Feb 25 10:21:24 crc kubenswrapper[4978]: I0225 10:21:24.006768 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_d2ad5e9e-1c73-4316-940f-5b58e5c39076/nova-api-api/0.log" Feb 25 10:21:24 crc kubenswrapper[4978]: I0225 10:21:24.303404 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_f2fb3fa7-1a7e-41e4-a11f-9ae6bd9380c4/nova-cell1-conductor-conductor/0.log" Feb 25 10:21:24 crc kubenswrapper[4978]: I0225 10:21:24.438049 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ddac0d7e-f9ca-4094-854a-e008472a187a/nova-cell1-novncproxy-novncproxy/0.log" Feb 25 10:21:24 crc kubenswrapper[4978]: I0225 10:21:24.788990 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-qp6rn_d39c9d7a-8206-4313-9915-aa3825c28444/nova-cell1-openstack-openstack-cell1/0.log" Feb 25 10:21:24 crc kubenswrapper[4978]: I0225 10:21:24.856174 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell2z9ht_03d45c16-5d3e-4342-b59f-453eeba6b33e/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Feb 25 10:21:25 crc kubenswrapper[4978]: I0225 10:21:25.260914 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d842555c-bc9c-492f-8a1a-db131c37f5ae/nova-metadata-log/0.log" Feb 25 10:21:25 crc kubenswrapper[4978]: I0225 10:21:25.457972 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_fd7e65e5-0901-4fa6-95ba-d4649dd9a62e/nova-scheduler-scheduler/0.log" Feb 25 10:21:25 crc kubenswrapper[4978]: I0225 10:21:25.705159 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c7be4547-b312-4601-93bf-9a3c62a13829/mysql-bootstrap/0.log" Feb 25 10:21:26 crc kubenswrapper[4978]: I0225 10:21:26.045500 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c7be4547-b312-4601-93bf-9a3c62a13829/mysql-bootstrap/0.log" Feb 25 10:21:26 crc kubenswrapper[4978]: I0225 10:21:26.159769 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_c7be4547-b312-4601-93bf-9a3c62a13829/galera/0.log" Feb 25 10:21:26 crc kubenswrapper[4978]: I0225 10:21:26.330256 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:21:26 crc kubenswrapper[4978]: I0225 10:21:26.447148 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_11a899ad-0164-4f8b-ae66-797557d34867/mysql-bootstrap/0.log" Feb 25 10:21:26 crc kubenswrapper[4978]: I0225 10:21:26.749961 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_11a899ad-0164-4f8b-ae66-797557d34867/mysql-bootstrap/0.log" Feb 25 10:21:26 crc kubenswrapper[4978]: I0225 10:21:26.915245 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_11a899ad-0164-4f8b-ae66-797557d34867/galera/0.log" Feb 25 10:21:26 crc kubenswrapper[4978]: I0225 10:21:26.996522 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_d842555c-bc9c-492f-8a1a-db131c37f5ae/nova-metadata-metadata/0.log" Feb 25 10:21:27 crc kubenswrapper[4978]: I0225 10:21:27.256755 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1a5e0a02-7fb2-4745-b49d-fd5b358c3122/openstackclient/0.log" Feb 25 10:21:27 crc kubenswrapper[4978]: I0225 10:21:27.294799 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c197b2cc-cc65-4113-ab1a-517539270d88/openstack-network-exporter/0.log" Feb 25 10:21:27 crc kubenswrapper[4978]: I0225 10:21:27.462814 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"95899c3bd9197bd5e205d8ad51463abe0d611cd768f6460f30bac2b6f6450c64"} Feb 25 10:21:27 crc kubenswrapper[4978]: I0225 10:21:27.669738 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c197b2cc-cc65-4113-ab1a-517539270d88/ovn-northd/0.log" Feb 25 10:21:27 crc kubenswrapper[4978]: I0225 10:21:27.754432 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-rq28r_7fdbbb5e-0829-45c0-8995-42267e75fa07/ovn-openstack-openstack-cell1/0.log" Feb 25 10:21:28 crc kubenswrapper[4978]: I0225 10:21:28.135345 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-networker-fxwqp_35bf1547-15c8-4591-945f-f185c31b9660/ovn-openstack-openstack-networker/0.log" Feb 25 10:21:28 crc kubenswrapper[4978]: I0225 10:21:28.285013 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a7f671c5-0151-4f3d-b919-e5c9ac280038/openstack-network-exporter/0.log" Feb 25 10:21:28 crc kubenswrapper[4978]: I0225 10:21:28.293328 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a7f671c5-0151-4f3d-b919-e5c9ac280038/ovsdbserver-nb/0.log" Feb 25 10:21:28 crc kubenswrapper[4978]: I0225 10:21:28.660740 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_aa16e3af-ea25-4a1c-83d1-3300ab386437/ovsdbserver-nb/0.log" Feb 25 10:21:28 crc kubenswrapper[4978]: I0225 10:21:28.846394 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_aa16e3af-ea25-4a1c-83d1-3300ab386437/openstack-network-exporter/0.log" Feb 25 10:21:28 crc kubenswrapper[4978]: I0225 10:21:28.961866 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_a9119e44-2b2c-479b-a465-405ec797b89d/openstack-network-exporter/0.log" Feb 25 10:21:29 crc kubenswrapper[4978]: I0225 10:21:29.149766 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_a9119e44-2b2c-479b-a465-405ec797b89d/ovsdbserver-nb/0.log" Feb 25 10:21:29 crc kubenswrapper[4978]: I0225 10:21:29.419166 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_99c42bcd-7266-4cf3-a7c4-b8db18645a4f/openstack-network-exporter/0.log" Feb 25 10:21:29 crc kubenswrapper[4978]: I0225 10:21:29.602254 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_99c42bcd-7266-4cf3-a7c4-b8db18645a4f/ovsdbserver-sb/0.log" Feb 25 10:21:29 crc kubenswrapper[4978]: I0225 10:21:29.782994 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_e21d9172-38ac-4816-b77d-db144cdfd65b/openstack-network-exporter/0.log" Feb 25 10:21:29 crc kubenswrapper[4978]: I0225 10:21:29.872528 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_e21d9172-38ac-4816-b77d-db144cdfd65b/ovsdbserver-sb/0.log" Feb 25 10:21:30 crc kubenswrapper[4978]: I0225 10:21:30.021146 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_8621c881-a82f-4922-9e91-43422305fcea/openstack-network-exporter/0.log" Feb 25 10:21:30 crc kubenswrapper[4978]: I0225 10:21:30.180063 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_8621c881-a82f-4922-9e91-43422305fcea/ovsdbserver-sb/0.log" Feb 25 10:21:30 crc kubenswrapper[4978]: I0225 10:21:30.602532 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-894b87bcd-94p22_d17346e6-d110-4b4e-bdc2-afd1eeea0b6e/placement-api/0.log" Feb 25 10:21:30 crc kubenswrapper[4978]: I0225 10:21:30.873437 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cpndph_a37f3a9e-cb0d-47f9-8cec-de20cf40a874/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Feb 25 10:21:30 crc kubenswrapper[4978]: I0225 10:21:30.924998 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-894b87bcd-94p22_d17346e6-d110-4b4e-bdc2-afd1eeea0b6e/placement-log/0.log" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.315714 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-n8rh5l_5deab02b-db5d-47f5-921b-a5b35412877a/pre-adoption-validation-openstack-pre-adoption-openstack-networ/0.log" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.532430 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2qbk2"] Feb 25 10:21:31 crc kubenswrapper[4978]: E0225 10:21:31.549619 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d4aee5d-a45a-4672-ae14-e1af0287f17d" containerName="oc" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.549650 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d4aee5d-a45a-4672-ae14-e1af0287f17d" containerName="oc" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.550083 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d4aee5d-a45a-4672-ae14-e1af0287f17d" containerName="oc" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.552124 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.561568 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qbk2"] Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.625441 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e3249bd-3d1b-41a2-9132-976e11161c91/init-config-reloader/0.log" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.657275 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-catalog-content\") pod \"redhat-marketplace-2qbk2\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.657647 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-utilities\") pod \"redhat-marketplace-2qbk2\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.657768 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvznr\" (UniqueName: \"kubernetes.io/projected/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-kube-api-access-fvznr\") pod \"redhat-marketplace-2qbk2\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.715762 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e3249bd-3d1b-41a2-9132-976e11161c91/init-config-reloader/0.log" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.762156 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-catalog-content\") pod \"redhat-marketplace-2qbk2\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.762217 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-utilities\") pod \"redhat-marketplace-2qbk2\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.762238 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvznr\" (UniqueName: \"kubernetes.io/projected/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-kube-api-access-fvznr\") pod \"redhat-marketplace-2qbk2\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.762991 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-utilities\") pod \"redhat-marketplace-2qbk2\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.763810 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-catalog-content\") pod \"redhat-marketplace-2qbk2\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.809232 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvznr\" (UniqueName: \"kubernetes.io/projected/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-kube-api-access-fvznr\") pod \"redhat-marketplace-2qbk2\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.897004 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:31 crc kubenswrapper[4978]: I0225 10:21:31.942563 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e3249bd-3d1b-41a2-9132-976e11161c91/config-reloader/0.log" Feb 25 10:21:32 crc kubenswrapper[4978]: I0225 10:21:32.478336 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e3249bd-3d1b-41a2-9132-976e11161c91/prometheus/0.log" Feb 25 10:21:32 crc kubenswrapper[4978]: I0225 10:21:32.531466 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e3249bd-3d1b-41a2-9132-976e11161c91/thanos-sidecar/0.log" Feb 25 10:21:32 crc kubenswrapper[4978]: I0225 10:21:32.675293 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_787e6909-959b-475a-a300-5ff751242c20/setup-container/0.log" Feb 25 10:21:32 crc kubenswrapper[4978]: I0225 10:21:32.955232 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qbk2"] Feb 25 10:21:33 crc kubenswrapper[4978]: I0225 10:21:33.285030 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_785d8daf-cc33-4011-9767-fa6049438fa8/setup-container/0.log" Feb 25 10:21:33 crc kubenswrapper[4978]: I0225 10:21:33.286469 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_787e6909-959b-475a-a300-5ff751242c20/rabbitmq/0.log" Feb 25 10:21:33 crc kubenswrapper[4978]: I0225 10:21:33.299955 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_787e6909-959b-475a-a300-5ff751242c20/setup-container/0.log" Feb 25 10:21:33 crc kubenswrapper[4978]: I0225 10:21:33.556898 4978 generic.go:334] "Generic (PLEG): container finished" podID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerID="8f37d36f5f081db4340bd589ce011cdf5f52e5988245e19309eb8ee813193b09" exitCode=0 Feb 25 10:21:33 crc kubenswrapper[4978]: I0225 10:21:33.556955 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qbk2" event={"ID":"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf","Type":"ContainerDied","Data":"8f37d36f5f081db4340bd589ce011cdf5f52e5988245e19309eb8ee813193b09"} Feb 25 10:21:33 crc kubenswrapper[4978]: I0225 10:21:33.556988 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qbk2" event={"ID":"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf","Type":"ContainerStarted","Data":"667112c1282156fd437f2371814e5f3d8e1c27198a3ef05da8a4c7b64937ad52"} Feb 25 10:21:33 crc kubenswrapper[4978]: I0225 10:21:33.560101 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 10:21:33 crc kubenswrapper[4978]: I0225 10:21:33.681499 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_785d8daf-cc33-4011-9767-fa6049438fa8/setup-container/0.log" Feb 25 10:21:33 crc kubenswrapper[4978]: I0225 10:21:33.922005 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-mbr4t_356ba162-9c45-4a68-b891-455d1e99065f/reboot-os-openstack-openstack-cell1/0.log" Feb 25 10:21:33 crc kubenswrapper[4978]: I0225 10:21:33.966922 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_785d8daf-cc33-4011-9767-fa6049438fa8/rabbitmq/0.log" Feb 25 10:21:34 crc kubenswrapper[4978]: I0225 10:21:34.366665 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-networker-ddfpk_87856568-5e2c-40f3-99cd-a159a8ae96e9/reboot-os-openstack-openstack-networker/0.log" Feb 25 10:21:34 crc kubenswrapper[4978]: I0225 10:21:34.825170 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-networker-2mwbf_ce547a55-9fd6-4692-b7e8-5e6303cd4293/run-os-openstack-openstack-networker/0.log" Feb 25 10:21:34 crc kubenswrapper[4978]: I0225 10:21:34.897249 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-tkbrw_1ca465d1-d912-48fc-984f-62f645da23cd/run-os-openstack-openstack-cell1/0.log" Feb 25 10:21:35 crc kubenswrapper[4978]: I0225 10:21:35.252743 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-9fl9m_06f07aaf-4685-457d-b8e5-da1b4f2daa9b/ssh-known-hosts-openstack/0.log" Feb 25 10:21:35 crc kubenswrapper[4978]: I0225 10:21:35.633345 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qbk2" event={"ID":"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf","Type":"ContainerStarted","Data":"96059f91fe99b6227e5c41fa9b42258f4cdf418604f9a3d4b73d10934d1d6a7d"} Feb 25 10:21:36 crc kubenswrapper[4978]: I0225 10:21:36.193860 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-tlbw2_751b1bbb-a2d3-4d1b-bc2e-60bb99fbd859/swift-ring-rebalance/0.log" Feb 25 10:21:36 crc kubenswrapper[4978]: I0225 10:21:36.344420 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-66cf8f6986-prvmn_a0d7bad3-d58a-4225-8390-e2785235d412/proxy-httpd/0.log" Feb 25 10:21:36 crc kubenswrapper[4978]: I0225 10:21:36.422545 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-66cf8f6986-prvmn_a0d7bad3-d58a-4225-8390-e2785235d412/proxy-server/0.log" Feb 25 10:21:36 crc kubenswrapper[4978]: I0225 10:21:36.544387 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/account-auditor/0.log" Feb 25 10:21:36 crc kubenswrapper[4978]: I0225 10:21:36.666262 4978 generic.go:334] "Generic (PLEG): container finished" podID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerID="96059f91fe99b6227e5c41fa9b42258f4cdf418604f9a3d4b73d10934d1d6a7d" exitCode=0 Feb 25 10:21:36 crc kubenswrapper[4978]: I0225 10:21:36.666327 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qbk2" event={"ID":"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf","Type":"ContainerDied","Data":"96059f91fe99b6227e5c41fa9b42258f4cdf418604f9a3d4b73d10934d1d6a7d"} Feb 25 10:21:36 crc kubenswrapper[4978]: I0225 10:21:36.880932 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/account-reaper/0.log" Feb 25 10:21:37 crc kubenswrapper[4978]: I0225 10:21:37.105845 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/account-replicator/0.log" Feb 25 10:21:37 crc kubenswrapper[4978]: I0225 10:21:37.512460 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/account-server/0.log" Feb 25 10:21:37 crc kubenswrapper[4978]: I0225 10:21:37.514877 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/container-replicator/0.log" Feb 25 10:21:37 crc kubenswrapper[4978]: I0225 10:21:37.526230 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/container-auditor/0.log" Feb 25 10:21:37 crc kubenswrapper[4978]: I0225 10:21:37.686554 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qbk2" event={"ID":"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf","Type":"ContainerStarted","Data":"44452ce3b29c3b55bc71bdd6403319f1df5837d75c748e2435348c95992ec93c"} Feb 25 10:21:37 crc kubenswrapper[4978]: I0225 10:21:37.702730 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2qbk2" podStartSLOduration=3.003501758 podStartE2EDuration="6.70271337s" podCreationTimestamp="2026-02-25 10:21:31 +0000 UTC" firstStartedPulling="2026-02-25 10:21:33.559887236 +0000 UTC m=+12986.999143695" lastFinishedPulling="2026-02-25 10:21:37.259098848 +0000 UTC m=+12990.698355307" observedRunningTime="2026-02-25 10:21:37.699753988 +0000 UTC m=+12991.139010457" watchObservedRunningTime="2026-02-25 10:21:37.70271337 +0000 UTC m=+12991.141969829" Feb 25 10:21:37 crc kubenswrapper[4978]: I0225 10:21:37.984043 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/object-auditor/0.log" Feb 25 10:21:38 crc kubenswrapper[4978]: I0225 10:21:38.331664 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/object-expirer/0.log" Feb 25 10:21:38 crc kubenswrapper[4978]: I0225 10:21:38.512733 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/container-updater/0.log" Feb 25 10:21:38 crc kubenswrapper[4978]: I0225 10:21:38.854553 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/object-replicator/0.log" Feb 25 10:21:39 crc kubenswrapper[4978]: I0225 10:21:39.074715 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/container-server/0.log" Feb 25 10:21:39 crc kubenswrapper[4978]: I0225 10:21:39.139356 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/object-updater/0.log" Feb 25 10:21:39 crc kubenswrapper[4978]: I0225 10:21:39.568515 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/object-server/0.log" Feb 25 10:21:39 crc kubenswrapper[4978]: I0225 10:21:39.711855 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/swift-recon-cron/0.log" Feb 25 10:21:39 crc kubenswrapper[4978]: I0225 10:21:39.970627 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/account-auditor/0.log" Feb 25 10:21:40 crc kubenswrapper[4978]: I0225 10:21:40.024508 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/account-reaper/0.log" Feb 25 10:21:40 crc kubenswrapper[4978]: I0225 10:21:40.033725 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f6550fa0-18f1-4c7a-a8e7-ab3960f5828b/rsync/0.log" Feb 25 10:21:40 crc kubenswrapper[4978]: I0225 10:21:40.435791 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/account-replicator/0.log" Feb 25 10:21:40 crc kubenswrapper[4978]: I0225 10:21:40.638682 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/container-auditor/0.log" Feb 25 10:21:40 crc kubenswrapper[4978]: I0225 10:21:40.702157 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/container-replicator/0.log" Feb 25 10:21:40 crc kubenswrapper[4978]: I0225 10:21:40.829099 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/account-server/0.log" Feb 25 10:21:41 crc kubenswrapper[4978]: I0225 10:21:41.245450 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/object-expirer/0.log" Feb 25 10:21:41 crc kubenswrapper[4978]: I0225 10:21:41.308764 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/container-updater/0.log" Feb 25 10:21:41 crc kubenswrapper[4978]: I0225 10:21:41.313477 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/object-auditor/0.log" Feb 25 10:21:41 crc kubenswrapper[4978]: I0225 10:21:41.450736 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/container-server/0.log" Feb 25 10:21:41 crc kubenswrapper[4978]: I0225 10:21:41.749607 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/object-replicator/0.log" Feb 25 10:21:41 crc kubenswrapper[4978]: I0225 10:21:41.790582 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/object-updater/0.log" Feb 25 10:21:41 crc kubenswrapper[4978]: I0225 10:21:41.898665 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:41 crc kubenswrapper[4978]: I0225 10:21:41.898715 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:41 crc kubenswrapper[4978]: I0225 10:21:41.954772 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:42 crc kubenswrapper[4978]: I0225 10:21:42.137775 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/swift-recon-cron/0.log" Feb 25 10:21:42 crc kubenswrapper[4978]: I0225 10:21:42.373982 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/object-server/0.log" Feb 25 10:21:42 crc kubenswrapper[4978]: I0225 10:21:42.576068 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/account-auditor/0.log" Feb 25 10:21:42 crc kubenswrapper[4978]: I0225 10:21:42.769280 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/account-reaper/0.log" Feb 25 10:21:42 crc kubenswrapper[4978]: I0225 10:21:42.770579 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-1_c9958cdd-7b41-48e4-b4fa-89b23d6e006c/rsync/0.log" Feb 25 10:21:42 crc kubenswrapper[4978]: I0225 10:21:42.870623 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:42 crc kubenswrapper[4978]: I0225 10:21:42.945875 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qbk2"] Feb 25 10:21:43 crc kubenswrapper[4978]: I0225 10:21:43.001138 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/account-replicator/0.log" Feb 25 10:21:43 crc kubenswrapper[4978]: I0225 10:21:43.187084 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/container-auditor/0.log" Feb 25 10:21:43 crc kubenswrapper[4978]: I0225 10:21:43.196093 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/account-server/0.log" Feb 25 10:21:43 crc kubenswrapper[4978]: I0225 10:21:43.411014 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/container-replicator/0.log" Feb 25 10:21:43 crc kubenswrapper[4978]: I0225 10:21:43.578608 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/container-updater/0.log" Feb 25 10:21:43 crc kubenswrapper[4978]: I0225 10:21:43.660677 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/object-auditor/0.log" Feb 25 10:21:43 crc kubenswrapper[4978]: I0225 10:21:43.864923 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/object-expirer/0.log" Feb 25 10:21:43 crc kubenswrapper[4978]: I0225 10:21:43.957096 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/container-server/0.log" Feb 25 10:21:44 crc kubenswrapper[4978]: I0225 10:21:44.042246 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/object-replicator/0.log" Feb 25 10:21:44 crc kubenswrapper[4978]: I0225 10:21:44.295746 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/object-updater/0.log" Feb 25 10:21:44 crc kubenswrapper[4978]: I0225 10:21:44.537353 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/swift-recon-cron/0.log" Feb 25 10:21:44 crc kubenswrapper[4978]: I0225 10:21:44.805721 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2qbk2" podUID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerName="registry-server" containerID="cri-o://44452ce3b29c3b55bc71bdd6403319f1df5837d75c748e2435348c95992ec93c" gracePeriod=2 Feb 25 10:21:45 crc kubenswrapper[4978]: I0225 10:21:45.127033 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-z7lrc_bc76d2fb-7bba-48f7-9658-e55f0c225807/telemetry-openstack-openstack-cell1/0.log" Feb 25 10:21:45 crc kubenswrapper[4978]: I0225 10:21:45.172325 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/object-server/0.log" Feb 25 10:21:45 crc kubenswrapper[4978]: I0225 10:21:45.430265 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-2_2e3f7363-8d31-4036-96bc-25f2d2c1b4bb/rsync/0.log" Feb 25 10:21:45 crc kubenswrapper[4978]: I0225 10:21:45.854471 4978 generic.go:334] "Generic (PLEG): container finished" podID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerID="44452ce3b29c3b55bc71bdd6403319f1df5837d75c748e2435348c95992ec93c" exitCode=0 Feb 25 10:21:45 crc kubenswrapper[4978]: I0225 10:21:45.854512 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qbk2" event={"ID":"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf","Type":"ContainerDied","Data":"44452ce3b29c3b55bc71bdd6403319f1df5837d75c748e2435348c95992ec93c"} Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.066295 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_8f150fab-adf6-49b7-8c94-447452f23850/test-operator-logs-container/0.log" Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.072988 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_ec6773fc-9a0a-4d54-a950-85b5be1fe6cd/tempest-tests-tempest-tests-runner/0.log" Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.516035 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-2vhwm_6153aa3b-10fb-4bba-8103-71d3f8347d0d/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.645968 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.848381 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-utilities\") pod \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.848892 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-catalog-content\") pod \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.848951 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvznr\" (UniqueName: \"kubernetes.io/projected/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-kube-api-access-fvznr\") pod \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\" (UID: \"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf\") " Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.860394 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-utilities" (OuterVolumeSpecName: "utilities") pod "ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" (UID: "ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.883635 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-kube-api-access-fvznr" (OuterVolumeSpecName: "kube-api-access-fvznr") pod "ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" (UID: "ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf"). InnerVolumeSpecName "kube-api-access-fvznr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.890901 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2qbk2" event={"ID":"ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf","Type":"ContainerDied","Data":"667112c1282156fd437f2371814e5f3d8e1c27198a3ef05da8a4c7b64937ad52"} Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.890979 4978 scope.go:117] "RemoveContainer" containerID="44452ce3b29c3b55bc71bdd6403319f1df5837d75c748e2435348c95992ec93c" Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.891200 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2qbk2" Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.951232 4978 scope.go:117] "RemoveContainer" containerID="96059f91fe99b6227e5c41fa9b42258f4cdf418604f9a3d4b73d10934d1d6a7d" Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.956258 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:21:46 crc kubenswrapper[4978]: I0225 10:21:46.956314 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvznr\" (UniqueName: \"kubernetes.io/projected/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-kube-api-access-fvznr\") on node \"crc\" DevicePath \"\"" Feb 25 10:21:47 crc kubenswrapper[4978]: I0225 10:21:46.989689 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" (UID: "ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:21:47 crc kubenswrapper[4978]: I0225 10:21:47.027489 4978 scope.go:117] "RemoveContainer" containerID="8f37d36f5f081db4340bd589ce011cdf5f52e5988245e19309eb8ee813193b09" Feb 25 10:21:47 crc kubenswrapper[4978]: I0225 10:21:47.032102 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-networker-qxng8_6abcdd51-69ae-42d5-9708-7088d9911458/tripleo-cleanup-tripleo-cleanup-openstack-networker/0.log" Feb 25 10:21:47 crc kubenswrapper[4978]: I0225 10:21:47.076242 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:21:47 crc kubenswrapper[4978]: I0225 10:21:47.255490 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qbk2"] Feb 25 10:21:47 crc kubenswrapper[4978]: I0225 10:21:47.277515 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2qbk2"] Feb 25 10:21:47 crc kubenswrapper[4978]: I0225 10:21:47.302539 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-czcmg_ad13325e-8670-43ca-9461-aec47779e55e/validate-network-openstack-openstack-cell1/0.log" Feb 25 10:21:47 crc kubenswrapper[4978]: I0225 10:21:47.354848 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" path="/var/lib/kubelet/pods/ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf/volumes" Feb 25 10:21:47 crc kubenswrapper[4978]: I0225 10:21:47.623093 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-networker-jl46h_4ccacd71-8214-46a5-b6a4-d31ec3b9fa71/validate-network-openstack-openstack-networker/0.log" Feb 25 10:21:48 crc kubenswrapper[4978]: I0225 10:21:48.062225 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2fd170eb-5a4e-4461-9619-58f110e6e8f1/memcached/0.log" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.141791 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533582-2bdwl"] Feb 25 10:22:00 crc kubenswrapper[4978]: E0225 10:22:00.142614 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerName="registry-server" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.142628 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerName="registry-server" Feb 25 10:22:00 crc kubenswrapper[4978]: E0225 10:22:00.142660 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerName="extract-utilities" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.142666 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerName="extract-utilities" Feb 25 10:22:00 crc kubenswrapper[4978]: E0225 10:22:00.142695 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerName="extract-content" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.142701 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerName="extract-content" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.142892 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea56538d-f20e-4c3c-8dfe-89bd81d5bcaf" containerName="registry-server" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.143596 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533582-2bdwl" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.146906 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.147051 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.153817 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533582-2bdwl"] Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.183202 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.250997 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbdwr\" (UniqueName: \"kubernetes.io/projected/b28b16ff-594c-4da7-aa74-e977658b02c0-kube-api-access-pbdwr\") pod \"auto-csr-approver-29533582-2bdwl\" (UID: \"b28b16ff-594c-4da7-aa74-e977658b02c0\") " pod="openshift-infra/auto-csr-approver-29533582-2bdwl" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.353181 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbdwr\" (UniqueName: \"kubernetes.io/projected/b28b16ff-594c-4da7-aa74-e977658b02c0-kube-api-access-pbdwr\") pod \"auto-csr-approver-29533582-2bdwl\" (UID: \"b28b16ff-594c-4da7-aa74-e977658b02c0\") " pod="openshift-infra/auto-csr-approver-29533582-2bdwl" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.371968 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbdwr\" (UniqueName: \"kubernetes.io/projected/b28b16ff-594c-4da7-aa74-e977658b02c0-kube-api-access-pbdwr\") pod \"auto-csr-approver-29533582-2bdwl\" (UID: \"b28b16ff-594c-4da7-aa74-e977658b02c0\") " pod="openshift-infra/auto-csr-approver-29533582-2bdwl" Feb 25 10:22:00 crc kubenswrapper[4978]: I0225 10:22:00.502341 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533582-2bdwl" Feb 25 10:22:01 crc kubenswrapper[4978]: I0225 10:22:01.310356 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533582-2bdwl"] Feb 25 10:22:02 crc kubenswrapper[4978]: I0225 10:22:02.036455 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533582-2bdwl" event={"ID":"b28b16ff-594c-4da7-aa74-e977658b02c0","Type":"ContainerStarted","Data":"d18845344b437d7d1b87cdf4e1d7a13bff56d19a17dd5f64ca839ed555765171"} Feb 25 10:22:06 crc kubenswrapper[4978]: I0225 10:22:06.083119 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533582-2bdwl" event={"ID":"b28b16ff-594c-4da7-aa74-e977658b02c0","Type":"ContainerStarted","Data":"b04f7b7becce137d277327e23bb6eda7b020ba22a1873c08a5648747300169b8"} Feb 25 10:22:06 crc kubenswrapper[4978]: I0225 10:22:06.103022 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533582-2bdwl" podStartSLOduration=2.465117037 podStartE2EDuration="6.103004518s" podCreationTimestamp="2026-02-25 10:22:00 +0000 UTC" firstStartedPulling="2026-02-25 10:22:01.332700684 +0000 UTC m=+13014.771957143" lastFinishedPulling="2026-02-25 10:22:04.970588165 +0000 UTC m=+13018.409844624" observedRunningTime="2026-02-25 10:22:06.092994768 +0000 UTC m=+13019.532251227" watchObservedRunningTime="2026-02-25 10:22:06.103004518 +0000 UTC m=+13019.542260977" Feb 25 10:22:07 crc kubenswrapper[4978]: I0225 10:22:07.093894 4978 generic.go:334] "Generic (PLEG): container finished" podID="b28b16ff-594c-4da7-aa74-e977658b02c0" containerID="b04f7b7becce137d277327e23bb6eda7b020ba22a1873c08a5648747300169b8" exitCode=0 Feb 25 10:22:07 crc kubenswrapper[4978]: I0225 10:22:07.093965 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533582-2bdwl" event={"ID":"b28b16ff-594c-4da7-aa74-e977658b02c0","Type":"ContainerDied","Data":"b04f7b7becce137d277327e23bb6eda7b020ba22a1873c08a5648747300169b8"} Feb 25 10:22:09 crc kubenswrapper[4978]: I0225 10:22:09.464425 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533582-2bdwl" Feb 25 10:22:09 crc kubenswrapper[4978]: I0225 10:22:09.553646 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbdwr\" (UniqueName: \"kubernetes.io/projected/b28b16ff-594c-4da7-aa74-e977658b02c0-kube-api-access-pbdwr\") pod \"b28b16ff-594c-4da7-aa74-e977658b02c0\" (UID: \"b28b16ff-594c-4da7-aa74-e977658b02c0\") " Feb 25 10:22:09 crc kubenswrapper[4978]: I0225 10:22:09.560104 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b28b16ff-594c-4da7-aa74-e977658b02c0-kube-api-access-pbdwr" (OuterVolumeSpecName: "kube-api-access-pbdwr") pod "b28b16ff-594c-4da7-aa74-e977658b02c0" (UID: "b28b16ff-594c-4da7-aa74-e977658b02c0"). InnerVolumeSpecName "kube-api-access-pbdwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:22:09 crc kubenswrapper[4978]: I0225 10:22:09.659826 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbdwr\" (UniqueName: \"kubernetes.io/projected/b28b16ff-594c-4da7-aa74-e977658b02c0-kube-api-access-pbdwr\") on node \"crc\" DevicePath \"\"" Feb 25 10:22:10 crc kubenswrapper[4978]: I0225 10:22:10.131364 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533582-2bdwl" event={"ID":"b28b16ff-594c-4da7-aa74-e977658b02c0","Type":"ContainerDied","Data":"d18845344b437d7d1b87cdf4e1d7a13bff56d19a17dd5f64ca839ed555765171"} Feb 25 10:22:10 crc kubenswrapper[4978]: I0225 10:22:10.131725 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d18845344b437d7d1b87cdf4e1d7a13bff56d19a17dd5f64ca839ed555765171" Feb 25 10:22:10 crc kubenswrapper[4978]: I0225 10:22:10.131429 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533582-2bdwl" Feb 25 10:22:10 crc kubenswrapper[4978]: I0225 10:22:10.530348 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533576-hr7r8"] Feb 25 10:22:10 crc kubenswrapper[4978]: I0225 10:22:10.542269 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533576-hr7r8"] Feb 25 10:22:11 crc kubenswrapper[4978]: I0225 10:22:11.338451 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd9f079d-35bf-45c0-850f-d8270674c947" path="/var/lib/kubelet/pods/cd9f079d-35bf-45c0-850f-d8270674c947/volumes" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.324335 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vfnpj"] Feb 25 10:22:17 crc kubenswrapper[4978]: E0225 10:22:17.325302 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b28b16ff-594c-4da7-aa74-e977658b02c0" containerName="oc" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.325316 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b28b16ff-594c-4da7-aa74-e977658b02c0" containerName="oc" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.325568 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b28b16ff-594c-4da7-aa74-e977658b02c0" containerName="oc" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.327270 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.341791 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vfnpj"] Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.464894 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50fabec8-ad45-4aac-a5f0-56496603e1e8-catalog-content\") pod \"community-operators-vfnpj\" (UID: \"50fabec8-ad45-4aac-a5f0-56496603e1e8\") " pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.465095 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72q68\" (UniqueName: \"kubernetes.io/projected/50fabec8-ad45-4aac-a5f0-56496603e1e8-kube-api-access-72q68\") pod \"community-operators-vfnpj\" (UID: \"50fabec8-ad45-4aac-a5f0-56496603e1e8\") " pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.465271 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50fabec8-ad45-4aac-a5f0-56496603e1e8-utilities\") pod \"community-operators-vfnpj\" (UID: \"50fabec8-ad45-4aac-a5f0-56496603e1e8\") " pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.567513 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72q68\" (UniqueName: \"kubernetes.io/projected/50fabec8-ad45-4aac-a5f0-56496603e1e8-kube-api-access-72q68\") pod \"community-operators-vfnpj\" (UID: \"50fabec8-ad45-4aac-a5f0-56496603e1e8\") " pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.567622 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50fabec8-ad45-4aac-a5f0-56496603e1e8-utilities\") pod \"community-operators-vfnpj\" (UID: \"50fabec8-ad45-4aac-a5f0-56496603e1e8\") " pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.567705 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50fabec8-ad45-4aac-a5f0-56496603e1e8-catalog-content\") pod \"community-operators-vfnpj\" (UID: \"50fabec8-ad45-4aac-a5f0-56496603e1e8\") " pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.568183 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50fabec8-ad45-4aac-a5f0-56496603e1e8-catalog-content\") pod \"community-operators-vfnpj\" (UID: \"50fabec8-ad45-4aac-a5f0-56496603e1e8\") " pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.568203 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50fabec8-ad45-4aac-a5f0-56496603e1e8-utilities\") pod \"community-operators-vfnpj\" (UID: \"50fabec8-ad45-4aac-a5f0-56496603e1e8\") " pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.586481 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72q68\" (UniqueName: \"kubernetes.io/projected/50fabec8-ad45-4aac-a5f0-56496603e1e8-kube-api-access-72q68\") pod \"community-operators-vfnpj\" (UID: \"50fabec8-ad45-4aac-a5f0-56496603e1e8\") " pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:17 crc kubenswrapper[4978]: I0225 10:22:17.666159 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:18 crc kubenswrapper[4978]: I0225 10:22:18.423128 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vfnpj"] Feb 25 10:22:19 crc kubenswrapper[4978]: I0225 10:22:19.295610 4978 generic.go:334] "Generic (PLEG): container finished" podID="50fabec8-ad45-4aac-a5f0-56496603e1e8" containerID="8659efbe6a3dcd275ca5cccca6527c9fa8b71b96b5e04600ba1ea3f17a90d9df" exitCode=0 Feb 25 10:22:19 crc kubenswrapper[4978]: I0225 10:22:19.295741 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfnpj" event={"ID":"50fabec8-ad45-4aac-a5f0-56496603e1e8","Type":"ContainerDied","Data":"8659efbe6a3dcd275ca5cccca6527c9fa8b71b96b5e04600ba1ea3f17a90d9df"} Feb 25 10:22:19 crc kubenswrapper[4978]: I0225 10:22:19.295892 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfnpj" event={"ID":"50fabec8-ad45-4aac-a5f0-56496603e1e8","Type":"ContainerStarted","Data":"5600c38ecaad3101aebda5b9b302d8e69d8c80e14a0630ff75e7d4e6b3d9ec99"} Feb 25 10:22:26 crc kubenswrapper[4978]: I0225 10:22:26.862149 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq_9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a/util/0.log" Feb 25 10:22:27 crc kubenswrapper[4978]: I0225 10:22:27.011397 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq_9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a/util/0.log" Feb 25 10:22:27 crc kubenswrapper[4978]: I0225 10:22:27.151528 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq_9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a/pull/0.log" Feb 25 10:22:27 crc kubenswrapper[4978]: I0225 10:22:27.308271 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq_9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a/pull/0.log" Feb 25 10:22:27 crc kubenswrapper[4978]: I0225 10:22:27.388093 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfnpj" event={"ID":"50fabec8-ad45-4aac-a5f0-56496603e1e8","Type":"ContainerStarted","Data":"f6b5f2456cb203369a524958a0d23331b8e6c85dd45383fc55390336634ad3ec"} Feb 25 10:22:27 crc kubenswrapper[4978]: I0225 10:22:27.527247 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq_9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a/pull/0.log" Feb 25 10:22:27 crc kubenswrapper[4978]: I0225 10:22:27.561074 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq_9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a/extract/0.log" Feb 25 10:22:27 crc kubenswrapper[4978]: I0225 10:22:27.593867 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_11a76e9741f3be63a88784b9f3f329441c07f3f3de97b4e48123ebda14bzlgq_9ba2f0c5-8952-4bd9-a3f9-87f72a7c0c6a/util/0.log" Feb 25 10:22:28 crc kubenswrapper[4978]: I0225 10:22:28.204254 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-jwrjk_6dc1c1a6-79f3-48bd-8eab-574e65c9152f/manager/0.log" Feb 25 10:22:28 crc kubenswrapper[4978]: I0225 10:22:28.706571 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-784b5bb6c5-rsf6k_2a48bde4-1ad7-41fd-937e-223509b1b712/manager/0.log" Feb 25 10:22:29 crc kubenswrapper[4978]: I0225 10:22:29.040766 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-gmzk4_b81c87fe-0c25-4ff3-b962-d1bab3e3a91a/manager/0.log" Feb 25 10:22:29 crc kubenswrapper[4978]: I0225 10:22:29.276772 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-whqp4_6299ec19-6e59-472a-a09c-a698bbc547e8/manager/0.log" Feb 25 10:22:29 crc kubenswrapper[4978]: I0225 10:22:29.410331 4978 generic.go:334] "Generic (PLEG): container finished" podID="50fabec8-ad45-4aac-a5f0-56496603e1e8" containerID="f6b5f2456cb203369a524958a0d23331b8e6c85dd45383fc55390336634ad3ec" exitCode=0 Feb 25 10:22:29 crc kubenswrapper[4978]: I0225 10:22:29.410393 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfnpj" event={"ID":"50fabec8-ad45-4aac-a5f0-56496603e1e8","Type":"ContainerDied","Data":"f6b5f2456cb203369a524958a0d23331b8e6c85dd45383fc55390336634ad3ec"} Feb 25 10:22:29 crc kubenswrapper[4978]: I0225 10:22:29.876586 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-xw8bv_3f1862db-5ef7-4749-b7a2-1ee2d65dd3c6/manager/0.log" Feb 25 10:22:30 crc kubenswrapper[4978]: I0225 10:22:30.423165 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfnpj" event={"ID":"50fabec8-ad45-4aac-a5f0-56496603e1e8","Type":"ContainerStarted","Data":"a404e18f665e0d8c7b08827b6f2947ec735bd3011bccf8ba028e6c17e3ab0bb7"} Feb 25 10:22:30 crc kubenswrapper[4978]: I0225 10:22:30.447841 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vfnpj" podStartSLOduration=2.937068258 podStartE2EDuration="13.447824781s" podCreationTimestamp="2026-02-25 10:22:17 +0000 UTC" firstStartedPulling="2026-02-25 10:22:19.297273936 +0000 UTC m=+13032.736530395" lastFinishedPulling="2026-02-25 10:22:29.808030459 +0000 UTC m=+13043.247286918" observedRunningTime="2026-02-25 10:22:30.444352043 +0000 UTC m=+13043.883608522" watchObservedRunningTime="2026-02-25 10:22:30.447824781 +0000 UTC m=+13043.887081230" Feb 25 10:22:30 crc kubenswrapper[4978]: I0225 10:22:30.646890 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-8k5bx_6aa1040d-b0e6-412b-ad79-67e0d2e90b63/manager/0.log" Feb 25 10:22:30 crc kubenswrapper[4978]: I0225 10:22:30.859246 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-7xrm6_2baccbb6-56c4-49ff-aa9b-d5f08b1c6b48/manager/0.log" Feb 25 10:22:31 crc kubenswrapper[4978]: I0225 10:22:31.151019 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-67d996989d-bqwwg_62fb8662-3720-4e7c-a556-08d9b441d1fe/manager/0.log" Feb 25 10:22:31 crc kubenswrapper[4978]: I0225 10:22:31.572141 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-c2vsw_02395a30-3d75-4d54-b057-422d43427f1c/manager/0.log" Feb 25 10:22:32 crc kubenswrapper[4978]: I0225 10:22:32.373121 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6bd4687957-wwpmj_19b7fc05-08e1-4982-9248-4f380e7733e6/manager/0.log" Feb 25 10:22:32 crc kubenswrapper[4978]: I0225 10:22:32.405768 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-grcw6_d1a081d5-4069-4ee0-9f2f-66460514ebe6/manager/0.log" Feb 25 10:22:32 crc kubenswrapper[4978]: I0225 10:22:32.665455 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-659dc6bbfc-pwgkf_72b3dc0a-de07-42e9-b0b6-0897aa5c17c0/manager/0.log" Feb 25 10:22:32 crc kubenswrapper[4978]: I0225 10:22:32.721994 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-579b7786b9ggmk5_fc4de411-fd98-4a38-b7a9-1137da7e8064/manager/0.log" Feb 25 10:22:33 crc kubenswrapper[4978]: I0225 10:22:33.104328 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-55c649df44-nggwh_18a604af-ab94-4884-9750-63f4dd830912/operator/0.log" Feb 25 10:22:33 crc kubenswrapper[4978]: I0225 10:22:33.708613 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-fhmb8_bbee41d9-6d3d-4c1a-ad09-2bc260c74b4e/registry-server/0.log" Feb 25 10:22:33 crc kubenswrapper[4978]: I0225 10:22:33.899238 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5955d8c787-vdnbw_3a6484c8-775f-4c06-817b-2f89501e1c71/manager/0.log" Feb 25 10:22:34 crc kubenswrapper[4978]: I0225 10:22:34.164859 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-9752r_29e606cc-8ae3-4065-8cf1-182f40ab4b98/manager/0.log" Feb 25 10:22:34 crc kubenswrapper[4978]: I0225 10:22:34.236246 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-55d77d7b5c-qk97g_43011269-2605-4cc7-9475-46996e0210e3/manager/0.log" Feb 25 10:22:34 crc kubenswrapper[4978]: I0225 10:22:34.493621 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-77v5j_af52fa17-14f1-4ff3-885e-12fa43af933d/operator/0.log" Feb 25 10:22:34 crc kubenswrapper[4978]: I0225 10:22:34.544293 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-qjw8s_555963fa-06fb-4dd4-854c-9dc86a6dec12/manager/0.log" Feb 25 10:22:35 crc kubenswrapper[4978]: I0225 10:22:35.076538 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5dc6794d5b-8xm7d_5de9178c-622e-4c57-93a8-dcda89354d5e/manager/0.log" Feb 25 10:22:35 crc kubenswrapper[4978]: I0225 10:22:35.288850 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-589c568786-hpd2q_b17d0775-1ea1-43b3-9813-a58f12645b64/manager/0.log" Feb 25 10:22:35 crc kubenswrapper[4978]: I0225 10:22:35.473172 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-bccc79885-njfb8_03f30036-f34a-4751-a133-4ed9cd390ac9/manager/0.log" Feb 25 10:22:37 crc kubenswrapper[4978]: I0225 10:22:37.605197 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5dc486cffc-nwxz6_5d604e16-a85f-4bca-a322-218306711870/manager/0.log" Feb 25 10:22:37 crc kubenswrapper[4978]: I0225 10:22:37.666744 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:37 crc kubenswrapper[4978]: I0225 10:22:37.667157 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:37 crc kubenswrapper[4978]: I0225 10:22:37.721929 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:38 crc kubenswrapper[4978]: I0225 10:22:38.063673 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-xr2cd_42305c1a-c5e6-467b-9f92-75b822ede523/manager/0.log" Feb 25 10:22:38 crc kubenswrapper[4978]: I0225 10:22:38.566984 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vfnpj" Feb 25 10:22:39 crc kubenswrapper[4978]: I0225 10:22:39.083762 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vfnpj"] Feb 25 10:22:39 crc kubenswrapper[4978]: I0225 10:22:39.452338 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mtwcf"] Feb 25 10:22:39 crc kubenswrapper[4978]: I0225 10:22:39.457821 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mtwcf" podUID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerName="registry-server" containerID="cri-o://6f10dbb2be6cba00db21a4ae7cc9bd70d3550e26bb8e5fce9eed23507f8de59c" gracePeriod=2 Feb 25 10:22:40 crc kubenswrapper[4978]: I0225 10:22:40.558512 4978 generic.go:334] "Generic (PLEG): container finished" podID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerID="6f10dbb2be6cba00db21a4ae7cc9bd70d3550e26bb8e5fce9eed23507f8de59c" exitCode=0 Feb 25 10:22:40 crc kubenswrapper[4978]: I0225 10:22:40.559031 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtwcf" event={"ID":"d1d097cc-7cac-41b8-9c33-0be1d52e9918","Type":"ContainerDied","Data":"6f10dbb2be6cba00db21a4ae7cc9bd70d3550e26bb8e5fce9eed23507f8de59c"} Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.345542 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtwcf" Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.461794 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2kb8\" (UniqueName: \"kubernetes.io/projected/d1d097cc-7cac-41b8-9c33-0be1d52e9918-kube-api-access-r2kb8\") pod \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.461832 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-utilities\") pod \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.461995 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-catalog-content\") pod \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\" (UID: \"d1d097cc-7cac-41b8-9c33-0be1d52e9918\") " Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.462726 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-utilities" (OuterVolumeSpecName: "utilities") pod "d1d097cc-7cac-41b8-9c33-0be1d52e9918" (UID: "d1d097cc-7cac-41b8-9c33-0be1d52e9918"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.473661 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d097cc-7cac-41b8-9c33-0be1d52e9918-kube-api-access-r2kb8" (OuterVolumeSpecName: "kube-api-access-r2kb8") pod "d1d097cc-7cac-41b8-9c33-0be1d52e9918" (UID: "d1d097cc-7cac-41b8-9c33-0be1d52e9918"). InnerVolumeSpecName "kube-api-access-r2kb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.522079 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1d097cc-7cac-41b8-9c33-0be1d52e9918" (UID: "d1d097cc-7cac-41b8-9c33-0be1d52e9918"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.563686 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.563715 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2kb8\" (UniqueName: \"kubernetes.io/projected/d1d097cc-7cac-41b8-9c33-0be1d52e9918-kube-api-access-r2kb8\") on node \"crc\" DevicePath \"\"" Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.563725 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d097cc-7cac-41b8-9c33-0be1d52e9918-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.570157 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mtwcf" event={"ID":"d1d097cc-7cac-41b8-9c33-0be1d52e9918","Type":"ContainerDied","Data":"98c41544e660bde518d03c1e9f91556c8f955c6a79d9aa811df17d0426ec55d1"} Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.570457 4978 scope.go:117] "RemoveContainer" containerID="6f10dbb2be6cba00db21a4ae7cc9bd70d3550e26bb8e5fce9eed23507f8de59c" Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.570266 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mtwcf" Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.609512 4978 scope.go:117] "RemoveContainer" containerID="138d1d06c48ea092c213ef55aaf340f0d4a712d611fae9de13cf47ce838f14c4" Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.615983 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mtwcf"] Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.629822 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mtwcf"] Feb 25 10:22:41 crc kubenswrapper[4978]: I0225 10:22:41.635178 4978 scope.go:117] "RemoveContainer" containerID="ee21fe05906659321c182ff687d4eea1413885e64e6af52f40a6ca3514e76d8f" Feb 25 10:22:43 crc kubenswrapper[4978]: I0225 10:22:43.340194 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" path="/var/lib/kubelet/pods/d1d097cc-7cac-41b8-9c33-0be1d52e9918/volumes" Feb 25 10:22:44 crc kubenswrapper[4978]: I0225 10:22:44.422616 4978 scope.go:117] "RemoveContainer" containerID="20885ce537e96d0ac8f4dd743fa139eaf4a5d77e14ada984740fbdece9e5a952" Feb 25 10:23:03 crc kubenswrapper[4978]: I0225 10:23:03.147549 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-rwfz5_20061d67-2059-4022-97ad-607c9f85540c/control-plane-machine-set-operator/0.log" Feb 25 10:23:03 crc kubenswrapper[4978]: I0225 10:23:03.477928 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gkwg7_411cf3c7-6262-4458-a3bd-245b2c002260/kube-rbac-proxy/0.log" Feb 25 10:23:03 crc kubenswrapper[4978]: I0225 10:23:03.521160 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gkwg7_411cf3c7-6262-4458-a3bd-245b2c002260/machine-api-operator/0.log" Feb 25 10:23:21 crc kubenswrapper[4978]: I0225 10:23:21.416591 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-z9ghg_89615955-fd40-4bf0-9bc1-b29188ff03c7/cert-manager-controller/0.log" Feb 25 10:23:21 crc kubenswrapper[4978]: I0225 10:23:21.810226 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-xplnd_f499cab6-b321-4fac-808f-c2ff15ca12ff/cert-manager-cainjector/0.log" Feb 25 10:23:21 crc kubenswrapper[4978]: I0225 10:23:21.874617 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-9sx64_8a262ae0-ec64-4848-9283-92b155feb047/cert-manager-webhook/0.log" Feb 25 10:23:37 crc kubenswrapper[4978]: I0225 10:23:37.757348 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-b7qb4_14d5cc0e-ed1f-4894-afde-ae5164197176/nmstate-console-plugin/0.log" Feb 25 10:23:37 crc kubenswrapper[4978]: I0225 10:23:37.966430 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vjh9q_5a297fa5-61f7-4feb-b6b0-532d3d8509f9/nmstate-handler/0.log" Feb 25 10:23:38 crc kubenswrapper[4978]: I0225 10:23:38.090154 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-m8ksd_31fedbf4-35ef-4440-9039-341b3a0ece67/nmstate-metrics/0.log" Feb 25 10:23:38 crc kubenswrapper[4978]: I0225 10:23:38.097358 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-m8ksd_31fedbf4-35ef-4440-9039-341b3a0ece67/kube-rbac-proxy/0.log" Feb 25 10:23:38 crc kubenswrapper[4978]: I0225 10:23:38.296094 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-6cv5l_107b6a29-7b92-4c24-a38e-5ae9ec345730/nmstate-operator/0.log" Feb 25 10:23:38 crc kubenswrapper[4978]: I0225 10:23:38.421896 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-hlrw5_81d07653-9826-45dd-a0fd-68a2bd37b8e2/nmstate-webhook/0.log" Feb 25 10:23:44 crc kubenswrapper[4978]: I0225 10:23:44.561313 4978 scope.go:117] "RemoveContainer" containerID="958ef308add9ecabcedb70be71262a09d658cbe64605e78301502d761a1bcf8c" Feb 25 10:23:46 crc kubenswrapper[4978]: I0225 10:23:46.540644 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:23:46 crc kubenswrapper[4978]: I0225 10:23:46.540987 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:23:57 crc kubenswrapper[4978]: I0225 10:23:57.529719 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-vw9ts_7701e422-276b-4919-ae42-2901a90faa26/prometheus-operator/0.log" Feb 25 10:23:57 crc kubenswrapper[4978]: I0225 10:23:57.789932 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg_dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5/prometheus-operator-admission-webhook/0.log" Feb 25 10:23:57 crc kubenswrapper[4978]: I0225 10:23:57.924710 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw_bb1b2828-81ee-4bc1-afdb-0306ae16d896/prometheus-operator-admission-webhook/0.log" Feb 25 10:23:58 crc kubenswrapper[4978]: I0225 10:23:58.181452 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-4vd82_79e63f5b-296b-4145-b3b9-83f7ddd2d901/perses-operator/0.log" Feb 25 10:23:58 crc kubenswrapper[4978]: I0225 10:23:58.236225 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-w8cvq_1a1db5e2-0139-4456-9e91-a3f31d7b38fb/operator/0.log" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.167566 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533584-8z88t"] Feb 25 10:24:00 crc kubenswrapper[4978]: E0225 10:24:00.172940 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerName="extract-utilities" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.173187 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerName="extract-utilities" Feb 25 10:24:00 crc kubenswrapper[4978]: E0225 10:24:00.173237 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerName="registry-server" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.173245 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerName="registry-server" Feb 25 10:24:00 crc kubenswrapper[4978]: E0225 10:24:00.173268 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerName="extract-content" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.173274 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerName="extract-content" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.173578 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d097cc-7cac-41b8-9c33-0be1d52e9918" containerName="registry-server" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.174577 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533584-8z88t" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.177157 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.177708 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.177955 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.205443 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533584-8z88t"] Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.329043 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql97j\" (UniqueName: \"kubernetes.io/projected/b5b94811-1a8c-419a-8101-ba50b4dad2a5-kube-api-access-ql97j\") pod \"auto-csr-approver-29533584-8z88t\" (UID: \"b5b94811-1a8c-419a-8101-ba50b4dad2a5\") " pod="openshift-infra/auto-csr-approver-29533584-8z88t" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.431465 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql97j\" (UniqueName: \"kubernetes.io/projected/b5b94811-1a8c-419a-8101-ba50b4dad2a5-kube-api-access-ql97j\") pod \"auto-csr-approver-29533584-8z88t\" (UID: \"b5b94811-1a8c-419a-8101-ba50b4dad2a5\") " pod="openshift-infra/auto-csr-approver-29533584-8z88t" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.458115 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql97j\" (UniqueName: \"kubernetes.io/projected/b5b94811-1a8c-419a-8101-ba50b4dad2a5-kube-api-access-ql97j\") pod \"auto-csr-approver-29533584-8z88t\" (UID: \"b5b94811-1a8c-419a-8101-ba50b4dad2a5\") " pod="openshift-infra/auto-csr-approver-29533584-8z88t" Feb 25 10:24:00 crc kubenswrapper[4978]: I0225 10:24:00.532172 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533584-8z88t" Feb 25 10:24:01 crc kubenswrapper[4978]: I0225 10:24:01.268628 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533584-8z88t"] Feb 25 10:24:01 crc kubenswrapper[4978]: I0225 10:24:01.339554 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533584-8z88t" event={"ID":"b5b94811-1a8c-419a-8101-ba50b4dad2a5","Type":"ContainerStarted","Data":"b74d2252b972116d8d37a642724e270cc4d155c035405387eae3ae9b4365e4e8"} Feb 25 10:24:03 crc kubenswrapper[4978]: I0225 10:24:03.357802 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533584-8z88t" event={"ID":"b5b94811-1a8c-419a-8101-ba50b4dad2a5","Type":"ContainerStarted","Data":"fe1e70f5d9d044846b69ebe72e4ab4be5581c19762b8fbabb2b7bc5882d627f5"} Feb 25 10:24:03 crc kubenswrapper[4978]: I0225 10:24:03.381739 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533584-8z88t" podStartSLOduration=2.529006689 podStartE2EDuration="3.381713472s" podCreationTimestamp="2026-02-25 10:24:00 +0000 UTC" firstStartedPulling="2026-02-25 10:24:01.276210964 +0000 UTC m=+13134.715467423" lastFinishedPulling="2026-02-25 10:24:02.128917737 +0000 UTC m=+13135.568174206" observedRunningTime="2026-02-25 10:24:03.371162525 +0000 UTC m=+13136.810418994" watchObservedRunningTime="2026-02-25 10:24:03.381713472 +0000 UTC m=+13136.820969931" Feb 25 10:24:04 crc kubenswrapper[4978]: I0225 10:24:04.375972 4978 generic.go:334] "Generic (PLEG): container finished" podID="b5b94811-1a8c-419a-8101-ba50b4dad2a5" containerID="fe1e70f5d9d044846b69ebe72e4ab4be5581c19762b8fbabb2b7bc5882d627f5" exitCode=0 Feb 25 10:24:04 crc kubenswrapper[4978]: I0225 10:24:04.376022 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533584-8z88t" event={"ID":"b5b94811-1a8c-419a-8101-ba50b4dad2a5","Type":"ContainerDied","Data":"fe1e70f5d9d044846b69ebe72e4ab4be5581c19762b8fbabb2b7bc5882d627f5"} Feb 25 10:24:06 crc kubenswrapper[4978]: I0225 10:24:06.911171 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533584-8z88t" Feb 25 10:24:06 crc kubenswrapper[4978]: I0225 10:24:06.997821 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql97j\" (UniqueName: \"kubernetes.io/projected/b5b94811-1a8c-419a-8101-ba50b4dad2a5-kube-api-access-ql97j\") pod \"b5b94811-1a8c-419a-8101-ba50b4dad2a5\" (UID: \"b5b94811-1a8c-419a-8101-ba50b4dad2a5\") " Feb 25 10:24:07 crc kubenswrapper[4978]: I0225 10:24:07.003496 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b94811-1a8c-419a-8101-ba50b4dad2a5-kube-api-access-ql97j" (OuterVolumeSpecName: "kube-api-access-ql97j") pod "b5b94811-1a8c-419a-8101-ba50b4dad2a5" (UID: "b5b94811-1a8c-419a-8101-ba50b4dad2a5"). InnerVolumeSpecName "kube-api-access-ql97j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:24:07 crc kubenswrapper[4978]: I0225 10:24:07.100620 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql97j\" (UniqueName: \"kubernetes.io/projected/b5b94811-1a8c-419a-8101-ba50b4dad2a5-kube-api-access-ql97j\") on node \"crc\" DevicePath \"\"" Feb 25 10:24:07 crc kubenswrapper[4978]: I0225 10:24:07.405057 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533584-8z88t" event={"ID":"b5b94811-1a8c-419a-8101-ba50b4dad2a5","Type":"ContainerDied","Data":"b74d2252b972116d8d37a642724e270cc4d155c035405387eae3ae9b4365e4e8"} Feb 25 10:24:07 crc kubenswrapper[4978]: I0225 10:24:07.405378 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b74d2252b972116d8d37a642724e270cc4d155c035405387eae3ae9b4365e4e8" Feb 25 10:24:07 crc kubenswrapper[4978]: I0225 10:24:07.405444 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533584-8z88t" Feb 25 10:24:07 crc kubenswrapper[4978]: I0225 10:24:07.981952 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533578-mpvsg"] Feb 25 10:24:07 crc kubenswrapper[4978]: I0225 10:24:07.994589 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533578-mpvsg"] Feb 25 10:24:09 crc kubenswrapper[4978]: I0225 10:24:09.339753 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3aa68c1-2c13-4bed-bc6e-87342302509f" path="/var/lib/kubelet/pods/a3aa68c1-2c13-4bed-bc6e-87342302509f/volumes" Feb 25 10:24:16 crc kubenswrapper[4978]: I0225 10:24:16.540520 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:24:16 crc kubenswrapper[4978]: I0225 10:24:16.541016 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:24:17 crc kubenswrapper[4978]: I0225 10:24:17.366358 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-xwfgl_3a29c614-8113-4dd5-8209-3bf914925474/frr-k8s-webhook-server/0.log" Feb 25 10:24:17 crc kubenswrapper[4978]: I0225 10:24:17.366606 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-6fkw5_22ded0ab-9ff4-42c2-80fd-7423e610c25b/kube-rbac-proxy/0.log" Feb 25 10:24:17 crc kubenswrapper[4978]: I0225 10:24:17.722347 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-6fkw5_22ded0ab-9ff4-42c2-80fd-7423e610c25b/controller/0.log" Feb 25 10:24:17 crc kubenswrapper[4978]: I0225 10:24:17.774670 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-frr-files/0.log" Feb 25 10:24:17 crc kubenswrapper[4978]: I0225 10:24:17.973056 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-reloader/0.log" Feb 25 10:24:17 crc kubenswrapper[4978]: I0225 10:24:17.978942 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-metrics/0.log" Feb 25 10:24:17 crc kubenswrapper[4978]: I0225 10:24:17.984327 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-frr-files/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.049159 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-reloader/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.242885 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-frr-files/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.267990 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-metrics/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.272214 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-metrics/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.293432 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-reloader/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.486854 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-frr-files/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.532438 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-reloader/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.554104 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/cp-metrics/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.582500 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/controller/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.762490 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/frr-metrics/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.808570 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/kube-rbac-proxy-frr/0.log" Feb 25 10:24:18 crc kubenswrapper[4978]: I0225 10:24:18.904109 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/kube-rbac-proxy/0.log" Feb 25 10:24:19 crc kubenswrapper[4978]: I0225 10:24:19.041263 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/reloader/0.log" Feb 25 10:24:19 crc kubenswrapper[4978]: I0225 10:24:19.145020 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5769477c66-cbbds_296657b4-46fe-4780-b077-65d11eeab81d/manager/0.log" Feb 25 10:24:19 crc kubenswrapper[4978]: I0225 10:24:19.631803 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5b647b685b-2jdt4_c4887490-f9e3-4a5f-be93-ba244beaf4e7/webhook-server/0.log" Feb 25 10:24:19 crc kubenswrapper[4978]: I0225 10:24:19.915115 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bkcdb_0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb/kube-rbac-proxy/0.log" Feb 25 10:24:20 crc kubenswrapper[4978]: I0225 10:24:20.691540 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bkcdb_0aaebfee-d253-4b6a-bdc0-acbd4ef12cdb/speaker/0.log" Feb 25 10:24:22 crc kubenswrapper[4978]: I0225 10:24:22.896934 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-xb4dk_65f4ea4a-d2f4-4180-adbd-e464b90c84a0/frr/0.log" Feb 25 10:24:38 crc kubenswrapper[4978]: I0225 10:24:38.430921 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v_6a65fcb7-9cf8-40b8-96d5-4b520ad4357e/util/0.log" Feb 25 10:24:38 crc kubenswrapper[4978]: I0225 10:24:38.728630 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v_6a65fcb7-9cf8-40b8-96d5-4b520ad4357e/util/0.log" Feb 25 10:24:38 crc kubenswrapper[4978]: I0225 10:24:38.758419 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v_6a65fcb7-9cf8-40b8-96d5-4b520ad4357e/pull/0.log" Feb 25 10:24:38 crc kubenswrapper[4978]: I0225 10:24:38.898749 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v_6a65fcb7-9cf8-40b8-96d5-4b520ad4357e/pull/0.log" Feb 25 10:24:39 crc kubenswrapper[4978]: I0225 10:24:39.113115 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v_6a65fcb7-9cf8-40b8-96d5-4b520ad4357e/extract/0.log" Feb 25 10:24:39 crc kubenswrapper[4978]: I0225 10:24:39.138335 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v_6a65fcb7-9cf8-40b8-96d5-4b520ad4357e/util/0.log" Feb 25 10:24:39 crc kubenswrapper[4978]: I0225 10:24:39.196737 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dcz6v_6a65fcb7-9cf8-40b8-96d5-4b520ad4357e/pull/0.log" Feb 25 10:24:39 crc kubenswrapper[4978]: I0225 10:24:39.349317 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n_9cdf7cf4-2d4a-481c-9d02-0f32e00f6035/util/0.log" Feb 25 10:24:39 crc kubenswrapper[4978]: I0225 10:24:39.613896 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n_9cdf7cf4-2d4a-481c-9d02-0f32e00f6035/util/0.log" Feb 25 10:24:39 crc kubenswrapper[4978]: I0225 10:24:39.650440 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n_9cdf7cf4-2d4a-481c-9d02-0f32e00f6035/pull/0.log" Feb 25 10:24:39 crc kubenswrapper[4978]: I0225 10:24:39.666354 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n_9cdf7cf4-2d4a-481c-9d02-0f32e00f6035/pull/0.log" Feb 25 10:24:39 crc kubenswrapper[4978]: I0225 10:24:39.908518 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n_9cdf7cf4-2d4a-481c-9d02-0f32e00f6035/pull/0.log" Feb 25 10:24:39 crc kubenswrapper[4978]: I0225 10:24:39.923790 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n_9cdf7cf4-2d4a-481c-9d02-0f32e00f6035/util/0.log" Feb 25 10:24:40 crc kubenswrapper[4978]: I0225 10:24:40.013004 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08fbn9n_9cdf7cf4-2d4a-481c-9d02-0f32e00f6035/extract/0.log" Feb 25 10:24:40 crc kubenswrapper[4978]: I0225 10:24:40.117581 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt_0c1cb1bf-977e-400b-a0c6-2058d62e1ff4/util/0.log" Feb 25 10:24:40 crc kubenswrapper[4978]: I0225 10:24:40.460488 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt_0c1cb1bf-977e-400b-a0c6-2058d62e1ff4/pull/0.log" Feb 25 10:24:40 crc kubenswrapper[4978]: I0225 10:24:40.463811 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt_0c1cb1bf-977e-400b-a0c6-2058d62e1ff4/pull/0.log" Feb 25 10:24:40 crc kubenswrapper[4978]: I0225 10:24:40.512521 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt_0c1cb1bf-977e-400b-a0c6-2058d62e1ff4/util/0.log" Feb 25 10:24:40 crc kubenswrapper[4978]: I0225 10:24:40.735888 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt_0c1cb1bf-977e-400b-a0c6-2058d62e1ff4/util/0.log" Feb 25 10:24:40 crc kubenswrapper[4978]: I0225 10:24:40.737946 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt_0c1cb1bf-977e-400b-a0c6-2058d62e1ff4/extract/0.log" Feb 25 10:24:40 crc kubenswrapper[4978]: I0225 10:24:40.872337 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213mb4nt_0c1cb1bf-977e-400b-a0c6-2058d62e1ff4/pull/0.log" Feb 25 10:24:40 crc kubenswrapper[4978]: I0225 10:24:40.951390 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dncwv_01c8bc46-7771-4b6a-b27c-49a507df6627/extract-utilities/0.log" Feb 25 10:24:41 crc kubenswrapper[4978]: I0225 10:24:41.194660 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dncwv_01c8bc46-7771-4b6a-b27c-49a507df6627/extract-utilities/0.log" Feb 25 10:24:41 crc kubenswrapper[4978]: I0225 10:24:41.254008 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dncwv_01c8bc46-7771-4b6a-b27c-49a507df6627/extract-content/0.log" Feb 25 10:24:41 crc kubenswrapper[4978]: I0225 10:24:41.269119 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dncwv_01c8bc46-7771-4b6a-b27c-49a507df6627/extract-content/0.log" Feb 25 10:24:41 crc kubenswrapper[4978]: I0225 10:24:41.490414 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dncwv_01c8bc46-7771-4b6a-b27c-49a507df6627/extract-content/0.log" Feb 25 10:24:41 crc kubenswrapper[4978]: I0225 10:24:41.502798 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dncwv_01c8bc46-7771-4b6a-b27c-49a507df6627/extract-utilities/0.log" Feb 25 10:24:41 crc kubenswrapper[4978]: I0225 10:24:41.720583 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-dncwv_01c8bc46-7771-4b6a-b27c-49a507df6627/registry-server/0.log" Feb 25 10:24:41 crc kubenswrapper[4978]: I0225 10:24:41.768084 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vfnpj_50fabec8-ad45-4aac-a5f0-56496603e1e8/extract-utilities/0.log" Feb 25 10:24:41 crc kubenswrapper[4978]: I0225 10:24:41.940069 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vfnpj_50fabec8-ad45-4aac-a5f0-56496603e1e8/extract-utilities/0.log" Feb 25 10:24:41 crc kubenswrapper[4978]: I0225 10:24:41.973905 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vfnpj_50fabec8-ad45-4aac-a5f0-56496603e1e8/extract-content/0.log" Feb 25 10:24:42 crc kubenswrapper[4978]: I0225 10:24:42.005331 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vfnpj_50fabec8-ad45-4aac-a5f0-56496603e1e8/extract-content/0.log" Feb 25 10:24:42 crc kubenswrapper[4978]: I0225 10:24:42.209019 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vfnpj_50fabec8-ad45-4aac-a5f0-56496603e1e8/extract-utilities/0.log" Feb 25 10:24:42 crc kubenswrapper[4978]: I0225 10:24:42.302157 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vfnpj_50fabec8-ad45-4aac-a5f0-56496603e1e8/extract-content/0.log" Feb 25 10:24:42 crc kubenswrapper[4978]: I0225 10:24:42.374068 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vfnpj_50fabec8-ad45-4aac-a5f0-56496603e1e8/registry-server/0.log" Feb 25 10:24:42 crc kubenswrapper[4978]: I0225 10:24:42.458743 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl_107428a7-7b98-43a1-9bd1-a6c08c914569/util/0.log" Feb 25 10:24:42 crc kubenswrapper[4978]: I0225 10:24:42.651944 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl_107428a7-7b98-43a1-9bd1-a6c08c914569/util/0.log" Feb 25 10:24:42 crc kubenswrapper[4978]: I0225 10:24:42.689497 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl_107428a7-7b98-43a1-9bd1-a6c08c914569/pull/0.log" Feb 25 10:24:42 crc kubenswrapper[4978]: I0225 10:24:42.762233 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl_107428a7-7b98-43a1-9bd1-a6c08c914569/pull/0.log" Feb 25 10:24:42 crc kubenswrapper[4978]: I0225 10:24:42.920160 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl_107428a7-7b98-43a1-9bd1-a6c08c914569/util/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.026832 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl_107428a7-7b98-43a1-9bd1-a6c08c914569/pull/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.030306 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-hpjsn_ecf97051-f58e-4888-8e7e-f03481e89bea/marketplace-operator/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.038502 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecaxzhbl_107428a7-7b98-43a1-9bd1-a6c08c914569/extract/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.223772 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dgg7r_e4a76118-7308-4b0f-aae5-41cd6888a845/extract-utilities/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.378661 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dgg7r_e4a76118-7308-4b0f-aae5-41cd6888a845/extract-utilities/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.380950 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dgg7r_e4a76118-7308-4b0f-aae5-41cd6888a845/extract-content/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.391242 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dgg7r_e4a76118-7308-4b0f-aae5-41cd6888a845/extract-content/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.584353 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dgg7r_e4a76118-7308-4b0f-aae5-41cd6888a845/extract-utilities/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.606123 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dgg7r_e4a76118-7308-4b0f-aae5-41cd6888a845/extract-content/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.678675 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jr4n7_a1af080e-55f9-4d48-82c6-02160123e37f/extract-utilities/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.852255 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jr4n7_a1af080e-55f9-4d48-82c6-02160123e37f/extract-utilities/0.log" Feb 25 10:24:43 crc kubenswrapper[4978]: I0225 10:24:43.966971 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jr4n7_a1af080e-55f9-4d48-82c6-02160123e37f/extract-content/0.log" Feb 25 10:24:44 crc kubenswrapper[4978]: I0225 10:24:44.002883 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dgg7r_e4a76118-7308-4b0f-aae5-41cd6888a845/registry-server/0.log" Feb 25 10:24:44 crc kubenswrapper[4978]: I0225 10:24:44.020737 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jr4n7_a1af080e-55f9-4d48-82c6-02160123e37f/extract-content/0.log" Feb 25 10:24:44 crc kubenswrapper[4978]: I0225 10:24:44.147656 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jr4n7_a1af080e-55f9-4d48-82c6-02160123e37f/extract-content/0.log" Feb 25 10:24:44 crc kubenswrapper[4978]: I0225 10:24:44.200241 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jr4n7_a1af080e-55f9-4d48-82c6-02160123e37f/extract-utilities/0.log" Feb 25 10:24:44 crc kubenswrapper[4978]: I0225 10:24:44.575393 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-jr4n7_a1af080e-55f9-4d48-82c6-02160123e37f/registry-server/0.log" Feb 25 10:24:44 crc kubenswrapper[4978]: I0225 10:24:44.642867 4978 scope.go:117] "RemoveContainer" containerID="a195044fb5c61a961d79848819aa5025ad19ca13589da9a8f91eadc29a2167a4" Feb 25 10:24:44 crc kubenswrapper[4978]: I0225 10:24:44.672974 4978 scope.go:117] "RemoveContainer" containerID="4e125d1aec87b4b1c2f404989d66d76ca9469688131870420a2372d61fed8027" Feb 25 10:24:46 crc kubenswrapper[4978]: I0225 10:24:46.540944 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:24:46 crc kubenswrapper[4978]: I0225 10:24:46.541245 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:24:46 crc kubenswrapper[4978]: I0225 10:24:46.541292 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 10:24:46 crc kubenswrapper[4978]: I0225 10:24:46.542089 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"95899c3bd9197bd5e205d8ad51463abe0d611cd768f6460f30bac2b6f6450c64"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 10:24:46 crc kubenswrapper[4978]: I0225 10:24:46.542159 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://95899c3bd9197bd5e205d8ad51463abe0d611cd768f6460f30bac2b6f6450c64" gracePeriod=600 Feb 25 10:24:46 crc kubenswrapper[4978]: I0225 10:24:46.805425 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="95899c3bd9197bd5e205d8ad51463abe0d611cd768f6460f30bac2b6f6450c64" exitCode=0 Feb 25 10:24:46 crc kubenswrapper[4978]: I0225 10:24:46.805467 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"95899c3bd9197bd5e205d8ad51463abe0d611cd768f6460f30bac2b6f6450c64"} Feb 25 10:24:46 crc kubenswrapper[4978]: I0225 10:24:46.805500 4978 scope.go:117] "RemoveContainer" containerID="1517eb9e3e05ae95c645c9adf3b363199c3be490907cec81feab85026f826638" Feb 25 10:24:47 crc kubenswrapper[4978]: I0225 10:24:47.815867 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988"} Feb 25 10:24:58 crc kubenswrapper[4978]: I0225 10:24:58.684432 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-vw9ts_7701e422-276b-4919-ae42-2901a90faa26/prometheus-operator/0.log" Feb 25 10:24:58 crc kubenswrapper[4978]: I0225 10:24:58.723259 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c6c668764-2fhdg_dbf2dd91-faf4-4ef7-86f9-1d29b411dbe5/prometheus-operator-admission-webhook/0.log" Feb 25 10:24:58 crc kubenswrapper[4978]: I0225 10:24:58.761059 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7c6c668764-d4xtw_bb1b2828-81ee-4bc1-afdb-0306ae16d896/prometheus-operator-admission-webhook/0.log" Feb 25 10:24:58 crc kubenswrapper[4978]: I0225 10:24:58.923649 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-4vd82_79e63f5b-296b-4145-b3b9-83f7ddd2d901/perses-operator/0.log" Feb 25 10:24:58 crc kubenswrapper[4978]: I0225 10:24:58.948885 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-w8cvq_1a1db5e2-0139-4456-9e91-a3f31d7b38fb/operator/0.log" Feb 25 10:25:03 crc kubenswrapper[4978]: E0225 10:25:03.860313 4978 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.146:50288->38.102.83.146:35177: write tcp 38.102.83.146:50288->38.102.83.146:35177: write: broken pipe Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.026453 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-clk2j"] Feb 25 10:25:15 crc kubenswrapper[4978]: E0225 10:25:15.027451 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b94811-1a8c-419a-8101-ba50b4dad2a5" containerName="oc" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.027463 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b94811-1a8c-419a-8101-ba50b4dad2a5" containerName="oc" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.027685 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b94811-1a8c-419a-8101-ba50b4dad2a5" containerName="oc" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.029073 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.036472 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-clk2j"] Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.151121 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-utilities\") pod \"redhat-operators-clk2j\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.151535 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch7rc\" (UniqueName: \"kubernetes.io/projected/80c73372-fc1a-4447-b345-1eaf32c0ef94-kube-api-access-ch7rc\") pod \"redhat-operators-clk2j\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.151677 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-catalog-content\") pod \"redhat-operators-clk2j\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.253647 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-utilities\") pod \"redhat-operators-clk2j\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.254051 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch7rc\" (UniqueName: \"kubernetes.io/projected/80c73372-fc1a-4447-b345-1eaf32c0ef94-kube-api-access-ch7rc\") pod \"redhat-operators-clk2j\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.254180 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-catalog-content\") pod \"redhat-operators-clk2j\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.254790 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-catalog-content\") pod \"redhat-operators-clk2j\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.254858 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-utilities\") pod \"redhat-operators-clk2j\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.275269 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch7rc\" (UniqueName: \"kubernetes.io/projected/80c73372-fc1a-4447-b345-1eaf32c0ef94-kube-api-access-ch7rc\") pod \"redhat-operators-clk2j\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:15 crc kubenswrapper[4978]: I0225 10:25:15.356348 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:16 crc kubenswrapper[4978]: I0225 10:25:16.245220 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-clk2j"] Feb 25 10:25:16 crc kubenswrapper[4978]: W0225 10:25:16.254787 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80c73372_fc1a_4447_b345_1eaf32c0ef94.slice/crio-a1fe81e38612a4ed8b8517726c5be69c9c49facfdb12b118d6f30f1d238c9584 WatchSource:0}: Error finding container a1fe81e38612a4ed8b8517726c5be69c9c49facfdb12b118d6f30f1d238c9584: Status 404 returned error can't find the container with id a1fe81e38612a4ed8b8517726c5be69c9c49facfdb12b118d6f30f1d238c9584 Feb 25 10:25:17 crc kubenswrapper[4978]: I0225 10:25:17.156565 4978 generic.go:334] "Generic (PLEG): container finished" podID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerID="714510dd3e5878a76c6d0b85b85184139331cace74646975a1c89d0f9d9ae37a" exitCode=0 Feb 25 10:25:17 crc kubenswrapper[4978]: I0225 10:25:17.156797 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clk2j" event={"ID":"80c73372-fc1a-4447-b345-1eaf32c0ef94","Type":"ContainerDied","Data":"714510dd3e5878a76c6d0b85b85184139331cace74646975a1c89d0f9d9ae37a"} Feb 25 10:25:17 crc kubenswrapper[4978]: I0225 10:25:17.156821 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clk2j" event={"ID":"80c73372-fc1a-4447-b345-1eaf32c0ef94","Type":"ContainerStarted","Data":"a1fe81e38612a4ed8b8517726c5be69c9c49facfdb12b118d6f30f1d238c9584"} Feb 25 10:25:19 crc kubenswrapper[4978]: I0225 10:25:19.181937 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clk2j" event={"ID":"80c73372-fc1a-4447-b345-1eaf32c0ef94","Type":"ContainerStarted","Data":"a159c47a522e853f435d6ad1a35104efa8a3ee13d75c910871ab396c88cc705c"} Feb 25 10:25:22 crc kubenswrapper[4978]: E0225 10:25:22.724963 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80c73372_fc1a_4447_b345_1eaf32c0ef94.slice/crio-conmon-a159c47a522e853f435d6ad1a35104efa8a3ee13d75c910871ab396c88cc705c.scope\": RecentStats: unable to find data in memory cache]" Feb 25 10:25:23 crc kubenswrapper[4978]: I0225 10:25:23.217659 4978 generic.go:334] "Generic (PLEG): container finished" podID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerID="a159c47a522e853f435d6ad1a35104efa8a3ee13d75c910871ab396c88cc705c" exitCode=0 Feb 25 10:25:23 crc kubenswrapper[4978]: I0225 10:25:23.217723 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clk2j" event={"ID":"80c73372-fc1a-4447-b345-1eaf32c0ef94","Type":"ContainerDied","Data":"a159c47a522e853f435d6ad1a35104efa8a3ee13d75c910871ab396c88cc705c"} Feb 25 10:25:24 crc kubenswrapper[4978]: I0225 10:25:24.229787 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clk2j" event={"ID":"80c73372-fc1a-4447-b345-1eaf32c0ef94","Type":"ContainerStarted","Data":"acb4059b02b5e89d61e0680d5159d8e850b19a2195b86ba2def236a4a27a075f"} Feb 25 10:25:24 crc kubenswrapper[4978]: I0225 10:25:24.247162 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-clk2j" podStartSLOduration=2.809045339 podStartE2EDuration="9.247140213s" podCreationTimestamp="2026-02-25 10:25:15 +0000 UTC" firstStartedPulling="2026-02-25 10:25:17.159665129 +0000 UTC m=+13210.598921588" lastFinishedPulling="2026-02-25 10:25:23.597760013 +0000 UTC m=+13217.037016462" observedRunningTime="2026-02-25 10:25:24.244607814 +0000 UTC m=+13217.683864313" watchObservedRunningTime="2026-02-25 10:25:24.247140213 +0000 UTC m=+13217.686396692" Feb 25 10:25:25 crc kubenswrapper[4978]: I0225 10:25:25.356838 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:25 crc kubenswrapper[4978]: I0225 10:25:25.357183 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:26 crc kubenswrapper[4978]: I0225 10:25:26.420486 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-clk2j" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerName="registry-server" probeResult="failure" output=< Feb 25 10:25:26 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 10:25:26 crc kubenswrapper[4978]: > Feb 25 10:25:36 crc kubenswrapper[4978]: I0225 10:25:36.412039 4978 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-clk2j" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerName="registry-server" probeResult="failure" output=< Feb 25 10:25:36 crc kubenswrapper[4978]: timeout: failed to connect service ":50051" within 1s Feb 25 10:25:36 crc kubenswrapper[4978]: > Feb 25 10:25:45 crc kubenswrapper[4978]: I0225 10:25:45.413846 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:45 crc kubenswrapper[4978]: I0225 10:25:45.489191 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:46 crc kubenswrapper[4978]: I0225 10:25:46.232470 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-clk2j"] Feb 25 10:25:46 crc kubenswrapper[4978]: I0225 10:25:46.437936 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-clk2j" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerName="registry-server" containerID="cri-o://acb4059b02b5e89d61e0680d5159d8e850b19a2195b86ba2def236a4a27a075f" gracePeriod=2 Feb 25 10:25:47 crc kubenswrapper[4978]: I0225 10:25:47.452030 4978 generic.go:334] "Generic (PLEG): container finished" podID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerID="acb4059b02b5e89d61e0680d5159d8e850b19a2195b86ba2def236a4a27a075f" exitCode=0 Feb 25 10:25:47 crc kubenswrapper[4978]: I0225 10:25:47.452537 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clk2j" event={"ID":"80c73372-fc1a-4447-b345-1eaf32c0ef94","Type":"ContainerDied","Data":"acb4059b02b5e89d61e0680d5159d8e850b19a2195b86ba2def236a4a27a075f"} Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.097867 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.232120 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-utilities\") pod \"80c73372-fc1a-4447-b345-1eaf32c0ef94\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.232254 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch7rc\" (UniqueName: \"kubernetes.io/projected/80c73372-fc1a-4447-b345-1eaf32c0ef94-kube-api-access-ch7rc\") pod \"80c73372-fc1a-4447-b345-1eaf32c0ef94\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.232323 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-catalog-content\") pod \"80c73372-fc1a-4447-b345-1eaf32c0ef94\" (UID: \"80c73372-fc1a-4447-b345-1eaf32c0ef94\") " Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.232977 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-utilities" (OuterVolumeSpecName: "utilities") pod "80c73372-fc1a-4447-b345-1eaf32c0ef94" (UID: "80c73372-fc1a-4447-b345-1eaf32c0ef94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.238889 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80c73372-fc1a-4447-b345-1eaf32c0ef94-kube-api-access-ch7rc" (OuterVolumeSpecName: "kube-api-access-ch7rc") pod "80c73372-fc1a-4447-b345-1eaf32c0ef94" (UID: "80c73372-fc1a-4447-b345-1eaf32c0ef94"). InnerVolumeSpecName "kube-api-access-ch7rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.334979 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.335014 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch7rc\" (UniqueName: \"kubernetes.io/projected/80c73372-fc1a-4447-b345-1eaf32c0ef94-kube-api-access-ch7rc\") on node \"crc\" DevicePath \"\"" Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.362813 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80c73372-fc1a-4447-b345-1eaf32c0ef94" (UID: "80c73372-fc1a-4447-b345-1eaf32c0ef94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.437904 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80c73372-fc1a-4447-b345-1eaf32c0ef94-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.466498 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-clk2j" event={"ID":"80c73372-fc1a-4447-b345-1eaf32c0ef94","Type":"ContainerDied","Data":"a1fe81e38612a4ed8b8517726c5be69c9c49facfdb12b118d6f30f1d238c9584"} Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.466558 4978 scope.go:117] "RemoveContainer" containerID="acb4059b02b5e89d61e0680d5159d8e850b19a2195b86ba2def236a4a27a075f" Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.466709 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-clk2j" Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.497039 4978 scope.go:117] "RemoveContainer" containerID="a159c47a522e853f435d6ad1a35104efa8a3ee13d75c910871ab396c88cc705c" Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.508224 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-clk2j"] Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.522706 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-clk2j"] Feb 25 10:25:48 crc kubenswrapper[4978]: I0225 10:25:48.546688 4978 scope.go:117] "RemoveContainer" containerID="714510dd3e5878a76c6d0b85b85184139331cace74646975a1c89d0f9d9ae37a" Feb 25 10:25:49 crc kubenswrapper[4978]: I0225 10:25:49.343042 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" path="/var/lib/kubelet/pods/80c73372-fc1a-4447-b345-1eaf32c0ef94/volumes" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.156353 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533586-c6l5k"] Feb 25 10:26:00 crc kubenswrapper[4978]: E0225 10:26:00.157794 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerName="registry-server" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.157820 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerName="registry-server" Feb 25 10:26:00 crc kubenswrapper[4978]: E0225 10:26:00.157863 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerName="extract-content" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.157874 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerName="extract-content" Feb 25 10:26:00 crc kubenswrapper[4978]: E0225 10:26:00.157928 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerName="extract-utilities" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.157942 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerName="extract-utilities" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.158330 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="80c73372-fc1a-4447-b345-1eaf32c0ef94" containerName="registry-server" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.159673 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533586-c6l5k" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.164345 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.164427 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.165524 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.165900 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533586-c6l5k"] Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.321028 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmwfc\" (UniqueName: \"kubernetes.io/projected/6da167fe-0017-4734-be04-9c649f761d27-kube-api-access-mmwfc\") pod \"auto-csr-approver-29533586-c6l5k\" (UID: \"6da167fe-0017-4734-be04-9c649f761d27\") " pod="openshift-infra/auto-csr-approver-29533586-c6l5k" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.423005 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmwfc\" (UniqueName: \"kubernetes.io/projected/6da167fe-0017-4734-be04-9c649f761d27-kube-api-access-mmwfc\") pod \"auto-csr-approver-29533586-c6l5k\" (UID: \"6da167fe-0017-4734-be04-9c649f761d27\") " pod="openshift-infra/auto-csr-approver-29533586-c6l5k" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.446739 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmwfc\" (UniqueName: \"kubernetes.io/projected/6da167fe-0017-4734-be04-9c649f761d27-kube-api-access-mmwfc\") pod \"auto-csr-approver-29533586-c6l5k\" (UID: \"6da167fe-0017-4734-be04-9c649f761d27\") " pod="openshift-infra/auto-csr-approver-29533586-c6l5k" Feb 25 10:26:00 crc kubenswrapper[4978]: I0225 10:26:00.515417 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533586-c6l5k" Feb 25 10:26:01 crc kubenswrapper[4978]: I0225 10:26:01.244315 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533586-c6l5k"] Feb 25 10:26:01 crc kubenswrapper[4978]: I0225 10:26:01.639695 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533586-c6l5k" event={"ID":"6da167fe-0017-4734-be04-9c649f761d27","Type":"ContainerStarted","Data":"0598f782617be58d5bbca7cd6211ed9a9f1c8156c63b1248abaa0d2490bb6a0c"} Feb 25 10:26:02 crc kubenswrapper[4978]: I0225 10:26:02.653401 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533586-c6l5k" event={"ID":"6da167fe-0017-4734-be04-9c649f761d27","Type":"ContainerStarted","Data":"063b61665c2ee996631b73afe9b0f055c2b932e6268e0772639d782937225569"} Feb 25 10:26:02 crc kubenswrapper[4978]: I0225 10:26:02.679278 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533586-c6l5k" podStartSLOduration=1.8270910900000001 podStartE2EDuration="2.679181875s" podCreationTimestamp="2026-02-25 10:26:00 +0000 UTC" firstStartedPulling="2026-02-25 10:26:01.258943649 +0000 UTC m=+13254.698200108" lastFinishedPulling="2026-02-25 10:26:02.111034434 +0000 UTC m=+13255.550290893" observedRunningTime="2026-02-25 10:26:02.666871153 +0000 UTC m=+13256.106127612" watchObservedRunningTime="2026-02-25 10:26:02.679181875 +0000 UTC m=+13256.118438334" Feb 25 10:26:03 crc kubenswrapper[4978]: I0225 10:26:03.666988 4978 generic.go:334] "Generic (PLEG): container finished" podID="6da167fe-0017-4734-be04-9c649f761d27" containerID="063b61665c2ee996631b73afe9b0f055c2b932e6268e0772639d782937225569" exitCode=0 Feb 25 10:26:03 crc kubenswrapper[4978]: I0225 10:26:03.667039 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533586-c6l5k" event={"ID":"6da167fe-0017-4734-be04-9c649f761d27","Type":"ContainerDied","Data":"063b61665c2ee996631b73afe9b0f055c2b932e6268e0772639d782937225569"} Feb 25 10:26:06 crc kubenswrapper[4978]: I0225 10:26:06.098493 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533586-c6l5k" Feb 25 10:26:06 crc kubenswrapper[4978]: I0225 10:26:06.157576 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmwfc\" (UniqueName: \"kubernetes.io/projected/6da167fe-0017-4734-be04-9c649f761d27-kube-api-access-mmwfc\") pod \"6da167fe-0017-4734-be04-9c649f761d27\" (UID: \"6da167fe-0017-4734-be04-9c649f761d27\") " Feb 25 10:26:06 crc kubenswrapper[4978]: I0225 10:26:06.165478 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6da167fe-0017-4734-be04-9c649f761d27-kube-api-access-mmwfc" (OuterVolumeSpecName: "kube-api-access-mmwfc") pod "6da167fe-0017-4734-be04-9c649f761d27" (UID: "6da167fe-0017-4734-be04-9c649f761d27"). InnerVolumeSpecName "kube-api-access-mmwfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:26:06 crc kubenswrapper[4978]: I0225 10:26:06.260859 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmwfc\" (UniqueName: \"kubernetes.io/projected/6da167fe-0017-4734-be04-9c649f761d27-kube-api-access-mmwfc\") on node \"crc\" DevicePath \"\"" Feb 25 10:26:06 crc kubenswrapper[4978]: I0225 10:26:06.704331 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533586-c6l5k" event={"ID":"6da167fe-0017-4734-be04-9c649f761d27","Type":"ContainerDied","Data":"0598f782617be58d5bbca7cd6211ed9a9f1c8156c63b1248abaa0d2490bb6a0c"} Feb 25 10:26:06 crc kubenswrapper[4978]: I0225 10:26:06.704391 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0598f782617be58d5bbca7cd6211ed9a9f1c8156c63b1248abaa0d2490bb6a0c" Feb 25 10:26:06 crc kubenswrapper[4978]: I0225 10:26:06.704444 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533586-c6l5k" Feb 25 10:26:07 crc kubenswrapper[4978]: I0225 10:26:07.180941 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533580-4f8k8"] Feb 25 10:26:07 crc kubenswrapper[4978]: I0225 10:26:07.193272 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533580-4f8k8"] Feb 25 10:26:07 crc kubenswrapper[4978]: I0225 10:26:07.343105 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d4aee5d-a45a-4672-ae14-e1af0287f17d" path="/var/lib/kubelet/pods/6d4aee5d-a45a-4672-ae14-e1af0287f17d/volumes" Feb 25 10:26:44 crc kubenswrapper[4978]: I0225 10:26:44.836599 4978 scope.go:117] "RemoveContainer" containerID="91092ca3188fe8ae4937f7f257581369ac360c51c531b685ae629e814456c6ba" Feb 25 10:26:46 crc kubenswrapper[4978]: I0225 10:26:46.540251 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:26:46 crc kubenswrapper[4978]: I0225 10:26:46.540711 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:27:16 crc kubenswrapper[4978]: I0225 10:27:16.540443 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:27:16 crc kubenswrapper[4978]: I0225 10:27:16.541136 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:27:46 crc kubenswrapper[4978]: I0225 10:27:46.540538 4978 patch_prober.go:28] interesting pod/machine-config-daemon-j496h container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 25 10:27:46 crc kubenswrapper[4978]: I0225 10:27:46.541234 4978 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 25 10:27:46 crc kubenswrapper[4978]: I0225 10:27:46.541293 4978 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-j496h" Feb 25 10:27:46 crc kubenswrapper[4978]: I0225 10:27:46.542092 4978 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988"} pod="openshift-machine-config-operator/machine-config-daemon-j496h" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 25 10:27:46 crc kubenswrapper[4978]: I0225 10:27:46.542152 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerName="machine-config-daemon" containerID="cri-o://45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" gracePeriod=600 Feb 25 10:27:46 crc kubenswrapper[4978]: E0225 10:27:46.722492 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:27:47 crc kubenswrapper[4978]: I0225 10:27:47.478726 4978 generic.go:334] "Generic (PLEG): container finished" podID="a5f01015-5dea-4e59-a9fc-326c84b85aed" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" exitCode=0 Feb 25 10:27:47 crc kubenswrapper[4978]: I0225 10:27:47.478817 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerDied","Data":"45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988"} Feb 25 10:27:47 crc kubenswrapper[4978]: I0225 10:27:47.479089 4978 scope.go:117] "RemoveContainer" containerID="95899c3bd9197bd5e205d8ad51463abe0d611cd768f6460f30bac2b6f6450c64" Feb 25 10:27:47 crc kubenswrapper[4978]: I0225 10:27:47.479965 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:27:47 crc kubenswrapper[4978]: E0225 10:27:47.480304 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:27:58 crc kubenswrapper[4978]: I0225 10:27:58.328764 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:27:58 crc kubenswrapper[4978]: E0225 10:27:58.329975 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.152563 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533588-f29nr"] Feb 25 10:28:00 crc kubenswrapper[4978]: E0225 10:28:00.153176 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6da167fe-0017-4734-be04-9c649f761d27" containerName="oc" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.153191 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6da167fe-0017-4734-be04-9c649f761d27" containerName="oc" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.153471 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6da167fe-0017-4734-be04-9c649f761d27" containerName="oc" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.154420 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533588-f29nr" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.158720 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.159077 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.161767 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.171139 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533588-f29nr"] Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.196991 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r55ph\" (UniqueName: \"kubernetes.io/projected/f48904f9-daf5-43c5-99a4-b1ac77dd9da9-kube-api-access-r55ph\") pod \"auto-csr-approver-29533588-f29nr\" (UID: \"f48904f9-daf5-43c5-99a4-b1ac77dd9da9\") " pod="openshift-infra/auto-csr-approver-29533588-f29nr" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.299223 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r55ph\" (UniqueName: \"kubernetes.io/projected/f48904f9-daf5-43c5-99a4-b1ac77dd9da9-kube-api-access-r55ph\") pod \"auto-csr-approver-29533588-f29nr\" (UID: \"f48904f9-daf5-43c5-99a4-b1ac77dd9da9\") " pod="openshift-infra/auto-csr-approver-29533588-f29nr" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.324310 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r55ph\" (UniqueName: \"kubernetes.io/projected/f48904f9-daf5-43c5-99a4-b1ac77dd9da9-kube-api-access-r55ph\") pod \"auto-csr-approver-29533588-f29nr\" (UID: \"f48904f9-daf5-43c5-99a4-b1ac77dd9da9\") " pod="openshift-infra/auto-csr-approver-29533588-f29nr" Feb 25 10:28:00 crc kubenswrapper[4978]: I0225 10:28:00.480068 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533588-f29nr" Feb 25 10:28:01 crc kubenswrapper[4978]: I0225 10:28:01.401928 4978 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 25 10:28:01 crc kubenswrapper[4978]: I0225 10:28:01.407970 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533588-f29nr"] Feb 25 10:28:01 crc kubenswrapper[4978]: I0225 10:28:01.641206 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533588-f29nr" event={"ID":"f48904f9-daf5-43c5-99a4-b1ac77dd9da9","Type":"ContainerStarted","Data":"66afc0ff713a3223403eed49c0d799d620618e10ba1f6e0cc4a0435842db6ec3"} Feb 25 10:28:02 crc kubenswrapper[4978]: I0225 10:28:02.651584 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533588-f29nr" event={"ID":"f48904f9-daf5-43c5-99a4-b1ac77dd9da9","Type":"ContainerStarted","Data":"6bde6f9159a1ee2471ae20e33167af7abf062170584a85412dcaaff6e4ce2ea5"} Feb 25 10:28:02 crc kubenswrapper[4978]: I0225 10:28:02.669298 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29533588-f29nr" podStartSLOduration=1.796240567 podStartE2EDuration="2.669277566s" podCreationTimestamp="2026-02-25 10:28:00 +0000 UTC" firstStartedPulling="2026-02-25 10:28:01.401682096 +0000 UTC m=+13374.840938555" lastFinishedPulling="2026-02-25 10:28:02.274719095 +0000 UTC m=+13375.713975554" observedRunningTime="2026-02-25 10:28:02.66234169 +0000 UTC m=+13376.101598149" watchObservedRunningTime="2026-02-25 10:28:02.669277566 +0000 UTC m=+13376.108534025" Feb 25 10:28:03 crc kubenswrapper[4978]: I0225 10:28:03.662975 4978 generic.go:334] "Generic (PLEG): container finished" podID="f48904f9-daf5-43c5-99a4-b1ac77dd9da9" containerID="6bde6f9159a1ee2471ae20e33167af7abf062170584a85412dcaaff6e4ce2ea5" exitCode=0 Feb 25 10:28:03 crc kubenswrapper[4978]: I0225 10:28:03.663025 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533588-f29nr" event={"ID":"f48904f9-daf5-43c5-99a4-b1ac77dd9da9","Type":"ContainerDied","Data":"6bde6f9159a1ee2471ae20e33167af7abf062170584a85412dcaaff6e4ce2ea5"} Feb 25 10:28:06 crc kubenswrapper[4978]: I0225 10:28:06.331712 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533588-f29nr" Feb 25 10:28:06 crc kubenswrapper[4978]: I0225 10:28:06.486055 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r55ph\" (UniqueName: \"kubernetes.io/projected/f48904f9-daf5-43c5-99a4-b1ac77dd9da9-kube-api-access-r55ph\") pod \"f48904f9-daf5-43c5-99a4-b1ac77dd9da9\" (UID: \"f48904f9-daf5-43c5-99a4-b1ac77dd9da9\") " Feb 25 10:28:06 crc kubenswrapper[4978]: I0225 10:28:06.497145 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f48904f9-daf5-43c5-99a4-b1ac77dd9da9-kube-api-access-r55ph" (OuterVolumeSpecName: "kube-api-access-r55ph") pod "f48904f9-daf5-43c5-99a4-b1ac77dd9da9" (UID: "f48904f9-daf5-43c5-99a4-b1ac77dd9da9"). InnerVolumeSpecName "kube-api-access-r55ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:28:06 crc kubenswrapper[4978]: I0225 10:28:06.588638 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r55ph\" (UniqueName: \"kubernetes.io/projected/f48904f9-daf5-43c5-99a4-b1ac77dd9da9-kube-api-access-r55ph\") on node \"crc\" DevicePath \"\"" Feb 25 10:28:06 crc kubenswrapper[4978]: I0225 10:28:06.699941 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533588-f29nr" event={"ID":"f48904f9-daf5-43c5-99a4-b1ac77dd9da9","Type":"ContainerDied","Data":"66afc0ff713a3223403eed49c0d799d620618e10ba1f6e0cc4a0435842db6ec3"} Feb 25 10:28:06 crc kubenswrapper[4978]: I0225 10:28:06.699979 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66afc0ff713a3223403eed49c0d799d620618e10ba1f6e0cc4a0435842db6ec3" Feb 25 10:28:06 crc kubenswrapper[4978]: I0225 10:28:06.700002 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533588-f29nr" Feb 25 10:28:07 crc kubenswrapper[4978]: I0225 10:28:07.408155 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533582-2bdwl"] Feb 25 10:28:07 crc kubenswrapper[4978]: I0225 10:28:07.436476 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533582-2bdwl"] Feb 25 10:28:09 crc kubenswrapper[4978]: I0225 10:28:09.340617 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b28b16ff-594c-4da7-aa74-e977658b02c0" path="/var/lib/kubelet/pods/b28b16ff-594c-4da7-aa74-e977658b02c0/volumes" Feb 25 10:28:11 crc kubenswrapper[4978]: I0225 10:28:11.327840 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:28:11 crc kubenswrapper[4978]: E0225 10:28:11.328464 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:28:24 crc kubenswrapper[4978]: I0225 10:28:24.327975 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:28:24 crc kubenswrapper[4978]: E0225 10:28:24.328623 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:28:35 crc kubenswrapper[4978]: I0225 10:28:35.328533 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:28:35 crc kubenswrapper[4978]: E0225 10:28:35.329253 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:28:44 crc kubenswrapper[4978]: I0225 10:28:44.994685 4978 scope.go:117] "RemoveContainer" containerID="b04f7b7becce137d277327e23bb6eda7b020ba22a1873c08a5648747300169b8" Feb 25 10:28:48 crc kubenswrapper[4978]: I0225 10:28:48.328538 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:28:48 crc kubenswrapper[4978]: E0225 10:28:48.329489 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:29:00 crc kubenswrapper[4978]: I0225 10:29:00.327494 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:29:00 crc kubenswrapper[4978]: E0225 10:29:00.332308 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:29:12 crc kubenswrapper[4978]: I0225 10:29:12.328113 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:29:12 crc kubenswrapper[4978]: E0225 10:29:12.328750 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:29:13 crc kubenswrapper[4978]: I0225 10:29:13.509438 4978 generic.go:334] "Generic (PLEG): container finished" podID="6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" containerID="a1f38cec30df06dca5c749dd908e691af42f62bf795536d8c4c0f1870be1cc2f" exitCode=0 Feb 25 10:29:13 crc kubenswrapper[4978]: I0225 10:29:13.509484 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pv922/must-gather-7jwqp" event={"ID":"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3","Type":"ContainerDied","Data":"a1f38cec30df06dca5c749dd908e691af42f62bf795536d8c4c0f1870be1cc2f"} Feb 25 10:29:13 crc kubenswrapper[4978]: I0225 10:29:13.509905 4978 scope.go:117] "RemoveContainer" containerID="a1f38cec30df06dca5c749dd908e691af42f62bf795536d8c4c0f1870be1cc2f" Feb 25 10:29:13 crc kubenswrapper[4978]: I0225 10:29:13.914333 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pv922_must-gather-7jwqp_6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3/gather/0.log" Feb 25 10:29:25 crc kubenswrapper[4978]: I0225 10:29:25.964924 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pv922/must-gather-7jwqp"] Feb 25 10:29:25 crc kubenswrapper[4978]: I0225 10:29:25.965994 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-pv922/must-gather-7jwqp" podUID="6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" containerName="copy" containerID="cri-o://17fe87146b908d0c834c610b28e9febcb8fda4c87e32a8ce509dc56de3a340ee" gracePeriod=2 Feb 25 10:29:25 crc kubenswrapper[4978]: I0225 10:29:25.981313 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pv922/must-gather-7jwqp"] Feb 25 10:29:26 crc kubenswrapper[4978]: I0225 10:29:26.327797 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:29:26 crc kubenswrapper[4978]: E0225 10:29:26.328410 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:29:26 crc kubenswrapper[4978]: I0225 10:29:26.661639 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pv922_must-gather-7jwqp_6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3/copy/0.log" Feb 25 10:29:26 crc kubenswrapper[4978]: I0225 10:29:26.662035 4978 generic.go:334] "Generic (PLEG): container finished" podID="6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" containerID="17fe87146b908d0c834c610b28e9febcb8fda4c87e32a8ce509dc56de3a340ee" exitCode=143 Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.452155 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pv922_must-gather-7jwqp_6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3/copy/0.log" Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.452801 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/must-gather-7jwqp" Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.474795 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-must-gather-output\") pod \"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3\" (UID: \"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3\") " Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.475093 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq4kl\" (UniqueName: \"kubernetes.io/projected/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-kube-api-access-cq4kl\") pod \"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3\" (UID: \"6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3\") " Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.481603 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-kube-api-access-cq4kl" (OuterVolumeSpecName: "kube-api-access-cq4kl") pod "6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" (UID: "6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3"). InnerVolumeSpecName "kube-api-access-cq4kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.579012 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq4kl\" (UniqueName: \"kubernetes.io/projected/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-kube-api-access-cq4kl\") on node \"crc\" DevicePath \"\"" Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.687822 4978 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pv922_must-gather-7jwqp_6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3/copy/0.log" Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.688148 4978 scope.go:117] "RemoveContainer" containerID="17fe87146b908d0c834c610b28e9febcb8fda4c87e32a8ce509dc56de3a340ee" Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.688306 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pv922/must-gather-7jwqp" Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.719739 4978 scope.go:117] "RemoveContainer" containerID="a1f38cec30df06dca5c749dd908e691af42f62bf795536d8c4c0f1870be1cc2f" Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.762015 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" (UID: "6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:29:27 crc kubenswrapper[4978]: I0225 10:29:27.786210 4978 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 25 10:29:29 crc kubenswrapper[4978]: I0225 10:29:29.366756 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" path="/var/lib/kubelet/pods/6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3/volumes" Feb 25 10:29:39 crc kubenswrapper[4978]: I0225 10:29:39.327714 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:29:39 crc kubenswrapper[4978]: E0225 10:29:39.328394 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:29:54 crc kubenswrapper[4978]: I0225 10:29:54.328258 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:29:54 crc kubenswrapper[4978]: E0225 10:29:54.329267 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.158458 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533590-cmrnw"] Feb 25 10:30:00 crc kubenswrapper[4978]: E0225 10:30:00.159126 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" containerName="copy" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.159136 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" containerName="copy" Feb 25 10:30:00 crc kubenswrapper[4978]: E0225 10:30:00.159152 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48904f9-daf5-43c5-99a4-b1ac77dd9da9" containerName="oc" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.159159 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48904f9-daf5-43c5-99a4-b1ac77dd9da9" containerName="oc" Feb 25 10:30:00 crc kubenswrapper[4978]: E0225 10:30:00.159179 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" containerName="gather" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.159185 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" containerName="gather" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.159417 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" containerName="copy" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.159439 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d32ba84-d8e7-4ac6-89f3-f037f5b81ee3" containerName="gather" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.159459 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="f48904f9-daf5-43c5-99a4-b1ac77dd9da9" containerName="oc" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.160121 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533590-cmrnw" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.162139 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.162776 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.163485 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.179217 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533590-cmrnw"] Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.196309 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg"] Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.198450 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.200361 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.200973 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.222275 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg"] Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.303396 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a98a49cf-e3dc-4631-886a-ed61e3f6501e-secret-volume\") pod \"collect-profiles-29533590-tjtmg\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.303540 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a98a49cf-e3dc-4631-886a-ed61e3f6501e-config-volume\") pod \"collect-profiles-29533590-tjtmg\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.303603 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4tsm\" (UniqueName: \"kubernetes.io/projected/a835415d-bf8b-4ad8-b223-1249cfbc3b51-kube-api-access-p4tsm\") pod \"auto-csr-approver-29533590-cmrnw\" (UID: \"a835415d-bf8b-4ad8-b223-1249cfbc3b51\") " pod="openshift-infra/auto-csr-approver-29533590-cmrnw" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.304065 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfzm4\" (UniqueName: \"kubernetes.io/projected/a98a49cf-e3dc-4631-886a-ed61e3f6501e-kube-api-access-dfzm4\") pod \"collect-profiles-29533590-tjtmg\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.406405 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfzm4\" (UniqueName: \"kubernetes.io/projected/a98a49cf-e3dc-4631-886a-ed61e3f6501e-kube-api-access-dfzm4\") pod \"collect-profiles-29533590-tjtmg\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.406483 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a98a49cf-e3dc-4631-886a-ed61e3f6501e-secret-volume\") pod \"collect-profiles-29533590-tjtmg\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.406552 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a98a49cf-e3dc-4631-886a-ed61e3f6501e-config-volume\") pod \"collect-profiles-29533590-tjtmg\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.406584 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4tsm\" (UniqueName: \"kubernetes.io/projected/a835415d-bf8b-4ad8-b223-1249cfbc3b51-kube-api-access-p4tsm\") pod \"auto-csr-approver-29533590-cmrnw\" (UID: \"a835415d-bf8b-4ad8-b223-1249cfbc3b51\") " pod="openshift-infra/auto-csr-approver-29533590-cmrnw" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.407383 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a98a49cf-e3dc-4631-886a-ed61e3f6501e-config-volume\") pod \"collect-profiles-29533590-tjtmg\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.417355 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a98a49cf-e3dc-4631-886a-ed61e3f6501e-secret-volume\") pod \"collect-profiles-29533590-tjtmg\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.424143 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfzm4\" (UniqueName: \"kubernetes.io/projected/a98a49cf-e3dc-4631-886a-ed61e3f6501e-kube-api-access-dfzm4\") pod \"collect-profiles-29533590-tjtmg\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.426040 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4tsm\" (UniqueName: \"kubernetes.io/projected/a835415d-bf8b-4ad8-b223-1249cfbc3b51-kube-api-access-p4tsm\") pod \"auto-csr-approver-29533590-cmrnw\" (UID: \"a835415d-bf8b-4ad8-b223-1249cfbc3b51\") " pod="openshift-infra/auto-csr-approver-29533590-cmrnw" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.482710 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533590-cmrnw" Feb 25 10:30:00 crc kubenswrapper[4978]: I0225 10:30:00.520901 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:01 crc kubenswrapper[4978]: I0225 10:30:01.243292 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533590-cmrnw"] Feb 25 10:30:01 crc kubenswrapper[4978]: I0225 10:30:01.702985 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg"] Feb 25 10:30:02 crc kubenswrapper[4978]: I0225 10:30:02.104216 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533590-cmrnw" event={"ID":"a835415d-bf8b-4ad8-b223-1249cfbc3b51","Type":"ContainerStarted","Data":"5216f5776ccf9a1407ff1b375ba01ab9d4749e00f06fa34f371957eccd588a1f"} Feb 25 10:30:02 crc kubenswrapper[4978]: I0225 10:30:02.106270 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" event={"ID":"a98a49cf-e3dc-4631-886a-ed61e3f6501e","Type":"ContainerStarted","Data":"6ed16e122d05eac9f73940c286f13f54b0fb3cb0337938637a71e91760a4a86f"} Feb 25 10:30:02 crc kubenswrapper[4978]: I0225 10:30:02.106319 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" event={"ID":"a98a49cf-e3dc-4631-886a-ed61e3f6501e","Type":"ContainerStarted","Data":"83a6b17d06fe63c1c246695ad394a98d071fde7f35e1b1f2cc4d8ec1fecfc9d9"} Feb 25 10:30:02 crc kubenswrapper[4978]: I0225 10:30:02.122688 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" podStartSLOduration=2.122668459 podStartE2EDuration="2.122668459s" podCreationTimestamp="2026-02-25 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-25 10:30:02.121875354 +0000 UTC m=+13495.561131823" watchObservedRunningTime="2026-02-25 10:30:02.122668459 +0000 UTC m=+13495.561924918" Feb 25 10:30:03 crc kubenswrapper[4978]: I0225 10:30:03.119594 4978 generic.go:334] "Generic (PLEG): container finished" podID="a98a49cf-e3dc-4631-886a-ed61e3f6501e" containerID="6ed16e122d05eac9f73940c286f13f54b0fb3cb0337938637a71e91760a4a86f" exitCode=0 Feb 25 10:30:03 crc kubenswrapper[4978]: I0225 10:30:03.119743 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" event={"ID":"a98a49cf-e3dc-4631-886a-ed61e3f6501e","Type":"ContainerDied","Data":"6ed16e122d05eac9f73940c286f13f54b0fb3cb0337938637a71e91760a4a86f"} Feb 25 10:30:04 crc kubenswrapper[4978]: I0225 10:30:04.133859 4978 generic.go:334] "Generic (PLEG): container finished" podID="a835415d-bf8b-4ad8-b223-1249cfbc3b51" containerID="330b467c5dd082c68d41aa1227c6f61a1064caa2814dc766cb421feda52af341" exitCode=0 Feb 25 10:30:04 crc kubenswrapper[4978]: I0225 10:30:04.133960 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533590-cmrnw" event={"ID":"a835415d-bf8b-4ad8-b223-1249cfbc3b51","Type":"ContainerDied","Data":"330b467c5dd082c68d41aa1227c6f61a1064caa2814dc766cb421feda52af341"} Feb 25 10:30:05 crc kubenswrapper[4978]: I0225 10:30:05.614100 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:05 crc kubenswrapper[4978]: I0225 10:30:05.726578 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a98a49cf-e3dc-4631-886a-ed61e3f6501e-config-volume\") pod \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " Feb 25 10:30:05 crc kubenswrapper[4978]: I0225 10:30:05.726860 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a98a49cf-e3dc-4631-886a-ed61e3f6501e-secret-volume\") pod \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " Feb 25 10:30:05 crc kubenswrapper[4978]: I0225 10:30:05.726896 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfzm4\" (UniqueName: \"kubernetes.io/projected/a98a49cf-e3dc-4631-886a-ed61e3f6501e-kube-api-access-dfzm4\") pod \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\" (UID: \"a98a49cf-e3dc-4631-886a-ed61e3f6501e\") " Feb 25 10:30:05 crc kubenswrapper[4978]: I0225 10:30:05.727138 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a98a49cf-e3dc-4631-886a-ed61e3f6501e-config-volume" (OuterVolumeSpecName: "config-volume") pod "a98a49cf-e3dc-4631-886a-ed61e3f6501e" (UID: "a98a49cf-e3dc-4631-886a-ed61e3f6501e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 25 10:30:05 crc kubenswrapper[4978]: I0225 10:30:05.728689 4978 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a98a49cf-e3dc-4631-886a-ed61e3f6501e-config-volume\") on node \"crc\" DevicePath \"\"" Feb 25 10:30:05 crc kubenswrapper[4978]: I0225 10:30:05.794547 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a98a49cf-e3dc-4631-886a-ed61e3f6501e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a98a49cf-e3dc-4631-886a-ed61e3f6501e" (UID: "a98a49cf-e3dc-4631-886a-ed61e3f6501e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 25 10:30:05 crc kubenswrapper[4978]: I0225 10:30:05.794960 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a98a49cf-e3dc-4631-886a-ed61e3f6501e-kube-api-access-dfzm4" (OuterVolumeSpecName: "kube-api-access-dfzm4") pod "a98a49cf-e3dc-4631-886a-ed61e3f6501e" (UID: "a98a49cf-e3dc-4631-886a-ed61e3f6501e"). InnerVolumeSpecName "kube-api-access-dfzm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:30:05 crc kubenswrapper[4978]: I0225 10:30:05.831203 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfzm4\" (UniqueName: \"kubernetes.io/projected/a98a49cf-e3dc-4631-886a-ed61e3f6501e-kube-api-access-dfzm4\") on node \"crc\" DevicePath \"\"" Feb 25 10:30:05 crc kubenswrapper[4978]: I0225 10:30:05.831247 4978 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a98a49cf-e3dc-4631-886a-ed61e3f6501e-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 25 10:30:06 crc kubenswrapper[4978]: I0225 10:30:06.155857 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" event={"ID":"a98a49cf-e3dc-4631-886a-ed61e3f6501e","Type":"ContainerDied","Data":"83a6b17d06fe63c1c246695ad394a98d071fde7f35e1b1f2cc4d8ec1fecfc9d9"} Feb 25 10:30:06 crc kubenswrapper[4978]: I0225 10:30:06.155899 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83a6b17d06fe63c1c246695ad394a98d071fde7f35e1b1f2cc4d8ec1fecfc9d9" Feb 25 10:30:06 crc kubenswrapper[4978]: I0225 10:30:06.155957 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29533590-tjtmg" Feb 25 10:30:06 crc kubenswrapper[4978]: I0225 10:30:06.327562 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:30:06 crc kubenswrapper[4978]: E0225 10:30:06.328079 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:30:06 crc kubenswrapper[4978]: I0225 10:30:06.698834 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533590-cmrnw" Feb 25 10:30:06 crc kubenswrapper[4978]: I0225 10:30:06.718521 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs"] Feb 25 10:30:06 crc kubenswrapper[4978]: I0225 10:30:06.729996 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29533545-s74zs"] Feb 25 10:30:06 crc kubenswrapper[4978]: I0225 10:30:06.851450 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4tsm\" (UniqueName: \"kubernetes.io/projected/a835415d-bf8b-4ad8-b223-1249cfbc3b51-kube-api-access-p4tsm\") pod \"a835415d-bf8b-4ad8-b223-1249cfbc3b51\" (UID: \"a835415d-bf8b-4ad8-b223-1249cfbc3b51\") " Feb 25 10:30:06 crc kubenswrapper[4978]: I0225 10:30:06.859200 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a835415d-bf8b-4ad8-b223-1249cfbc3b51-kube-api-access-p4tsm" (OuterVolumeSpecName: "kube-api-access-p4tsm") pod "a835415d-bf8b-4ad8-b223-1249cfbc3b51" (UID: "a835415d-bf8b-4ad8-b223-1249cfbc3b51"). InnerVolumeSpecName "kube-api-access-p4tsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:30:06 crc kubenswrapper[4978]: I0225 10:30:06.953157 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4tsm\" (UniqueName: \"kubernetes.io/projected/a835415d-bf8b-4ad8-b223-1249cfbc3b51-kube-api-access-p4tsm\") on node \"crc\" DevicePath \"\"" Feb 25 10:30:07 crc kubenswrapper[4978]: I0225 10:30:07.166939 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533590-cmrnw" event={"ID":"a835415d-bf8b-4ad8-b223-1249cfbc3b51","Type":"ContainerDied","Data":"5216f5776ccf9a1407ff1b375ba01ab9d4749e00f06fa34f371957eccd588a1f"} Feb 25 10:30:07 crc kubenswrapper[4978]: I0225 10:30:07.166976 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5216f5776ccf9a1407ff1b375ba01ab9d4749e00f06fa34f371957eccd588a1f" Feb 25 10:30:07 crc kubenswrapper[4978]: I0225 10:30:07.167037 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533590-cmrnw" Feb 25 10:30:07 crc kubenswrapper[4978]: I0225 10:30:07.343957 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64b5569b-078c-45d8-89c9-8a4f32e2aad7" path="/var/lib/kubelet/pods/64b5569b-078c-45d8-89c9-8a4f32e2aad7/volumes" Feb 25 10:30:07 crc kubenswrapper[4978]: I0225 10:30:07.760180 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533584-8z88t"] Feb 25 10:30:07 crc kubenswrapper[4978]: I0225 10:30:07.770619 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533584-8z88t"] Feb 25 10:30:09 crc kubenswrapper[4978]: I0225 10:30:09.347284 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b94811-1a8c-419a-8101-ba50b4dad2a5" path="/var/lib/kubelet/pods/b5b94811-1a8c-419a-8101-ba50b4dad2a5/volumes" Feb 25 10:30:21 crc kubenswrapper[4978]: I0225 10:30:21.327443 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:30:21 crc kubenswrapper[4978]: E0225 10:30:21.328164 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:30:36 crc kubenswrapper[4978]: I0225 10:30:36.328219 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:30:36 crc kubenswrapper[4978]: E0225 10:30:36.329029 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:30:37 crc kubenswrapper[4978]: I0225 10:30:37.916786 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vx842"] Feb 25 10:30:37 crc kubenswrapper[4978]: E0225 10:30:37.917252 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a98a49cf-e3dc-4631-886a-ed61e3f6501e" containerName="collect-profiles" Feb 25 10:30:37 crc kubenswrapper[4978]: I0225 10:30:37.917265 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a98a49cf-e3dc-4631-886a-ed61e3f6501e" containerName="collect-profiles" Feb 25 10:30:37 crc kubenswrapper[4978]: E0225 10:30:37.917283 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a835415d-bf8b-4ad8-b223-1249cfbc3b51" containerName="oc" Feb 25 10:30:37 crc kubenswrapper[4978]: I0225 10:30:37.917289 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="a835415d-bf8b-4ad8-b223-1249cfbc3b51" containerName="oc" Feb 25 10:30:37 crc kubenswrapper[4978]: I0225 10:30:37.917524 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a98a49cf-e3dc-4631-886a-ed61e3f6501e" containerName="collect-profiles" Feb 25 10:30:37 crc kubenswrapper[4978]: I0225 10:30:37.917543 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="a835415d-bf8b-4ad8-b223-1249cfbc3b51" containerName="oc" Feb 25 10:30:37 crc kubenswrapper[4978]: I0225 10:30:37.918935 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:37 crc kubenswrapper[4978]: I0225 10:30:37.929362 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vx842"] Feb 25 10:30:38 crc kubenswrapper[4978]: I0225 10:30:38.106021 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-catalog-content\") pod \"certified-operators-vx842\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:38 crc kubenswrapper[4978]: I0225 10:30:38.106226 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-utilities\") pod \"certified-operators-vx842\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:38 crc kubenswrapper[4978]: I0225 10:30:38.106273 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg6qn\" (UniqueName: \"kubernetes.io/projected/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-kube-api-access-hg6qn\") pod \"certified-operators-vx842\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:38 crc kubenswrapper[4978]: I0225 10:30:38.208123 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-utilities\") pod \"certified-operators-vx842\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:38 crc kubenswrapper[4978]: I0225 10:30:38.208189 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg6qn\" (UniqueName: \"kubernetes.io/projected/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-kube-api-access-hg6qn\") pod \"certified-operators-vx842\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:38 crc kubenswrapper[4978]: I0225 10:30:38.208227 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-catalog-content\") pod \"certified-operators-vx842\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:38 crc kubenswrapper[4978]: I0225 10:30:38.208696 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-utilities\") pod \"certified-operators-vx842\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:38 crc kubenswrapper[4978]: I0225 10:30:38.208711 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-catalog-content\") pod \"certified-operators-vx842\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:38 crc kubenswrapper[4978]: I0225 10:30:38.236028 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg6qn\" (UniqueName: \"kubernetes.io/projected/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-kube-api-access-hg6qn\") pod \"certified-operators-vx842\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:38 crc kubenswrapper[4978]: I0225 10:30:38.240232 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:39 crc kubenswrapper[4978]: I0225 10:30:39.060544 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vx842"] Feb 25 10:30:39 crc kubenswrapper[4978]: I0225 10:30:39.547508 4978 generic.go:334] "Generic (PLEG): container finished" podID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerID="bc84361e7acbe7ef42d0c037aedf5263a8f70808fbd502db44113c2431ae4d79" exitCode=0 Feb 25 10:30:39 crc kubenswrapper[4978]: I0225 10:30:39.547596 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vx842" event={"ID":"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0","Type":"ContainerDied","Data":"bc84361e7acbe7ef42d0c037aedf5263a8f70808fbd502db44113c2431ae4d79"} Feb 25 10:30:39 crc kubenswrapper[4978]: I0225 10:30:39.548206 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vx842" event={"ID":"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0","Type":"ContainerStarted","Data":"8f0786fd9e57ab19a55ed8e0a0c3675d44aadb2df3d96da8dbd2d2f88b26c7de"} Feb 25 10:30:40 crc kubenswrapper[4978]: I0225 10:30:40.561824 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vx842" event={"ID":"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0","Type":"ContainerStarted","Data":"1056ca2a345418fd5bc8ddf9aa68669fec4dd75378086e5c207cb8f95ab40905"} Feb 25 10:30:41 crc kubenswrapper[4978]: E0225 10:30:41.864079 4978 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c80d668_ae70_4010_a8a0_c3a74e0aa8b0.slice/crio-conmon-1056ca2a345418fd5bc8ddf9aa68669fec4dd75378086e5c207cb8f95ab40905.scope\": RecentStats: unable to find data in memory cache]" Feb 25 10:30:42 crc kubenswrapper[4978]: I0225 10:30:42.583354 4978 generic.go:334] "Generic (PLEG): container finished" podID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerID="1056ca2a345418fd5bc8ddf9aa68669fec4dd75378086e5c207cb8f95ab40905" exitCode=0 Feb 25 10:30:42 crc kubenswrapper[4978]: I0225 10:30:42.583401 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vx842" event={"ID":"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0","Type":"ContainerDied","Data":"1056ca2a345418fd5bc8ddf9aa68669fec4dd75378086e5c207cb8f95ab40905"} Feb 25 10:30:43 crc kubenswrapper[4978]: I0225 10:30:43.595107 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vx842" event={"ID":"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0","Type":"ContainerStarted","Data":"e7b34c722a972ceb6b30440ad3e6d1eb7a1e5fb84455adef1dac54be7c768bf9"} Feb 25 10:30:43 crc kubenswrapper[4978]: I0225 10:30:43.631197 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vx842" podStartSLOduration=3.203662385 podStartE2EDuration="6.631175593s" podCreationTimestamp="2026-02-25 10:30:37 +0000 UTC" firstStartedPulling="2026-02-25 10:30:39.549525981 +0000 UTC m=+13532.988782440" lastFinishedPulling="2026-02-25 10:30:42.977039169 +0000 UTC m=+13536.416295648" observedRunningTime="2026-02-25 10:30:43.621252904 +0000 UTC m=+13537.060509403" watchObservedRunningTime="2026-02-25 10:30:43.631175593 +0000 UTC m=+13537.070432062" Feb 25 10:30:45 crc kubenswrapper[4978]: I0225 10:30:45.142228 4978 scope.go:117] "RemoveContainer" containerID="fe1e70f5d9d044846b69ebe72e4ab4be5581c19762b8fbabb2b7bc5882d627f5" Feb 25 10:30:45 crc kubenswrapper[4978]: I0225 10:30:45.188089 4978 scope.go:117] "RemoveContainer" containerID="9ea691c0ccee30301c453c41eacc320eaee2233143b7e11d76b1408816e9ca9a" Feb 25 10:30:47 crc kubenswrapper[4978]: I0225 10:30:47.338264 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:30:47 crc kubenswrapper[4978]: E0225 10:30:47.338911 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:30:48 crc kubenswrapper[4978]: I0225 10:30:48.241149 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:48 crc kubenswrapper[4978]: I0225 10:30:48.241646 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:48 crc kubenswrapper[4978]: I0225 10:30:48.307236 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:48 crc kubenswrapper[4978]: I0225 10:30:48.729432 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:48 crc kubenswrapper[4978]: I0225 10:30:48.797286 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vx842"] Feb 25 10:30:50 crc kubenswrapper[4978]: I0225 10:30:50.670255 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vx842" podUID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerName="registry-server" containerID="cri-o://e7b34c722a972ceb6b30440ad3e6d1eb7a1e5fb84455adef1dac54be7c768bf9" gracePeriod=2 Feb 25 10:30:51 crc kubenswrapper[4978]: I0225 10:30:51.684888 4978 generic.go:334] "Generic (PLEG): container finished" podID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerID="e7b34c722a972ceb6b30440ad3e6d1eb7a1e5fb84455adef1dac54be7c768bf9" exitCode=0 Feb 25 10:30:51 crc kubenswrapper[4978]: I0225 10:30:51.685060 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vx842" event={"ID":"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0","Type":"ContainerDied","Data":"e7b34c722a972ceb6b30440ad3e6d1eb7a1e5fb84455adef1dac54be7c768bf9"} Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.395295 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.462321 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-catalog-content\") pod \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.462403 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg6qn\" (UniqueName: \"kubernetes.io/projected/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-kube-api-access-hg6qn\") pod \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.462567 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-utilities\") pod \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\" (UID: \"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0\") " Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.463838 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-utilities" (OuterVolumeSpecName: "utilities") pod "9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" (UID: "9c80d668-ae70-4010-a8a0-c3a74e0aa8b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.477263 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-kube-api-access-hg6qn" (OuterVolumeSpecName: "kube-api-access-hg6qn") pod "9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" (UID: "9c80d668-ae70-4010-a8a0-c3a74e0aa8b0"). InnerVolumeSpecName "kube-api-access-hg6qn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.542253 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" (UID: "9c80d668-ae70-4010-a8a0-c3a74e0aa8b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.565649 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.565685 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg6qn\" (UniqueName: \"kubernetes.io/projected/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-kube-api-access-hg6qn\") on node \"crc\" DevicePath \"\"" Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.565695 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.697890 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vx842" event={"ID":"9c80d668-ae70-4010-a8a0-c3a74e0aa8b0","Type":"ContainerDied","Data":"8f0786fd9e57ab19a55ed8e0a0c3675d44aadb2df3d96da8dbd2d2f88b26c7de"} Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.697947 4978 scope.go:117] "RemoveContainer" containerID="e7b34c722a972ceb6b30440ad3e6d1eb7a1e5fb84455adef1dac54be7c768bf9" Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.697962 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vx842" Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.739136 4978 scope.go:117] "RemoveContainer" containerID="1056ca2a345418fd5bc8ddf9aa68669fec4dd75378086e5c207cb8f95ab40905" Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.739692 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vx842"] Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.752083 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vx842"] Feb 25 10:30:52 crc kubenswrapper[4978]: I0225 10:30:52.775776 4978 scope.go:117] "RemoveContainer" containerID="bc84361e7acbe7ef42d0c037aedf5263a8f70808fbd502db44113c2431ae4d79" Feb 25 10:30:53 crc kubenswrapper[4978]: I0225 10:30:53.340911 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" path="/var/lib/kubelet/pods/9c80d668-ae70-4010-a8a0-c3a74e0aa8b0/volumes" Feb 25 10:30:59 crc kubenswrapper[4978]: I0225 10:30:59.328624 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:30:59 crc kubenswrapper[4978]: E0225 10:30:59.329828 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:31:11 crc kubenswrapper[4978]: I0225 10:31:11.327650 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:31:11 crc kubenswrapper[4978]: E0225 10:31:11.328645 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:31:25 crc kubenswrapper[4978]: I0225 10:31:25.327796 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:31:25 crc kubenswrapper[4978]: E0225 10:31:25.328544 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:31:36 crc kubenswrapper[4978]: I0225 10:31:36.327458 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:31:36 crc kubenswrapper[4978]: E0225 10:31:36.328282 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:31:49 crc kubenswrapper[4978]: I0225 10:31:49.328011 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:31:49 crc kubenswrapper[4978]: E0225 10:31:49.328704 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.152911 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29533592-797xd"] Feb 25 10:32:00 crc kubenswrapper[4978]: E0225 10:32:00.153902 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerName="extract-utilities" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.153920 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerName="extract-utilities" Feb 25 10:32:00 crc kubenswrapper[4978]: E0225 10:32:00.153937 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerName="registry-server" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.153944 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerName="registry-server" Feb 25 10:32:00 crc kubenswrapper[4978]: E0225 10:32:00.153966 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerName="extract-content" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.153974 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerName="extract-content" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.154234 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c80d668-ae70-4010-a8a0-c3a74e0aa8b0" containerName="registry-server" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.155118 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533592-797xd" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.207802 4978 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-t7zdt" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.211767 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.216506 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533592-797xd"] Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.234726 4978 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.251962 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzj5v\" (UniqueName: \"kubernetes.io/projected/4eba01dd-e4de-42b9-b4b5-af399ca9fb9b-kube-api-access-vzj5v\") pod \"auto-csr-approver-29533592-797xd\" (UID: \"4eba01dd-e4de-42b9-b4b5-af399ca9fb9b\") " pod="openshift-infra/auto-csr-approver-29533592-797xd" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.354848 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzj5v\" (UniqueName: \"kubernetes.io/projected/4eba01dd-e4de-42b9-b4b5-af399ca9fb9b-kube-api-access-vzj5v\") pod \"auto-csr-approver-29533592-797xd\" (UID: \"4eba01dd-e4de-42b9-b4b5-af399ca9fb9b\") " pod="openshift-infra/auto-csr-approver-29533592-797xd" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.391406 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzj5v\" (UniqueName: \"kubernetes.io/projected/4eba01dd-e4de-42b9-b4b5-af399ca9fb9b-kube-api-access-vzj5v\") pod \"auto-csr-approver-29533592-797xd\" (UID: \"4eba01dd-e4de-42b9-b4b5-af399ca9fb9b\") " pod="openshift-infra/auto-csr-approver-29533592-797xd" Feb 25 10:32:00 crc kubenswrapper[4978]: I0225 10:32:00.536362 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533592-797xd" Feb 25 10:32:01 crc kubenswrapper[4978]: I0225 10:32:01.212015 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29533592-797xd"] Feb 25 10:32:01 crc kubenswrapper[4978]: I0225 10:32:01.460642 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533592-797xd" event={"ID":"4eba01dd-e4de-42b9-b4b5-af399ca9fb9b","Type":"ContainerStarted","Data":"fdd43393413dce2a5f1d7a8e9e4b1733ce55412378b98473cb4f628eb98b70ce"} Feb 25 10:32:02 crc kubenswrapper[4978]: I0225 10:32:02.328817 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:32:02 crc kubenswrapper[4978]: E0225 10:32:02.329042 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:32:03 crc kubenswrapper[4978]: I0225 10:32:03.482088 4978 generic.go:334] "Generic (PLEG): container finished" podID="4eba01dd-e4de-42b9-b4b5-af399ca9fb9b" containerID="e2462b9c7962a63bc8b136fbd94c2a5ac687ebdf7fbebca4905ce134241131b3" exitCode=0 Feb 25 10:32:03 crc kubenswrapper[4978]: I0225 10:32:03.482131 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533592-797xd" event={"ID":"4eba01dd-e4de-42b9-b4b5-af399ca9fb9b","Type":"ContainerDied","Data":"e2462b9c7962a63bc8b136fbd94c2a5ac687ebdf7fbebca4905ce134241131b3"} Feb 25 10:32:05 crc kubenswrapper[4978]: I0225 10:32:05.991200 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533592-797xd" Feb 25 10:32:06 crc kubenswrapper[4978]: I0225 10:32:06.081603 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzj5v\" (UniqueName: \"kubernetes.io/projected/4eba01dd-e4de-42b9-b4b5-af399ca9fb9b-kube-api-access-vzj5v\") pod \"4eba01dd-e4de-42b9-b4b5-af399ca9fb9b\" (UID: \"4eba01dd-e4de-42b9-b4b5-af399ca9fb9b\") " Feb 25 10:32:06 crc kubenswrapper[4978]: I0225 10:32:06.090694 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eba01dd-e4de-42b9-b4b5-af399ca9fb9b-kube-api-access-vzj5v" (OuterVolumeSpecName: "kube-api-access-vzj5v") pod "4eba01dd-e4de-42b9-b4b5-af399ca9fb9b" (UID: "4eba01dd-e4de-42b9-b4b5-af399ca9fb9b"). InnerVolumeSpecName "kube-api-access-vzj5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:32:06 crc kubenswrapper[4978]: I0225 10:32:06.184754 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzj5v\" (UniqueName: \"kubernetes.io/projected/4eba01dd-e4de-42b9-b4b5-af399ca9fb9b-kube-api-access-vzj5v\") on node \"crc\" DevicePath \"\"" Feb 25 10:32:06 crc kubenswrapper[4978]: I0225 10:32:06.517267 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29533592-797xd" event={"ID":"4eba01dd-e4de-42b9-b4b5-af399ca9fb9b","Type":"ContainerDied","Data":"fdd43393413dce2a5f1d7a8e9e4b1733ce55412378b98473cb4f628eb98b70ce"} Feb 25 10:32:06 crc kubenswrapper[4978]: I0225 10:32:06.517653 4978 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdd43393413dce2a5f1d7a8e9e4b1733ce55412378b98473cb4f628eb98b70ce" Feb 25 10:32:06 crc kubenswrapper[4978]: I0225 10:32:06.517355 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29533592-797xd" Feb 25 10:32:07 crc kubenswrapper[4978]: I0225 10:32:07.077794 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29533586-c6l5k"] Feb 25 10:32:07 crc kubenswrapper[4978]: I0225 10:32:07.091794 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29533586-c6l5k"] Feb 25 10:32:07 crc kubenswrapper[4978]: I0225 10:32:07.349710 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6da167fe-0017-4734-be04-9c649f761d27" path="/var/lib/kubelet/pods/6da167fe-0017-4734-be04-9c649f761d27/volumes" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.625338 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rjfq6"] Feb 25 10:32:13 crc kubenswrapper[4978]: E0225 10:32:13.627968 4978 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eba01dd-e4de-42b9-b4b5-af399ca9fb9b" containerName="oc" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.627989 4978 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eba01dd-e4de-42b9-b4b5-af399ca9fb9b" containerName="oc" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.628300 4978 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eba01dd-e4de-42b9-b4b5-af399ca9fb9b" containerName="oc" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.630572 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.647864 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rjfq6"] Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.741480 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ad6726-a12b-4f14-ac82-1b9fe5d0cb36-utilities\") pod \"redhat-marketplace-rjfq6\" (UID: \"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36\") " pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.741794 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ad6726-a12b-4f14-ac82-1b9fe5d0cb36-catalog-content\") pod \"redhat-marketplace-rjfq6\" (UID: \"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36\") " pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.741870 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrj87\" (UniqueName: \"kubernetes.io/projected/51ad6726-a12b-4f14-ac82-1b9fe5d0cb36-kube-api-access-mrj87\") pod \"redhat-marketplace-rjfq6\" (UID: \"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36\") " pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.844720 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ad6726-a12b-4f14-ac82-1b9fe5d0cb36-utilities\") pod \"redhat-marketplace-rjfq6\" (UID: \"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36\") " pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.844873 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ad6726-a12b-4f14-ac82-1b9fe5d0cb36-catalog-content\") pod \"redhat-marketplace-rjfq6\" (UID: \"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36\") " pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.844910 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrj87\" (UniqueName: \"kubernetes.io/projected/51ad6726-a12b-4f14-ac82-1b9fe5d0cb36-kube-api-access-mrj87\") pod \"redhat-marketplace-rjfq6\" (UID: \"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36\") " pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.845585 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ad6726-a12b-4f14-ac82-1b9fe5d0cb36-utilities\") pod \"redhat-marketplace-rjfq6\" (UID: \"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36\") " pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.845841 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ad6726-a12b-4f14-ac82-1b9fe5d0cb36-catalog-content\") pod \"redhat-marketplace-rjfq6\" (UID: \"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36\") " pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.883483 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrj87\" (UniqueName: \"kubernetes.io/projected/51ad6726-a12b-4f14-ac82-1b9fe5d0cb36-kube-api-access-mrj87\") pod \"redhat-marketplace-rjfq6\" (UID: \"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36\") " pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:13 crc kubenswrapper[4978]: I0225 10:32:13.963938 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:14 crc kubenswrapper[4978]: I0225 10:32:14.742802 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rjfq6"] Feb 25 10:32:15 crc kubenswrapper[4978]: I0225 10:32:15.614572 4978 generic.go:334] "Generic (PLEG): container finished" podID="51ad6726-a12b-4f14-ac82-1b9fe5d0cb36" containerID="2e0d9f36d0dda3b0559fdceedf0b332f79078ac6ed8ff5bc4facddbed52f241a" exitCode=0 Feb 25 10:32:15 crc kubenswrapper[4978]: I0225 10:32:15.615197 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rjfq6" event={"ID":"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36","Type":"ContainerDied","Data":"2e0d9f36d0dda3b0559fdceedf0b332f79078ac6ed8ff5bc4facddbed52f241a"} Feb 25 10:32:15 crc kubenswrapper[4978]: I0225 10:32:15.615222 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rjfq6" event={"ID":"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36","Type":"ContainerStarted","Data":"24358c1f4e7383acd70c103834154fd379df896b0936354cadab0a73f0cb720d"} Feb 25 10:32:16 crc kubenswrapper[4978]: I0225 10:32:16.328332 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:32:16 crc kubenswrapper[4978]: E0225 10:32:16.328726 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:32:20 crc kubenswrapper[4978]: I0225 10:32:20.686583 4978 generic.go:334] "Generic (PLEG): container finished" podID="51ad6726-a12b-4f14-ac82-1b9fe5d0cb36" containerID="2562572ece99001b125d843c7713b4750913aa4a2daf936991620ce3a254326b" exitCode=0 Feb 25 10:32:20 crc kubenswrapper[4978]: I0225 10:32:20.686686 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rjfq6" event={"ID":"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36","Type":"ContainerDied","Data":"2562572ece99001b125d843c7713b4750913aa4a2daf936991620ce3a254326b"} Feb 25 10:32:21 crc kubenswrapper[4978]: I0225 10:32:21.703088 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rjfq6" event={"ID":"51ad6726-a12b-4f14-ac82-1b9fe5d0cb36","Type":"ContainerStarted","Data":"4d07a6804890be36b37b3f5a288ac934c1cd43d65878af7f62a5f02c10888ff9"} Feb 25 10:32:21 crc kubenswrapper[4978]: I0225 10:32:21.746684 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rjfq6" podStartSLOduration=3.260333421 podStartE2EDuration="8.746658405s" podCreationTimestamp="2026-02-25 10:32:13 +0000 UTC" firstStartedPulling="2026-02-25 10:32:15.617430736 +0000 UTC m=+13629.056687205" lastFinishedPulling="2026-02-25 10:32:21.10375572 +0000 UTC m=+13634.543012189" observedRunningTime="2026-02-25 10:32:21.732926186 +0000 UTC m=+13635.172182675" watchObservedRunningTime="2026-02-25 10:32:21.746658405 +0000 UTC m=+13635.185914874" Feb 25 10:32:23 crc kubenswrapper[4978]: I0225 10:32:23.965711 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:23 crc kubenswrapper[4978]: I0225 10:32:23.966167 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.033384 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.297340 4978 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n6vjt"] Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.303600 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.321104 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n6vjt"] Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.483720 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5zb6\" (UniqueName: \"kubernetes.io/projected/b5a55801-d5ec-4111-8b56-77ec20b93592-kube-api-access-c5zb6\") pod \"community-operators-n6vjt\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.483978 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-catalog-content\") pod \"community-operators-n6vjt\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.484144 4978 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-utilities\") pod \"community-operators-n6vjt\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.586079 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-catalog-content\") pod \"community-operators-n6vjt\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.586143 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-utilities\") pod \"community-operators-n6vjt\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.586298 4978 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5zb6\" (UniqueName: \"kubernetes.io/projected/b5a55801-d5ec-4111-8b56-77ec20b93592-kube-api-access-c5zb6\") pod \"community-operators-n6vjt\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.586767 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-catalog-content\") pod \"community-operators-n6vjt\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.586802 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-utilities\") pod \"community-operators-n6vjt\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.616813 4978 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5zb6\" (UniqueName: \"kubernetes.io/projected/b5a55801-d5ec-4111-8b56-77ec20b93592-kube-api-access-c5zb6\") pod \"community-operators-n6vjt\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:24 crc kubenswrapper[4978]: I0225 10:32:24.638284 4978 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:25 crc kubenswrapper[4978]: I0225 10:32:25.398526 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n6vjt"] Feb 25 10:32:25 crc kubenswrapper[4978]: W0225 10:32:25.407865 4978 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5a55801_d5ec_4111_8b56_77ec20b93592.slice/crio-35e580c5d7086c8a32f6a0679506b54858326855bc635763906d1467f3e86937 WatchSource:0}: Error finding container 35e580c5d7086c8a32f6a0679506b54858326855bc635763906d1467f3e86937: Status 404 returned error can't find the container with id 35e580c5d7086c8a32f6a0679506b54858326855bc635763906d1467f3e86937 Feb 25 10:32:25 crc kubenswrapper[4978]: I0225 10:32:25.771245 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6vjt" event={"ID":"b5a55801-d5ec-4111-8b56-77ec20b93592","Type":"ContainerStarted","Data":"35e580c5d7086c8a32f6a0679506b54858326855bc635763906d1467f3e86937"} Feb 25 10:32:26 crc kubenswrapper[4978]: I0225 10:32:26.784501 4978 generic.go:334] "Generic (PLEG): container finished" podID="b5a55801-d5ec-4111-8b56-77ec20b93592" containerID="856d1ed7be404670eb356521fde0eb73b1a96c89adda8653f8d6d8f958238e00" exitCode=0 Feb 25 10:32:26 crc kubenswrapper[4978]: I0225 10:32:26.784709 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6vjt" event={"ID":"b5a55801-d5ec-4111-8b56-77ec20b93592","Type":"ContainerDied","Data":"856d1ed7be404670eb356521fde0eb73b1a96c89adda8653f8d6d8f958238e00"} Feb 25 10:32:28 crc kubenswrapper[4978]: I0225 10:32:28.805296 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6vjt" event={"ID":"b5a55801-d5ec-4111-8b56-77ec20b93592","Type":"ContainerStarted","Data":"69af9bbacef4193ceebbd8a48e28ba4dadb2ac67daf4a294171dd1bad0665937"} Feb 25 10:32:30 crc kubenswrapper[4978]: I0225 10:32:30.327686 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:32:30 crc kubenswrapper[4978]: E0225 10:32:30.327967 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:32:30 crc kubenswrapper[4978]: I0225 10:32:30.834486 4978 generic.go:334] "Generic (PLEG): container finished" podID="b5a55801-d5ec-4111-8b56-77ec20b93592" containerID="69af9bbacef4193ceebbd8a48e28ba4dadb2ac67daf4a294171dd1bad0665937" exitCode=0 Feb 25 10:32:30 crc kubenswrapper[4978]: I0225 10:32:30.834561 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6vjt" event={"ID":"b5a55801-d5ec-4111-8b56-77ec20b93592","Type":"ContainerDied","Data":"69af9bbacef4193ceebbd8a48e28ba4dadb2ac67daf4a294171dd1bad0665937"} Feb 25 10:32:32 crc kubenswrapper[4978]: I0225 10:32:32.860037 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6vjt" event={"ID":"b5a55801-d5ec-4111-8b56-77ec20b93592","Type":"ContainerStarted","Data":"cef92814ca2bcd3f218f95bda29f62d5cfe31c87fb377f41818c571e765e9895"} Feb 25 10:32:32 crc kubenswrapper[4978]: I0225 10:32:32.887494 4978 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n6vjt" podStartSLOduration=3.925477718 podStartE2EDuration="8.887471487s" podCreationTimestamp="2026-02-25 10:32:24 +0000 UTC" firstStartedPulling="2026-02-25 10:32:26.786462118 +0000 UTC m=+13640.225718577" lastFinishedPulling="2026-02-25 10:32:31.748455877 +0000 UTC m=+13645.187712346" observedRunningTime="2026-02-25 10:32:32.884621578 +0000 UTC m=+13646.323878057" watchObservedRunningTime="2026-02-25 10:32:32.887471487 +0000 UTC m=+13646.326727946" Feb 25 10:32:34 crc kubenswrapper[4978]: I0225 10:32:34.010114 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rjfq6" Feb 25 10:32:34 crc kubenswrapper[4978]: I0225 10:32:34.086753 4978 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rjfq6"] Feb 25 10:32:34 crc kubenswrapper[4978]: I0225 10:32:34.146171 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgg7r"] Feb 25 10:32:34 crc kubenswrapper[4978]: I0225 10:32:34.146464 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dgg7r" podUID="e4a76118-7308-4b0f-aae5-41cd6888a845" containerName="registry-server" containerID="cri-o://5cbfd50bb21d245634688454de6092339a95aed5c8edfda3ec1af5e802077baf" gracePeriod=2 Feb 25 10:32:34 crc kubenswrapper[4978]: I0225 10:32:34.638673 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:34 crc kubenswrapper[4978]: I0225 10:32:34.638714 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:34 crc kubenswrapper[4978]: I0225 10:32:34.693574 4978 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:34 crc kubenswrapper[4978]: I0225 10:32:34.931674 4978 generic.go:334] "Generic (PLEG): container finished" podID="e4a76118-7308-4b0f-aae5-41cd6888a845" containerID="5cbfd50bb21d245634688454de6092339a95aed5c8edfda3ec1af5e802077baf" exitCode=0 Feb 25 10:32:34 crc kubenswrapper[4978]: I0225 10:32:34.932846 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgg7r" event={"ID":"e4a76118-7308-4b0f-aae5-41cd6888a845","Type":"ContainerDied","Data":"5cbfd50bb21d245634688454de6092339a95aed5c8edfda3ec1af5e802077baf"} Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.748837 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.902158 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn94s\" (UniqueName: \"kubernetes.io/projected/e4a76118-7308-4b0f-aae5-41cd6888a845-kube-api-access-sn94s\") pod \"e4a76118-7308-4b0f-aae5-41cd6888a845\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.902233 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-catalog-content\") pod \"e4a76118-7308-4b0f-aae5-41cd6888a845\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.902713 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-utilities\") pod \"e4a76118-7308-4b0f-aae5-41cd6888a845\" (UID: \"e4a76118-7308-4b0f-aae5-41cd6888a845\") " Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.903120 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-utilities" (OuterVolumeSpecName: "utilities") pod "e4a76118-7308-4b0f-aae5-41cd6888a845" (UID: "e4a76118-7308-4b0f-aae5-41cd6888a845"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.903412 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.919598 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4a76118-7308-4b0f-aae5-41cd6888a845-kube-api-access-sn94s" (OuterVolumeSpecName: "kube-api-access-sn94s") pod "e4a76118-7308-4b0f-aae5-41cd6888a845" (UID: "e4a76118-7308-4b0f-aae5-41cd6888a845"). InnerVolumeSpecName "kube-api-access-sn94s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.931037 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e4a76118-7308-4b0f-aae5-41cd6888a845" (UID: "e4a76118-7308-4b0f-aae5-41cd6888a845"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.945904 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dgg7r" event={"ID":"e4a76118-7308-4b0f-aae5-41cd6888a845","Type":"ContainerDied","Data":"bea11dacbf3072b566b229e9a09f5d37cd7f3247015fff92cec2410b75e42c3e"} Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.945952 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dgg7r" Feb 25 10:32:35 crc kubenswrapper[4978]: I0225 10:32:35.945974 4978 scope.go:117] "RemoveContainer" containerID="5cbfd50bb21d245634688454de6092339a95aed5c8edfda3ec1af5e802077baf" Feb 25 10:32:36 crc kubenswrapper[4978]: I0225 10:32:36.004960 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn94s\" (UniqueName: \"kubernetes.io/projected/e4a76118-7308-4b0f-aae5-41cd6888a845-kube-api-access-sn94s\") on node \"crc\" DevicePath \"\"" Feb 25 10:32:36 crc kubenswrapper[4978]: I0225 10:32:36.004994 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e4a76118-7308-4b0f-aae5-41cd6888a845-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:32:36 crc kubenswrapper[4978]: I0225 10:32:36.006317 4978 scope.go:117] "RemoveContainer" containerID="64c441d2520e340c4c91ad918adeec0be7d87420a2b916096c443dd6208f6e81" Feb 25 10:32:36 crc kubenswrapper[4978]: I0225 10:32:36.024074 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgg7r"] Feb 25 10:32:36 crc kubenswrapper[4978]: I0225 10:32:36.039813 4978 scope.go:117] "RemoveContainer" containerID="0920ec34fa9aa9b1576d58fbbdfa0613a2278d444564b9929aa479fb7fec7bfa" Feb 25 10:32:36 crc kubenswrapper[4978]: I0225 10:32:36.045975 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dgg7r"] Feb 25 10:32:37 crc kubenswrapper[4978]: I0225 10:32:37.337970 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4a76118-7308-4b0f-aae5-41cd6888a845" path="/var/lib/kubelet/pods/e4a76118-7308-4b0f-aae5-41cd6888a845/volumes" Feb 25 10:32:41 crc kubenswrapper[4978]: I0225 10:32:41.327758 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:32:41 crc kubenswrapper[4978]: E0225 10:32:41.328311 4978 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-j496h_openshift-machine-config-operator(a5f01015-5dea-4e59-a9fc-326c84b85aed)\"" pod="openshift-machine-config-operator/machine-config-daemon-j496h" podUID="a5f01015-5dea-4e59-a9fc-326c84b85aed" Feb 25 10:32:44 crc kubenswrapper[4978]: I0225 10:32:44.728141 4978 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:44 crc kubenswrapper[4978]: I0225 10:32:44.794607 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n6vjt"] Feb 25 10:32:45 crc kubenswrapper[4978]: I0225 10:32:45.031628 4978 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n6vjt" podUID="b5a55801-d5ec-4111-8b56-77ec20b93592" containerName="registry-server" containerID="cri-o://cef92814ca2bcd3f218f95bda29f62d5cfe31c87fb377f41818c571e765e9895" gracePeriod=2 Feb 25 10:32:45 crc kubenswrapper[4978]: I0225 10:32:45.331009 4978 scope.go:117] "RemoveContainer" containerID="063b61665c2ee996631b73afe9b0f055c2b932e6268e0772639d782937225569" Feb 25 10:32:46 crc kubenswrapper[4978]: I0225 10:32:46.044025 4978 generic.go:334] "Generic (PLEG): container finished" podID="b5a55801-d5ec-4111-8b56-77ec20b93592" containerID="cef92814ca2bcd3f218f95bda29f62d5cfe31c87fb377f41818c571e765e9895" exitCode=0 Feb 25 10:32:46 crc kubenswrapper[4978]: I0225 10:32:46.044097 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6vjt" event={"ID":"b5a55801-d5ec-4111-8b56-77ec20b93592","Type":"ContainerDied","Data":"cef92814ca2bcd3f218f95bda29f62d5cfe31c87fb377f41818c571e765e9895"} Feb 25 10:32:46 crc kubenswrapper[4978]: I0225 10:32:46.791294 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:46 crc kubenswrapper[4978]: I0225 10:32:46.926866 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-catalog-content\") pod \"b5a55801-d5ec-4111-8b56-77ec20b93592\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " Feb 25 10:32:46 crc kubenswrapper[4978]: I0225 10:32:46.927064 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-utilities\") pod \"b5a55801-d5ec-4111-8b56-77ec20b93592\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " Feb 25 10:32:46 crc kubenswrapper[4978]: I0225 10:32:46.927228 4978 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5zb6\" (UniqueName: \"kubernetes.io/projected/b5a55801-d5ec-4111-8b56-77ec20b93592-kube-api-access-c5zb6\") pod \"b5a55801-d5ec-4111-8b56-77ec20b93592\" (UID: \"b5a55801-d5ec-4111-8b56-77ec20b93592\") " Feb 25 10:32:46 crc kubenswrapper[4978]: I0225 10:32:46.928190 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-utilities" (OuterVolumeSpecName: "utilities") pod "b5a55801-d5ec-4111-8b56-77ec20b93592" (UID: "b5a55801-d5ec-4111-8b56-77ec20b93592"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:32:46 crc kubenswrapper[4978]: I0225 10:32:46.934773 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5a55801-d5ec-4111-8b56-77ec20b93592-kube-api-access-c5zb6" (OuterVolumeSpecName: "kube-api-access-c5zb6") pod "b5a55801-d5ec-4111-8b56-77ec20b93592" (UID: "b5a55801-d5ec-4111-8b56-77ec20b93592"). InnerVolumeSpecName "kube-api-access-c5zb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 25 10:32:46 crc kubenswrapper[4978]: I0225 10:32:46.981848 4978 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5a55801-d5ec-4111-8b56-77ec20b93592" (UID: "b5a55801-d5ec-4111-8b56-77ec20b93592"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.030090 4978 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.030125 4978 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5a55801-d5ec-4111-8b56-77ec20b93592-utilities\") on node \"crc\" DevicePath \"\"" Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.030135 4978 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5zb6\" (UniqueName: \"kubernetes.io/projected/b5a55801-d5ec-4111-8b56-77ec20b93592-kube-api-access-c5zb6\") on node \"crc\" DevicePath \"\"" Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.058518 4978 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6vjt" Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.058510 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6vjt" event={"ID":"b5a55801-d5ec-4111-8b56-77ec20b93592","Type":"ContainerDied","Data":"35e580c5d7086c8a32f6a0679506b54858326855bc635763906d1467f3e86937"} Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.058652 4978 scope.go:117] "RemoveContainer" containerID="cef92814ca2bcd3f218f95bda29f62d5cfe31c87fb377f41818c571e765e9895" Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.078717 4978 scope.go:117] "RemoveContainer" containerID="69af9bbacef4193ceebbd8a48e28ba4dadb2ac67daf4a294171dd1bad0665937" Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.100763 4978 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n6vjt"] Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.115142 4978 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n6vjt"] Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.119178 4978 scope.go:117] "RemoveContainer" containerID="856d1ed7be404670eb356521fde0eb73b1a96c89adda8653f8d6d8f958238e00" Feb 25 10:32:47 crc kubenswrapper[4978]: I0225 10:32:47.341690 4978 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5a55801-d5ec-4111-8b56-77ec20b93592" path="/var/lib/kubelet/pods/b5a55801-d5ec-4111-8b56-77ec20b93592/volumes" Feb 25 10:32:53 crc kubenswrapper[4978]: I0225 10:32:53.328061 4978 scope.go:117] "RemoveContainer" containerID="45892f68947c731fef191827fc5d51a85f24d63847d58e2e80a3b079e0f4e988" Feb 25 10:32:54 crc kubenswrapper[4978]: I0225 10:32:54.170144 4978 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-j496h" event={"ID":"a5f01015-5dea-4e59-a9fc-326c84b85aed","Type":"ContainerStarted","Data":"28e4682c24820a4d63eab6cf1c1fd975f899b5e66b89aa201430d25d6e788702"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515147550036024453 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015147550036017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015147514250016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015147514250015461 5ustar corecore